Optimal Tests of Hypotheses (Hogg Chapter Eight)
|
|
- Lilian Byrd
- 5 years ago
- Views:
Transcription
1 Optimal Tests of Hypotheses Hogg hapter Eight STAT 406-0: Mathematical Statistics II Spring Semester 06 ontents Most Powerful Tests. Review of Hypothesis Testing The Neyman-Pearson Lemma Example: Gamma Distribution Uniformly Most Powerful Tests Example: One-Sided Hypothesis Example: Two-Sided Hypothesis Example: omposite Null Hypothesis The Karlin-Rubin Theorem Example: Exponential Family Likelihood Ratio Tests and UMP Unbiased Tests 7. Gamma Example ontinued Unbiased Tests Illustration for Gamma example omposite Hypothesis Testing with Priors 0 3. Example: Mean of a Normal Distribution ase : Uniform Prior opyright 06, John T. Whelan, and all that 3.. ase : Non-Uniform Prior Tuesday 6 April 06 Read Section 8. of Hogg Most Powerful Tests We turn now to another application of likelihood methods: as tests for hypotheses. First, a reminder of some of the notation and definitions associated with classical frequentist hypothesis testing.. Review of Hypothesis Testing In the classical framework, we define a test to reject or not reject a null hypothesis H 0 in the context of an alternative hypothesis H, based on a realization x of the n-dimensional random vector X which is often but not always a sample drawn from a univariate distribution. We refer to the support space for the whole sample as S, i.e., X S. We can define the test as partitioning the sample space into a critical region and its compelement.
2 If x S we reject H 0, while if x / i.e., x c we do not reject H 0. Strictly speaking, we should not refer to the outcome of the test as accepting one hypothesis or the other. We often think of this in terms of a parametrized distribution f X x; θ and define H 0 to specify a value θ = θ 0 Hogg calls this θ for the parameter and H to specify some other value θ = θ which Hogg calls θ. Since each hypothesis only specifies a single value for θ, we call these point hypotheses, and the parameter formalism is actually unneccessary; f X x; θ 0 = f X x H 0 and f X x; θ = f X x H can just be thought of as two different probsbility distributions relevant to H 0 and H, respectively. We define the size α of the critical region, also known as the significance or false-alarm probability of the test as the probability that we will reject H 0 if it is true: α = P X H 0 = f X x H 0 d n x. The power of the test is the probability that we ll reject H 0 if H is true: γ = P X H = f X x H d n x. We can also write this in terms of β = γ, which is known as the false dismissal probability. Given different tests with the same false-alarm probability α, we d naturally rather use the one with the lowest false-dismissal probability β = γ, i.e., the highest power γ. This is known as the most powerful test.. The Neyman-Pearson Lemma There is a theorem, usually known as the Neyman-Pearson lemma, that shows how the most powerful test to of one point hypothesis H 0 against another H can be constructed from the likelihood ratio Λx = f Xx H 0 f X x H = Lθ 0; x Lθ ; x.3 We define so that x if and only if Λx k where k is defined by P ΛX k H 0 = f X x H 0 d n x = α.4 which ensures that the critical region is of size α. I.e., we reject H 0 if and only if Λx k. The Neyman-Pearson lemma states that the power of this test γ = P ΛX k H = f X x H d n x.5 is greater than or equal to the power of any other test with the same significance. I.e., if A is some other critical region with size α, so that f X x H 0 d n x = α.6 the Neyman-Pearson lemma says that γ = f X x H d n x A A f X x H d n x.7 The demonstration of the Neyman-Pearson lemma involves breaking up the regions and A in terms of their overlap A. Evidentally, we can write = A c A A = c A A.8a.8b
3 The contribution to both α and γ from A cancel out of any comparison between and A. So the Neyman-Pearson lemma is equivalent to the condition that γ f X x H d n x A = A c f X x H d n x c A If we can prove that, we ve proved the lemma Now, by definition so while so A c c A f X x H 0 f X x H f X x H d n x f X x H d n x.9 k for x.0 f X x H 0 d n x. k A c f X x H 0 f X x H k for x c. f X x H d n x k c A f X x H 0 d n x.3 But since the tests defined by and A both have the same significance α, the integrals of f X x H 0 over the non-overlapping regions must be the same: α f X x H 0 d n x A = A c f X x H 0 d n x = c A f X x H 0 d n x.4 so the right-hand sides of.3 and. must be equal, which means f X x H d n x f X x H d n x.5 A c c A which, as we ve argued above, means that the power of the likelihood ratio test defined by is greater than or equal to that defined by A..3 Example: Gamma Distribution fx; θ = x5 θ 5 e x/θ, 0 < x <. Let H 0 : θ = and H : θ =. Likelihood is so likelihood ratio is Λ = Lθ; x = n i= x i 4 exp θ 5n n x i /θ i=.6 [ L; x L; x = 5n exp ] n x i = 5n e y/.7 i= where y = n i= x i. Note that the likelihood ratio depends only on the sufficient statistic y. A moment s thought will show that this will be true in general. If there are one or more sufficient statistics for the parameters which distinguish H 0 from H, the likelihood ratio must be a function of those alone. Note that in this case, Y is Gamman, of H 0 is true and Gamman, of H is true, so we can figure out the significance and power of the test by using the percentiles of the Gamma or equivalently chi-square distribution. 3
4 Thursday 8 April 06 Read Section 8. of Hogg.4 Uniformly Most Powerful Tests The Neyman-Pearson Lemma shows that when comparing point hypotheses H 0 and H, a test based on the likelihood ratio f X x H 0 /f X x H gives the most powerful test highest γ = P X H at a given significance α = P X H 0. We often wish to compare composite hypotheses, in which H and/or H 0 can correspond to parametrized families of distributions. In particular, if X is a sample drawn from a distribution fx; θ, we may have hypotheses H 0 : θ ω and H : θ Ω. If we first consider the case where H 0 is a point hypothesis θ = θ 0 and H is a composite hypothesis parametrized by θ, the power of a given test will depend on θ: γθ = P X ; θ = f X x; θd n x.8 If the same test is most powerful at a given α for every choice of θ, we say it is uniformly most powerful UMP. For example, recall the example from the last class, where X is a sample of size n from a Gamma5, θ distribution and H 0 is θ = θ 0 =. Now let H be the composite hypothesis θ >. We know from the Neyman-Pearson lemma that at each θ the most powerful test will be the one given by the likelihood ratio: : Λx; θ = Lθ 0; x Lθ; x where the constant kθ is defined by kθ.9 P ΛX; θ kθ H 0 = α.0.4. Example: One-Sided Hypothesis We saw last time that Lθ; x = n i= x4 i θ 5n e y/θ. where y is the sufficient statistic y = n i= x i, and Y is a Gamma5n, θ random variable. The likelihood ratio is Λθ; x = Lθ 5n [ 0; x θ Lθ; x = exp ] y θ 0 θ 0 θ = θ 5n exp θ. y θ As long as θ >, this is a decreasing function of y, so a test which rejects H 0 if Λ kθ will be equivalent to one which does so if Y a for some corresponding a. This can be defined so that the critical region is of size α: α = P Y a H 0 = γθ 0 = where Γs, x = a x y 5n Γ5nθ 5n 0 e y/θ 0 dy = Γ5n, a Γ5.3 t s e t dt.4 is the upper incomplete Gamma function. In any event, the value of a doesn t depend on θ, as long as θ >, so Y a is the UMP test of H 0 in light of the composite hypothesis H..4. Example: Two-Sided Hypothesis Note that there does not always exist a UMP test. Suppose instead the hypothesis H is θ θ 0 =. For θ >, the most 4
5 powerful test is to reject H 0 if Y a as defined above. For 0 < θ <, the likelihood ratio θ Λθ; x = θ 5n exp y.5 θ is a monotonically increasing function of y, so Λ kθ is equivalent to Y b, where b is defined by α = P Y b H 0 = γθ 0 = where b Γs Γs, x = y 5n Γ5nθ 5n 0 x e y/θ 0 dy = Γ5n Γ5n, b Γ5n.6 t s e t dt.7 is the lower incomplete Gamma function. Thus the most powerful test depends on the value of θ: rejecting H 0 when Y a is most powerful if θ > θ 0, while rejecting H when Y b is most powerful if θ < θ 0, To illustrate, we ll plot the power function for the two tests assuming α = 0.0 and n = 4. We ll get the percentiles and tail probabilities for the Gamma distribution using the SciPy stats package. For any distribution, the method sf is the survival function one minus the cdf, cdf is the cdf, and isf and ppf are the inverses of these functions. In []: from scipy.stats import gamma as gammadist In []: alpha = 0.0 In [3]: n = 4 In [4]: a = gammadist5*n.isfalpha In [5]: b = gammadist5*n.ppfalpha We check that the significance of the two tests is indeed 0%: In [6]: print gammadist5*n.sfa 0. In [7]: print gammadist5*n.cdfb 0. In [8]: theta = linspace0,3,000[:] The [:] is to omit the first element in the array, so that we re not actually trying to use θ = 0 as a value of the scale parameter. In [9]: gamma_a = gammadist5*n,scale=theta.sfa In [0]: gamma_b = gammadist5*n,scale=theta.cdfb In []: figure; In []: plottheta,gamma_a,'b-',label=r'$y\geq a$'; In [3]: plottheta,gamma_b,'r--',lw=,label=r'$y\leq b$'; In [4]: legendloc='center right'; In [5]: xlabelr'$\theta$'; In [6]: ylabelr'$\gamma\theta$'; In [7]: gridtrue In [8]: savefig'notes08_power.eps',bbox_inches='tight' Here are the two power curves: 5
6 γθ Y a Y b If we consider the test Y a, P X ; θ = a y 5n Γ5nθ 5n e y/θ dy = a/θ t 5n Γ5n e t dt.9 Since the integrand is positive definite and independent of θ, we see that we can maximize the integral by making θ as large as possible subject to H 0 so that the lower limit a/θ is as small as possible and the integral is maximized, i.e., θ = θ 0 =. So the sigificance of the test for null hypothesis H 0 : θ θ 0 = is the same as for H 0 : θ = θ 0 =, and the rest of the problem proceeds as before, i.e., Y a defines the UMP for H 0 : θ vs H : θ >..4.4 The Karlin-Rubin Theorem Note that they intersect at γ = 0., which makes sense since any test with significance α will satisfy γθ 0 = α. And as we can see, the power of the test Y a solid blue is higher for θ >, while the power of the test Y b dashed red is higher for θ <..4.3 Example: omposite Null Hypothesis Returning to the case where H is θ >, we can extend the discussion to comparison between two composite hypotheses by letting H 0 be θ. As a matter of convention, we define the significance α for a test with a composite null hypothesis as the worst-case false alarm rate: θ α = max P X ; θ.8 θ ω This is example is an illustration of a result known as the Karlin- Rubin theorem, which extends the Neyman-Pearson lemma to tests of one-sided composite hypotheses. If the likelihood function Lθ; x can be written in terms of a single sufficient statistic y = ux and has the property that, for any θ 0 and θ in the parameter space and satisfying θ 0 < θ, the likelihood ratio Λx; θ 0, θ = Lθ 0; θ Lθ ; θ.30 is a monotonically increasing or monotonically decreasing function of y, then there exists a uniformly most powerful test for H 0 : θ θ versus H : θ > θ. We call this situation a monotone likelihood ratio mlr..4.5 Example: Exponential Family The conditions for a monotone likelihood ratio seem a little restrictive, but there is a decent class of models which satisfy them. 6
7 If we consider a distribution which is a member of the regular exponential family, so fx; θ = exp ηθkx + Hx + qθ.3 then the likelihood for a sample of size n is Lθ; x = exp ηθ n Kx i + i= and the likelihood ratio is n Hx i + nqθ i=.3 Λx; θ 0, θ = Lθ 0; θ Lθ ; θ = exp [ηθ 0 ηθ ]y + n[qθ 0 qθ ].33 where the dependence on the data is via the sufficient statistic y = n i= Kx i. We see that: If ηθ is a monotonically increasing function of θ, then ηθ 0 ηθ < 0 for θ 0 < θ, and Λx; θ 0, θ is a monotonically decreasing function of y. If ηθ is a monotonically decreasing function of θ, then ηθ 0 ηθ > 0 for θ 0 < θ, and Λx; θ 0, θ is a monotonically increasing function of y. I.e., if ηθ is a monotone function of θ, we have a monotone likelihood ratio, and the Karlin-Rubin theorem tells us there s a UMP test to distinguish between one-sided hypotheses. Tuesday 3 May 06 Read Section 8.3 of Hogg Likelihood Ratio Tests and UMP Unbiased Tests. Gamma Example ontinued Recall that last time, we saw that, for a sample of size n drawn from a Gamma5, θ distribution, a test which rejected H 0 : θ = θ 0 = if n i= X i a was uniformly most powerful for comparing H 0 to the one-sided alternative hypothesis θ > θ 0 =, while one with rejected H 0 if n i= X i b was UMP if the alternative hypothesis was θ < θ 0 =, but if the alternative hypothesis was H = θ θ 0 =, there was no single test which was most powerful for every θ allowed under H. One method we ve considered before for comparing composite hypotheses is a likelihood ratio test, where we maximize the likelihood subject to each hypothesis. In this case, where the null hypothesis is still the point hypothesis θ = θ 0, this becomes specifically in this example, Λx = Lθ 0; x L θx, x Lθ; x =. n i= x4 i [Γ5] n θ 5n e y/θ. and the maximum likelihood solution θ is the solution to 0 = θ 5n ln θ y = 5n θ θ + y.3 θ 7
8 i.e., θ = y, which makes the likelihood ratio 5n Λx = L; x y 5n L y, x = e y+5n 5n 5n.4 Since this goes to zero as y goes to zero or infinity, a test which rejects H 0 if Λx k will be a two-sided test on Y, rejecting if Y c or Y c.5 where y = c and y = c correspond to equal values of the likelihood, i.e., c 5n e c = c 5n e c.6 which, together with the specified significance α determines c and c via α = c c f Y y; dy = Γ5n c c y 5n e y dy.7 Note that this likelihood ratio test is slightly different from the equal-tailed test defined by c sym 0 f Y y; dy = α = c sym f Y y; dy.8 which is slightly easier to construct. c sym and c sym can be written as the 00αth and 00 αth percentiles of the Gamma5n, distribution which applies to Y under the null hypothesis H 0. If the distribution function f Y y; and the likelihood ratio Λx were symmetric functions in y, of course, the likelihood ratio test would be equal-tailed. This is the case in several of the examples in Hogg. We can examine the power functions of the different options the two one-sided tests, the equal-tailed test, and the likelihood ratio test by continuing our plotting example from last time: In [9]: c_sym = gammadist5*n.isf0.5*alpha In [0]: c_sym = gammadist5*n.ppf0.5*alpha In []: gamma_sym =...: gammadist5*n,scale=theta.cdfc_sym...: + gammadist5*n,scale=theta.sfc_sym In []: plottheta,gamma_sym,'k:',lw=,...: label=r'$y\leq c_^{\rm sym}$ '...: + r'or $Y\geq c_^{\rm sym}$'; For the maximum likelihood test, we have to find the test which satisfies.6 from among the choices for which P Y c H 0 + P Y c H 0 = α: In [3]: alphaleft = linspace0,alpha,000[:-] In [4]: c = gammadist5*n.ppfalphaleft In [5]: c = gammadist5*n.isfalpha-alphaleft c and c are both 998-element arrays giving the lower and uppwe end of the range of y values for which we should not reject H 0. Rather than doing some sophisticated root-finding, we just brute-force it and pick the pair which has the lowest value of c 5n e c c 5n e c : In [6]: ind_ml = argminabsc**5*n*exp-c...: -c**5*n*exp-c In [7]: c_ml = c[ind_ml] In [8]: c_ml = c[ind_ml] 8
9 In [9]: gamma_ml =...: gammadist5*n,scale=theta.cdfc_ml...: + gammadist5*n,scale=theta.sfc_ml In [30]: plottheta,gamma_ml,'g-',...: label=r'$y\leq c_^{\rm ML}$ '...: + r'or $Y\geq c_^{\rm ML}$'; In [3]: legendloc='center right'; In [3]: savefig'notes08_ml.eps',bbox_inches='tight' γθ Y a Y b Y c sym or Y c sym Y c ML or Y c ML function that goes to unity as you go away from θ = θ 0 in either direction, as opposed to the one-sided tests, whose power function goes to zero on the wrong side of θ = θ 0.. Unbiased Tests One of the unsatisfactory properties of the one-sided hypothesis tests is that while they are the most powerful tests for some ranges of the parameter θ, for other ranges, the power is actually below the false alarm probability α. For instance, the test which rejects H 0 when Y a has γθ < α when θ < θ 0 =. This means the test is more likely to reject H 0 when it s true than when H is true, if it happens that θ <. A test with this undesirable property is called biased. onversely, an unbiased test is one for which γθ α for all θ allowed by H. It is often the case that, while no uniformly most powerful test exists, there is an unbiased test which, at all values of θ, is more powerful than any other unbiased test. This is known as the uniformly most powerful unbiased test. It can be shown that in many cases specifically the regular exponential family, the UMPU test is a two-sided test on the sufficient statistic Y, which rejects H 0 if Y c or Y c We see that the equal-tailed test is slightly different from the maximum likelihood test, thanks to the asymmetric distribution, but it s close. Both have the property of having a power θ.. Illustration for Gamma example Returning to the example of a sample from the Gamma5, θ distribution, consider a test which satisfies γθ α when θ θ 0 =. Since γθ 0 = α, that means the power function γθ must have a minimum at θ = θ 0. Recalling that 9
10 Y Gamma5n, θ, we have 0.0 γθ = c = Γ5n f Y y; θ dy = c Γ5n c /θ c /θ t 5n e t dt c c y 5n e y/θ dy θ 5n.9 where we have made the substitution t = y/θ in the last step. Requiring this to be a minimum at θ = θ 0 gives us 0 = γ θ 0 = c c 5n Γ5n θ0 which gives us, using θ 0 =, e c c θ 0 c 5n e c.0 γθ θ Y a Y b Y c sym or Y c sym Y c ML or Y c ML c 5n e c = c 5n e c. If we look back, we see this is just the condition.6 associated with the likelihood ratio test, so that test is in fact the UMP unbiased test for this model. We can verify this by zooming in on the power function plot above: In [33]: xlim[.95,.05]; In [34]: ylim[0.08,0.]; In [35]: savefig'notes08_mlzoom.eps',bbox_inches='tight' We see that the equal-tailed test has a power function which dips below α = 0.0 for a small range of θ values, so in fact that test is not unbiased. We see that the likelihood ratio test is indeed unbiased, and it turns out to tbe the UMPU test. Thursday 5 May 06 See Searle, 3 omposite Hypothesis Testing with Priors onsider now a slightly different situation. Suppose that the null hypothesis H 0 is a point hypothesis, but the alternative hypothesis H is a composite hypothesis which allows for a range 0
11 of values of the model parameters θ, but comes with a prior distribution f Θ θ H on those parameters. We include the possibility of a p-dimensional parameter space, but it may also be that p =. If we define a test which rejects H 0 when f X x H 0 f X x H = Lθ 0 ; x Lθ; x fθ θ H d p θ k 3. the Neyman-Pearson lemma tells us this is the most powerful test of H 0 versus H. This is most powerful in the sense of maximizing the power function γh = P X H = f X x H d n x = Lθ; x f Θ θ H d p θ d n x = γθ f Θ θ H d p θ A few points to note: 3. The test statistic in 3. is just the Bayes factor which we ve already motivated using Bayes s theorem in the form to write P H i x = f Xx H i P H i f X x P H 0 x P H x = f Xx H 0 P H 0 f X x H P H If there is a uniformly most powerful test which covers any θ in the support of f Θ θ H, this will also be the most powerful test of H 0 against H for any prior distribution. A possible objection is that the frequentist hypothesis testing formalism only applies to the outcomes of repeated experiments, and isn t supposed to know about any prior distribution. The preprint by Searle referenced above actually refers to the outcome of a Monte arlo experiment, where θ is also randomly generated along with the realization of X, so the relevant joint distribution is f XΘ x, θ H = f X x; θ f Θ θ H. In that context, the Bayes factor constructed using the same prior as the Monte arlo simulation gives the most powerful test, when attention is restricted to tests which define using only X and not Θ. 3. Example: Mean of a Normal Distribution Suppose that the data vector is X = X X but rather than being a sample from a given distribution, they are independent random variables X Nθ, and X Nθ,. Let the null hypothesis be H 0 : θ = 0 = θ and the alternative hypothesis be H : θ θ 0 0 Lθ; x = π exp. The likelihood function is x θ x θ 3.5 The maximum likelihood solution is θ = x and θ = x, so the likelihood ratio test would reject H 0 if L0, x Λx = L θ, x = exp x + x k 3.6 which means, for a test of significance α, rejecting H 0 if X + X χ,α 3.7 On the other hand, if we have a prior f Θ θ, θ H the Neyman-Pearson lemma tells us the optimal test statistic is the Bayes factor exp x +x f Θθ, θ H exp x θ x θ 3.8 dθ dθ
12 3.. ase : Uniform Prior Suppose that the prior f Θ θ, θ H is uniform in θ and θ. In that case the denominator is proportional to exp x θ x θ dθ dθ = constant 3.9 and the Bayes factor is proportional to the maximum-likelihood ratio exp x + x 3.0 so using x + x as a test statistic does give the optimal test in light of the prior. 3.. ase : Non-Uniform Prior It could be that the prior on θ and θ is more complicated. For instance, suppose f Θ θ, θ θ θ. Then the denominator of the Bayes factor is proportional to θθ exp x θ x θ dθ dθ + x + x 3. This is an improper prior, so technically the normalization factor will be zero and Bayes factor will be infinite, but we re only interested in the x dependence of the Bayes factor anyway. If we want to be careful, we can use a prior that is e.g., Gaussian with some large width, much greater than any of the other values in the problem. E.g., for the gravitational wave signal from a rotating or orbiting system, the amplitudes of the left- and right-circularly polarized parts of the signal are proportional to ±cos ι where ι is an inclination angle, which is generally unknown with a uniform prior on cos ι [, ], which produces a non-trivial prior distribution for the two amplitudes. See Whelan et al, lassical and Quantum Gravity 3, ; which means the Bayes factor is proportional to exp x +x + x + x and the optimal test rejects H 0 if 3. X + X + ln + X + ln + X c 3.3 We can show the difference between the two families of critical regions on a contour plot: In []: x = linspace-4,4,00 In []: xgrid, xgrid = meshgridx,x In [3]: Fgrid = xgrid** + xgrid** In [4]: Bgrid = Fgrid...: + *log+xgrid** + *log+xgrid** In [5]: xlevels = arange6 In [6]: Blevels = xlevels** + *log+xlevels** In [7]: Flevels = xlevels** In [8]: figurefigsize=5,5; In [9]: contourxgrid,xgrid,fgrid,flevels,colors='r',...: linewidths=,linestyles='dashed'; In [0]: contourxgrid,xgrid,bgrid,blevels,colors='b'; In []: gridtrue
13 In []: xlabelr'$x_$'; In [3]: ylabelr'$x_$'; In [4]: savefig'notes08_contours.eps',...: bbox_inches='tight' 4 3 to estimate the significance of the Bayes-factor test numerically. See figure 3 of Whelan et al 04 for an example of this. Tuesday 0 May 06 Review for Final Exam The exam is comprehensive, but with relatively more emphasis on chapter eight and the second half of chapter seven. Please come with questions and topics you d like to go over. Thursday May 06 Review for Final Exam The exam is comprehensive, but with relatively more emphasis on chapter eight and the second half of chapter seven. Please come with questions and topics you d like to go over. x x Note that we drew the contours at arbitrary levels; to draw them at the same significance for the two tests, we d probably need 3
Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes
Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis
More informationLecture 21. Hypothesis Testing II
Lecture 21. Hypothesis Testing II December 7, 2011 In the previous lecture, we dened a few key concepts of hypothesis testing and introduced the framework for parametric hypothesis testing. In the parametric
More informationHypothesis Testing - Frequentist
Frequentist Hypothesis Testing - Frequentist Compare two hypotheses to see which one better explains the data. Or, alternatively, what is the best way to separate events into two classes, those originating
More informationTheory of Statistical Tests
Ch 9. Theory of Statistical Tests 9.1 Certain Best Tests How to construct good testing. For simple hypothesis H 0 : θ = θ, H 1 : θ = θ, Page 1 of 100 where Θ = {θ, θ } 1. Define the best test for H 0 H
More informationDefinition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.
Hypothesis Testing Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Suppose the family of population distributions is indexed
More information10. Composite Hypothesis Testing. ECE 830, Spring 2014
10. Composite Hypothesis Testing ECE 830, Spring 2014 1 / 25 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve unknown parameters
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More information557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES
557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES Example Suppose that X,..., X n N, ). To test H 0 : 0 H : the most powerful test at level α is based on the statistic λx) f π) X x ) n/ exp
More informationHypothesis Testing: The Generalized Likelihood Ratio Test
Hypothesis Testing: The Generalized Likelihood Ratio Test Consider testing the hypotheses H 0 : θ Θ 0 H 1 : θ Θ \ Θ 0 Definition: The Generalized Likelihood Ratio (GLR Let L(θ be a likelihood for a random
More informationBEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.
BEST TESTS Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. 1. Most powerful test Let {f θ } θ Θ be a family of pdfs. We will consider
More informationInterval Estimation. Chapter 9
Chapter 9 Interval Estimation 9.1 Introduction Definition 9.1.1 An interval estimate of a real-values parameter θ is any pair of functions, L(x 1,..., x n ) and U(x 1,..., x n ), of a sample that satisfy
More informationSTAT 830 Hypothesis Testing
STAT 830 Hypothesis Testing Richard Lockhart Simon Fraser University STAT 830 Fall 2018 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 1 / 30 Purposes of These
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationHomework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.
Stat 300A Theory of Statistics Homework 7: Solutions Nikos Ignatiadis Due on November 28, 208 Solutions should be complete and concisely written. Please, use a separate sheet or set of sheets for each
More informationReview Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the
Review Quiz 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Cramér Rao lower bound (CRLB). That is, if where { } and are scalars, then
More informationMethods of evaluating tests
Methods of evaluating tests Let X,, 1 Xn be i.i.d. Bernoulli( p ). Then 5 j= 1 j ( 5, ) T = X Binomial p. We test 1 H : p vs. 1 1 H : p>. We saw that a LRT is 1 if t k* φ ( x ) =. otherwise (t is the observed
More informationλ(x + 1)f g (x) > θ 0
Stat 8111 Final Exam December 16 Eleven students took the exam, the scores were 92, 78, 4 in the 5 s, 1 in the 4 s, 1 in the 3 s and 3 in the 2 s. 1. i) Let X 1, X 2,..., X n be iid each Bernoulli(θ) where
More informationChapter 4. Theory of Tests. 4.1 Introduction
Chapter 4 Theory of Tests 4.1 Introduction Parametric model: (X, B X, P θ ), P θ P = {P θ θ Θ} where Θ = H 0 +H 1 X = K +A : K: critical region = rejection region / A: acceptance region A decision rule
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationSTAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015
STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis
More informationSTAT 830 Hypothesis Testing
STAT 830 Hypothesis Testing Hypothesis testing is a statistical problem where you must choose, on the basis of data X, between two alternatives. We formalize this as the problem of choosing between two
More informationEconomics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,
Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem
More informationCh. 5 Hypothesis Testing
Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,
More informationMAT 271E Probability and Statistics
MAT 71E Probability and Statistics Spring 013 Instructor : Class Meets : Office Hours : Textbook : Supp. Text : İlker Bayram EEB 1103 ibayram@itu.edu.tr 13.30 1.30, Wednesday EEB 5303 10.00 1.00, Wednesday
More informationUniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence
Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence Valen E. Johnson Texas A&M University February 27, 2014 Valen E. Johnson Texas A&M University Uniformly most powerful Bayes
More informationEco517 Fall 2004 C. Sims MIDTERM EXAM
Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering
More informationSTA 732: Inference. Notes 2. Neyman-Pearsonian Classical Hypothesis Testing B&D 4
STA 73: Inference Notes. Neyman-Pearsonian Classical Hypothesis Testing B&D 4 1 Testing as a rule Fisher s quantification of extremeness of observed evidence clearly lacked rigorous mathematical interpretation.
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationStatistics Ph.D. Qualifying Exam: Part I October 18, 2003
Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer
More informationSTAT 514 Solutions to Assignment #6
STAT 514 Solutions to Assignment #6 Question 1: Suppose that X 1,..., X n are a simple random sample from a Weibull distribution with density function f θ x) = θcx c 1 exp{ θx c }I{x > 0} for some fixed
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationHYPOTHESIS TESTING: FREQUENTIST APPROACH.
HYPOTHESIS TESTING: FREQUENTIST APPROACH. These notes summarize the lectures on (the frequentist approach to) hypothesis testing. You should be familiar with the standard hypothesis testing from previous
More informationStatistics Masters Comprehensive Exam March 21, 2003
Statistics Masters Comprehensive Exam March 21, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer
More informationSTAT 135 Lab 5 Bootstrapping and Hypothesis Testing
STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,
More informationPrimer on statistics:
Primer on statistics: MLE, Confidence Intervals, and Hypothesis Testing ryan.reece@gmail.com http://rreece.github.io/ Insight Data Science - AI Fellows Workshop Feb 16, 018 Outline 1. Maximum likelihood
More informationE. Santovetti lesson 4 Maximum likelihood Interval estimation
E. Santovetti lesson 4 Maximum likelihood Interval estimation 1 Extended Maximum Likelihood Sometimes the number of total events measurements of the experiment n is not fixed, but, for example, is a Poisson
More informationLecture 2: Basic Concepts of Statistical Decision Theory
EE378A Statistical Signal Processing Lecture 2-03/31/2016 Lecture 2: Basic Concepts of Statistical Decision Theory Lecturer: Jiantao Jiao, Tsachy Weissman Scribe: John Miller and Aran Nayebi In this lecture
More informationDirection: This test is worth 250 points and each problem worth points. DO ANY SIX
Term Test 3 December 5, 2003 Name Math 52 Student Number Direction: This test is worth 250 points and each problem worth 4 points DO ANY SIX PROBLEMS You are required to complete this test within 50 minutes
More informationLecture 12 November 3
STATS 300A: Theory of Statistics Fall 2015 Lecture 12 November 3 Lecturer: Lester Mackey Scribe: Jae Hyuck Park, Christian Fong Warning: These notes may contain factual and/or typographic errors. 12.1
More information7. Estimation and hypothesis testing. Objective. Recommended reading
7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing
More informationParameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn
Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation
More informationThe University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80
The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 71. Decide in each case whether the hypothesis is simple
More informationStatistical Inference
Statistical Inference Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Spring, 2006 1. DeGroot 1973 In (DeGroot 1973), Morrie DeGroot considers testing the
More informationMathematical Statistics
Mathematical Statistics MAS 713 Chapter 8 Previous lecture: 1 Bayesian Inference 2 Decision theory 3 Bayesian Vs. Frequentist 4 Loss functions 5 Conjugate priors Any questions? Mathematical Statistics
More informationLecture 16 November Application of MoUM to our 2-sided testing problem
STATS 300A: Theory of Statistics Fall 2015 Lecture 16 November 17 Lecturer: Lester Mackey Scribe: Reginald Long, Colin Wei Warning: These notes may contain factual and/or typographic errors. 16.1 Recap
More information40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology
Singapore University of Design and Technology Lecture 9: Hypothesis testing, uniformly most powerful tests. The Neyman-Pearson framework Let P be the family of distributions of concern. The Neyman-Pearson
More informationComposite Hypotheses and Generalized Likelihood Ratio Tests
Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve
More informationStatistical Theory MT 2007 Problems 4: Solution sketches
Statistical Theory MT 007 Problems 4: Solution sketches 1. Consider a 1-parameter exponential family model with density f(x θ) = f(x)g(θ)exp{cφ(θ)h(x)}, x X. Suppose that the prior distribution has the
More informationTopic 15: Simple Hypotheses
Topic 15: November 10, 2009 In the simplest set-up for a statistical hypothesis, we consider two values θ 0, θ 1 in the parameter space. We write the test as H 0 : θ = θ 0 versus H 1 : θ = θ 1. H 0 is
More informationAnswers to the 8th problem set. f(x θ = θ 0 ) L(θ 0 )
Answers to the 8th problem set The likelihood ratio with which we worked in this problem set is: Λ(x) = f(x θ = θ 1 ) L(θ 1 ) =. f(x θ = θ 0 ) L(θ 0 ) With a lower-case x, this defines a function. With
More informationStat 5421 Lecture Notes Fuzzy P-Values and Confidence Intervals Charles J. Geyer March 12, Discreteness versus Hypothesis Tests
Stat 5421 Lecture Notes Fuzzy P-Values and Confidence Intervals Charles J. Geyer March 12, 2016 1 Discreteness versus Hypothesis Tests You cannot do an exact level α test for any α when the data are discrete.
More informationLECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.
Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the
More informationHypothesis Testing. BS2 Statistical Inference, Lecture 11 Michaelmas Term Steffen Lauritzen, University of Oxford; November 15, 2004
Hypothesis Testing BS2 Statistical Inference, Lecture 11 Michaelmas Term 2004 Steffen Lauritzen, University of Oxford; November 15, 2004 Hypothesis testing We consider a family of densities F = {f(x; θ),
More informationStat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS
Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS 1a. Under the null hypothesis X has the binomial (100,.5) distribution with E(X) = 50 and SE(X) = 5. So P ( X 50 > 10) is (approximately) two tails
More informationLECTURE NOTES 57. Lecture 9
LECTURE NOTES 57 Lecture 9 17. Hypothesis testing A special type of decision problem is hypothesis testing. We partition the parameter space into H [ A with H \ A = ;. Wewrite H 2 H A 2 A. A decision problem
More information4.5.1 The use of 2 log Λ when θ is scalar
4.5. ASYMPTOTIC FORM OF THE G.L.R.T. 97 4.5.1 The use of 2 log Λ when θ is scalar Suppose we wish to test the hypothesis NH : θ = θ where θ is a given value against the alternative AH : θ θ on the basis
More informationMcGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper
McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section
More informationECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters
ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters D. Richard Brown III Worcester Polytechnic Institute 26-February-2009 Worcester Polytechnic Institute D. Richard Brown III 26-February-2009
More information6.867 Machine Learning
6.867 Machine Learning Problem set 1 Due Thursday, September 19, in class What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove.
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationMAT 271E Probability and Statistics
MAT 7E Probability and Statistics Spring 6 Instructor : Class Meets : Office Hours : Textbook : İlker Bayram EEB 3 ibayram@itu.edu.tr 3.3 6.3, Wednesday EEB 6.., Monday D. B. Bertsekas, J. N. Tsitsiklis,
More informationLecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary
ECE 830 Spring 207 Instructor: R. Willett Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we saw that the likelihood
More informationLecture 26: Likelihood ratio tests
Lecture 26: Likelihood ratio tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f θ0 (X) > c 0 for
More information4 Hypothesis testing. 4.1 Types of hypothesis and types of error 4 HYPOTHESIS TESTING 49
4 HYPOTHESIS TESTING 49 4 Hypothesis testing In sections 2 and 3 we considered the problem of estimating a single parameter of interest, θ. In this section we consider the related problem of testing whether
More informationBayesian Statistics (Hogg Chapter Eleven)
Bayesian Statistics (Hogg Chapter Eleven) STAT 406-0: Mathematical Statistics II Spring Semester 206 Contents 0 Preliminaries 2 0. Administrata.................... 2 0.2 Outline........................
More informationLecture Testing Hypotheses: The Neyman-Pearson Paradigm
Math 408 - Mathematical Statistics Lecture 29-30. Testing Hypotheses: The Neyman-Pearson Paradigm April 12-15, 2013 Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 1 / 12 Agenda Example:
More informationLine Broadening. φ(ν) = Γ/4π 2 (ν ν 0 ) 2 + (Γ/4π) 2, (3) where now Γ = γ +2ν col includes contributions from both natural broadening and collisions.
Line Broadening Spectral lines are not arbitrarily sharp. There are a variety of mechanisms that give them finite width, and some of those mechanisms contain significant information. We ll consider a few
More information5.1 Uniformly Most Accurate Families of Confidence
Chapter 5 Confidence Estimation Let (X, B X, P ), P {P θ θ Θ} be sample space and {Θ, B Θ, µ} be measure space with some σ finite measure µ. 5.1 Uniformly Most Accurate Families of Confidence Sets Def.
More informationMore Empirical Process Theory
More Empirical Process heory 4.384 ime Series Analysis, Fall 2008 Recitation by Paul Schrimpf Supplementary to lectures given by Anna Mikusheva October 24, 2008 Recitation 8 More Empirical Process heory
More informationHypothesis testing: theory and methods
Statistical Methods Warsaw School of Economics November 3, 2017 Statistical hypothesis is the name of any conjecture about unknown parameters of a population distribution. The hypothesis should be verifiable
More information14.30 Introduction to Statistical Methods in Economics Spring 2009
MIT OpenCourseWare http://ocw.mit.edu.30 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. .30
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More informationPartitioning the Parameter Space. Topic 18 Composite Hypotheses
Topic 18 Composite Hypotheses Partitioning the Parameter Space 1 / 10 Outline Partitioning the Parameter Space 2 / 10 Partitioning the Parameter Space Simple hypotheses limit us to a decision between one
More informationUNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences FINAL EXAMINATION, APRIL 2013
UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences FINAL EXAMINATION, APRIL 2013 STAB57H3 Introduction to Statistics Duration: 3 hours Last Name: First Name: Student number:
More informationECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing
ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing D. Richard Brown III Worcester Polytechnic Institute Worcester Polytechnic Institute D. Richard Brown III 1 / 8 Basics Hypotheses H 0
More informationStat 710: Mathematical Statistics Lecture 27
Stat 710: Mathematical Statistics Lecture 27 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 27 April 3, 2009 1 / 10 Lecture 27:
More informationStatistical Theory MT 2006 Problems 4: Solution sketches
Statistical Theory MT 006 Problems 4: Solution sketches 1. Suppose that X has a Poisson distribution with unknown mean θ. Determine the conjugate prior, and associate posterior distribution, for θ. Determine
More information2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?
ECE 830 / CS 76 Spring 06 Instructors: R. Willett & R. Nowak Lecture 3: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we
More information6.867 Machine Learning
6.867 Machine Learning Problem set 1 Solutions Thursday, September 19 What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove.
More informationDistribution-Free Procedures (Devore Chapter Fifteen)
Distribution-Free Procedures (Devore Chapter Fifteen) MATH-5-01: Probability and Statistics II Spring 018 Contents 1 Nonparametric Hypothesis Tests 1 1.1 The Wilcoxon Rank Sum Test........... 1 1. Normal
More informationLet us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided
Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or
More informationStatistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That
Statistics Lecture 4 August 9, 2000 Frank Porter Caltech The plan for these lectures: 1. The Fundamentals; Point Estimation 2. Maximum Likelihood, Least Squares and All That 3. What is a Confidence Interval?
More information1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D.
probabilities, we ll use Bayes formula. We can easily compute the reverse probabilities A short introduction to Bayesian statistics, part I Math 17 Probability and Statistics Prof. D. Joyce, Fall 014 I
More informationStatistical Data Analysis Stat 3: p-values, parameter estimation
Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,
More informationReview. December 4 th, Review
December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter
More informationST495: Survival Analysis: Hypothesis testing and confidence intervals
ST495: Survival Analysis: Hypothesis testing and confidence intervals Eric B. Laber Department of Statistics, North Carolina State University April 3, 2014 I remember that one fateful day when Coach took
More informationTUTORIAL 8 SOLUTIONS #
TUTORIAL 8 SOLUTIONS #9.11.21 Suppose that a single observation X is taken from a uniform density on [0,θ], and consider testing H 0 : θ = 1 versus H 1 : θ =2. (a) Find a test that has significance level
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More informationFrequentist-Bayesian Model Comparisons: A Simple Example
Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal
More informationCSC321 Lecture 18: Learning Probabilistic Models
CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling
More information3 Continuous Random Variables
Jinguo Lian Math437 Notes January 15, 016 3 Continuous Random Variables Remember that discrete random variables can take only a countable number of possible values. On the other hand, a continuous random
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationhttp://www.math.uah.edu/stat/hypothesis/.xhtml 1 of 5 7/29/2009 3:14 PM Virtual Laboratories > 9. Hy pothesis Testing > 1 2 3 4 5 6 7 1. The Basic Statistical Model As usual, our starting point is a random
More informationProblem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30
Problem Set MAS 6J/1.16J: Pattern Recognition and Analysis Due: 5:00 p.m. on September 30 [Note: All instructions to plot data or write a program should be carried out using Matlab. In order to maintain
More informationCharles Geyer University of Minnesota. joint work with. Glen Meeden University of Minnesota.
Fuzzy Confidence Intervals and P -values Charles Geyer University of Minnesota joint work with Glen Meeden University of Minnesota http://www.stat.umn.edu/geyer/fuzz 1 Ordinary Confidence Intervals OK
More informationDerivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques
Vol:7, No:0, 203 Derivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques D. A. Farinde International Science Index, Mathematical and Computational Sciences Vol:7,
More information8 Basics of Hypothesis Testing
8 Basics of Hypothesis Testing 4 Problems Problem : The stochastic signal S is either 0 or E with equal probability, for a known value E > 0. Consider an observation X = x of the stochastic variable X
More informationThis does not cover everything on the final. Look at the posted practice problems for other topics.
Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry
More informationDetection Theory. Composite tests
Composite tests Chapter 5: Correction Thu I claimed that the above, which is the most general case, was captured by the below Thu Chapter 5: Correction Thu I claimed that the above, which is the most general
More informationStatistical Distribution Assumptions of General Linear Models
Statistical Distribution Assumptions of General Linear Models Applied Multilevel Models for Cross Sectional Data Lecture 4 ICPSR Summer Workshop University of Colorado Boulder Lecture 4: Statistical Distributions
More information