A REVERSE TO THE JEFFREYS LINDLEY PARADOX

Similar documents
7. Estimation and hypothesis testing. Objective. Recommended reading

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

A Note on Hypothesis Testing with Random Sample Sizes and its Relationship to Bayes Factors

1 Hypothesis Testing and Model Selection

P Values and Nuisance Parameters

7. Estimation and hypothesis testing. Objective. Recommended reading

Invariant HPD credible sets and MAP estimators

Part III. A Decision-Theoretic Approach and Bayesian testing

Second Workshop, Third Summary

Parametric Techniques Lecture 3

Integrated Objective Bayesian Estimation and Hypothesis Testing

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Bayesian Social Learning with Random Decision Making in Sequential Systems

Statistical Inference

On the Bayesianity of Pereira-Stern tests

Package severity. February 20, 2015

H 2 : otherwise. that is simply the proportion of the sample points below level x. For any fixed point x the law of large numbers gives that

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence

David Giles Bayesian Econometrics

Hypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33

Understanding Ding s Apparent Paradox

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Bayesian Assessment of Hypotheses and Models

Testing Simple Hypotheses R.L. Wolpert Institute of Statistics and Decision Sciences Duke University, Box Durham, NC 27708, USA

A CONDITION TO OBTAIN THE SAME DECISION IN THE HOMOGENEITY TEST- ING PROBLEM FROM THE FREQUENTIST AND BAYESIAN POINT OF VIEW

Ch. 5 Hypothesis Testing

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Parametric Techniques

Decision theory. 1 We may also consider randomized decision rules, where δ maps observed data D to a probability distribution over

Data Mining Chapter 4: Data Analysis and Uncertainty Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Sequential Monitoring of Clinical Trials Session 4 - Bayesian Evaluation of Group Sequential Designs

Bayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida

Measuring Statistical Evidence Using Relative Belief

Comparison of Three Calculation Methods for a Bayesian Inference of Two Poisson Parameters

Including historical data in the analysis of clinical trials using the modified power priors: theoretical overview and sampling algorithms

Hypothesis Testing. Part I. James J. Heckman University of Chicago. Econ 312 This draft, April 20, 2006

Testing Restrictions and Comparing Models

Bayesian vs frequentist techniques for the analysis of binary outcome data

Hierarchical Bayesian Inversion

Bayesian inference. Justin Chumbley ETH and UZH. (Thanks to Jean Denizeau for slides)

HYPOTHESIS TESTING. Hypothesis Testing

Bayesian hypothesis testing

MODEL COMPARISON CHRISTOPHER A. SIMS PRINCETON UNIVERSITY

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Statistical Methods in Particle Physics. Lecture 2

Uniformly and Restricted Most Powerful Bayesian Tests

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests

Lecture notes on statistical decision theory Econ 2110, fall 2013

Remarks on Improper Ignorance Priors

Bayesian Hypothesis Testing: Redux

Bios 6649: Clinical Trials - Statistical Design and Monitoring

Objective Bayesian Hypothesis Testing

Review. December 4 th, Review

Bayesian Computation

Summary of Chapters 7-9

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Stable Limit Laws for Marginal Probabilities from MCMC Streams: Acceleration of Convergence

Group Sequential Designs: Theory, Computation and Optimisation

Bayesian tests of hypotheses

Lecture 2: Basic Concepts of Statistical Decision Theory

Topic 10: Hypothesis Testing

PROBABILITY AND STATISTICS Vol. I - Random Variables and Their Distributions - Wiebe R. Pestman RANDOM VARIABLES AND THEIR DISTRIBUTIONS

Answers and expectations

Statistical Inference

Seminar über Statistik FS2008: Model Selection

Bayesian Regression (1/31/13)

STATISTICS SYLLABUS UNIT I

Chapter 2 Signal Processing at Receivers: Detection Theory

Frequentist Statistics and Hypothesis Testing Spring

Relationship between Least Squares Approximation and Maximum Likelihood Hypotheses

Chapters 10. Hypothesis Testing

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics

ROBUST TESTS BASED ON MINIMUM DENSITY POWER DIVERGENCE ESTIMATORS AND SADDLEPOINT APPROXIMATIONS

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Statistical Methods for Astronomy

Noninformative Priors for the Ratio of the Scale Parameters in the Inverted Exponential Distributions

Statistical Methods in Particle Physics

KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS

LECTURE NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO

Eco517 Fall 2004 C. Sims MIDTERM EXAM

HYPOTHESIS TESTING: FREQUENTIST APPROACH.

Fiducial Inference and Generalizations

Political Science 236 Hypothesis Testing: Review and Bootstrapping

Statistical Methods for Particle Physics (I)

2. A Basic Statistical Toolbox

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

(1) Introduction to Bayesian statistics

IMPROVING TWO RESULTS IN MULTIPLE TESTING

An invariance result for Hammersley s process with sources and sinks

Summary: the confidence interval for the mean (σ 2 known) with gaussian assumption

Topic 10: Hypothesis Testing

Testing Hypotheses in Particle Physics: Plots of p 0 Versus p 1

Constructing Ensembles of Pseudo-Experiments

ON TWO RESULTS IN MULTIPLE TESTING

A BAYESIAN MATHEMATICAL STATISTICS PRIMER. José M. Bernardo Universitat de València, Spain

Model comparison. Christopher A. Sims Princeton University October 18, 2016

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

QED. Queen s Economics Department Working Paper No Hypothesis Testing for Arbitrary Bounds. Jeffrey Penney Queen s University

Transcription:

PROBABILITY AND MATHEMATICAL STATISTICS Vol. 38, Fasc. 1 (2018), pp. 243 247 doi:10.19195/0208-4147.38.1.13 A REVERSE TO THE JEFFREYS LINDLEY PARADOX BY WIEBE R. P E S T M A N (LEUVEN), FRANCIS T U E R L I N C K X (LEUVEN), AND WOLF VA N PA E M E L (LEUVEN) Abstract. In this paper the seminal Jeffreys Lindley paradox is regarded from a mathematical point of view. We show that in certain scenarios the paradox may emerge in a reverse direction. 2010 AMS Mathematics Subject Classification: Primary: 62A01, 62C10; Secondary: 60A10. Key words and phrases: Hypothesis testing, Bayes factor, p-value, reverse, paradox, Lindley, Jeffreys. 1. INTRODUCTION The Jeffreys Lindley paradox (see [2], [6], [9]) describes a discordance between frequentist and Bayesian hypothesis testing. When comparing a simple null hypothesis against a diffuse alternative, it has been found that a given sample may simultaneously lead to a frequentist rejection of the null hypothesis (because the p-value is smaller than a critical alpha) and a Bayesian support for the null hypothesis (because the value of the Bayes factor exceeds some critical threshold). The paradox has been the subject of intensive debate in the statistical literature and this debate is still ongoing ([6], [10], [11]). The classical example used to illustrate the discordance involves Gaussian populations with known variance and a point null hypothesis for the mean versus a diffuse alternative hypothesis. In words, the argument usually goes as follows: A value for the test statistic is chosen that gives a small but constant p-value with increasing sample size (hence the test value scales with the sample size), thus leading to a systematic rejection of the null hypothesis in frequentist statistics. Under the same scenario, it can then be shown that, with increasing sample sizes, the Bayes factor of the alternative hypothesis over the null goes to zero. Thus, asymptotically, the Bayesian will favour the null hypothesis. The same phenomenon can be The research leading to the results reported in this paper was supported in part by the Research Fund of the KU Leuven (OT/11/032 and CREA/11/005).

244 W. R. Pestman et al. observed with several non-gaussian populations. In this paper we show, however, that there also exist scenarios in which the Jeffreys Lindley paradox may appear in a reversed way. That is to say, we show that it may occur that the frequentist asymptotically maintains the null hypothesis whereas the Bayesian rejects it. 2. THE JEFFREYS LINDLEY PARADOX REVERSED We will construct an example where the Jeffreys Lindley paradox appears in a reversed way. As a first step, we define an auxiliary function f : R [0, + ) as follows: { } A (2.1) f(x) = x exp log(1/ x ) log ( log(1/ x ) ) if 0 < x < 1 e 2, 0 elsewhere. The constant A in the above formula is chosen in such a way that the function f integrates to one. Thus f can be considered a probability density in the strict sense of the word. This density has the following properties: f is symmetric around the origin. f is piecewise continuous. f has compact support and thus has moments of all orders. Due to a singularity at x = 0, all convolution powers f n are unbounded. A proof of the last property can be found in [4]. In the following, the variance of f will be denoted by σ 2. A family {f( θ) θ R} of densities is defined as (2.2) f(x θ) = f(x θ). Suppose now that a population is given with a probability density f( θ), where the parameter θ, presenting the population mean, is unknown. In the following two subsections we will compare the Bayesian and frequentist approach when testing the null hypothesis H 0 : θ = 0 against the alternative hypothesis H 1 : θ 0. To this end, in both scenarios, a sample X 1, X 2,..., X n is drawn from the population, and the sample mean Y n, defined as (2.3) Y n = X 1 + X 2 +... + X n, n is chosen to be the test statistic. 2.1. The Bayesian approach. In a Bayesian framework of hypothesis testing we need to specify priors for the null and the alternative hypothesis. In our example the null prior is chosen to be the Dirac measure δ 0 in θ = 0. In order to reverse the Jeffreys Lindley paradox, the alternative prior p is constructed as follows. First we define a sequence I 1, I 2, I 3,... of subsets of R by I n = { θ R f n (σ n nθ) n2 n f n (σ n) }.

A reverse to the Jeffreys Lindley paradox 245 Now, on the one hand, the functions f n being unbounded and piecewise continuous, the subsets I n must have non-empty interior, and therefore must be of strictly positive Lebesgue measure. On the other hand, the f n being integrable, the subsets I n must be of bounded Lebesgue measure. Altogether one may talk about the uniform distribution on I n and about its density p n. In terms of these p n, we define the prior p as (2.4) p(θ) = k=1 2 k p k (θ). From Lebesgue s convergence theorems (see, for example, [3], [7], [8]) it follows that p is a well-defined density with total probability mass equal to one. With the likelihood given by (2.2) the hypotheses to be tested are H 0 : prior is δ 0 against H 1 : prior is p. Bayesians will base their decision on the Bayes factor, which is in this scenario, for an arbitrary outcome y of Y n, given by (2.5) BF n (y) = + f Yn (y θ) p(θ) dθ f Yn (y 0) In the above formula, f Yn ( θ) stands for the probability density of the variable Y n, given θ. If this Bayes factor exceeds some prescribed threshold, then the Bayesian rejects the hypothesis H 0. Suppose now that we observe the outcome. y n = σ n for the sample mean Y n. The value of the density f Yn in y n may be expressed (see [3], [5]) in terms of a convolution power of f as Using this in (2.5), one arrives at BF n (y n ) = = + k=1 f Yn (y n θ) = nf n (σ n nθ). + f Yn (y n θ) f Yn (y n 0) p(θ) dθ = f n (σ n nθ) I k f n (σ 2 k p k (θ) dθ n) f n (σ n nθ) I n f n (σ 2 n p n (θ) dθ n) n2 n 2 n p n (θ) dθ = n. I n f n (σ n nθ) f n (σ n) p(θ) dθ

246 W. R. Pestman et al. It thus appears that (2.6) lim n BF n(y n ) =. Thus, asymptotically, the Bayesian will reject H 0 when confronted with outcomes y n of the sample mean Y n. 2.2. The frequentist approach. The frequentist approach is based on the p- value. If this p-value is below some prescribed threshold (typically 0.05), then the frequentist will reject the hypothesis H 0. The p-value is computed starting from the null-hypothesized value for θ, in our scenario the value θ = 0. Contrary to the Bayesian approach, the form of the alternative hypothesis H 1 does not play an essential role in the frequentist decision procedure. When testing in a two-sided way, the p-value in our particular scenario, given the outcome y n = σ/ n for the sample mean Y n, would be determined as (2.7) PV n = 2 {1 F Yn (y n H 0 )}, where F Yn ( H 0 ) stands for the cumulative distribution function of Y n under H 0. In order to evaluate the p-value asymptotically, we define the variable Z n as Z n = Y n σ/ n. Note that Z n is under H 0 precisely the standardization of Y n. Given the outcome y n for the sample mean Y n one may rewrite (2.7) as follows in terms of the Z n : PV n = 2 {1 F Zn (1 H 0 )}. The second moment of the population being finite, the classical central limit theorem may be applied. Thus, denoting the cumulative distribution function of the standard Gaussian distribution by Φ, the asymptotic p-value turns out to be (2.8) lim n PV n = 2 {1 Φ(1)} = 0.32. When observing a p-value of this size the frequentist will generally decide not to reject H 0. Hence, asymptotically, the frequentist will maintain H 0 when confronted with an outcome y n of the sample mean Y n. 3. CLOSING REMARKS We have constructed an example in which the Bayesian (basing his decision on (2.6)) will asymptotically reject H 0 whereas the frequentist (basing his decision on (2.8)) will asymptotically not reject this hypothesis. It thus appears that, in the absence of sufficient regularity of likelihood or prior, the Jeffreys Lindley paradox

A reverse to the Jeffreys Lindley paradox 247 may manifest itself in a reversed way. It should be noted that, rather than the specific function f defined by (2.1), any persistently unbounded probability density (see [4]) will lead to a reversion of the paradox. In the likelihood defined by (2.2), the parameter θ presents the population mean. For this reason, at first sight, it may seem natural to use its empirical counterpart, the sample mean Y n defined by (2.3), as the test statistic. However, in the proposed scenario, the sample mean fails to be a sufficient estimator for the parameter θ. Future work needs to be done to construct an example in which the paradox is reversed through a sufficient test statistic. REFERENCES [1] H. Jeffreys, The Theory of Probability, Oxford University Press, Oxford 1961. [2] D. V. Lindley, A statistical paradox, Biometrika 44 (1957), pp. 187 192. [3] W. R. Pestman, Mathematical Statistics, second edition, De Gruyter Verlag, Berlin 2009. [4] W. R. Pestman, F. Tuerlinckx, and W. Vanpaemel, Persistently unbounded probability densities, Statist. Probab. Lett. 118 (2016), pp. 135 138. [5] A. Rényi, Foundations of Probability, Dover Publications, Inc., New York 2007. [6] C. P. Robert, On the Jeffreys Lindley paradox, Philos. Sci. 81 (2014), pp. 216 232. [7] W. Rudin, Real and Complex Analysis, McGraw-Hill, New York 1987. [8] L. Schwartz, Mathematics for the Physical Sciences, Dover Publications, New York 2008. [9] G. Shafer, Lindley s paradox, J. Amer. Statist. Assoc. 77 (378) (1982), pp. 325 334. [10] A. Spanos, Who should be afraid of the Jeffreys Lindley paradox?, Philos. Sci. 80 (2013), pp. 73 93. [11] J. Sprenger, Testing a precise null hypothesis: The case of Lindley s paradox, Philos. Sci. 80 (2013), pp. 733 744. Wiebe R. Pestman University of Leuven Tiensestraat 102, 3000 Leuven, Belgium E-mail: wrpestman@gmail.com Francis Tuerlinckx E-mail: francis.tuerlinckx@kuleuven.be Wolf Vanpaemel E-mail: wolf.vanpaemel@kuleuven.be Received on 16.9.2016; revised version on 24.1.2017