Divergence Based priors for the problem of hypothesis testing

Size: px
Start display at page:

Download "Divergence Based priors for the problem of hypothesis testing"

Transcription

1 Divergence Based priors for the problem of hypothesis testing gonzalo garcía-donato and susie Bayarri May 22, 2009 gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

2 Jeffreys and testing: the essentials Much of the material presented in this talk is based on the work of Jeffreys. gonzalo garcı a-donato and susie Bayarri () DB priors May 22, / 46

3 Jeffreys and testing: the essentials Much of the material presented in this talk is based on the work of Jeffreys. Sir Harold Jeffreys born in England in 1891 and died in 1989 (almost at the age of hundred). He was Professor of Astronomy at Cambridge. gonzalo garcı a-donato and susie Bayarri () DB priors May 22, / 46

4 Jeffreys and testing: the essentials Much of the material presented in this talk is based on the work of Jeffreys. Sir Harold Jeffreys born in England in 1891 and died in 1989 (almost at the age of hundred). He was Professor of Astronomy at Cambridge. His contributions to statistics through a Bayesian thinking are very popular (eg. Jeffreys prior) and are the basis of many important methods. gonzalo garcı a-donato and susie Bayarri () DB priors May 22, / 46

5 Jeffreys and testing: the essentials We focus here on his ideas on testing gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

6 Jeffreys and testing: the essentials We focus here on his ideas on testing specially on the solutions he provided to testing problem, but also and very importantly, gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

7 Jeffreys and testing: the essentials We focus here on his ideas on testing specially on the solutions he provided to testing problem, but also and very importantly, on his particular perception of the problem. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

8 Jeffreys and testing: the essentials We focus here on his ideas on testing specially on the solutions he provided to testing problem, but also and very importantly, on his particular perception of the problem. This has motivated the design of this presentation: PART I. Looking at testing problem through Jeffreys eyes, gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

9 Jeffreys and testing: the essentials We focus here on his ideas on testing specially on the solutions he provided to testing problem, but also and very importantly, on his particular perception of the problem. This has motivated the design of this presentation: PART I. Looking at testing problem through Jeffreys eyes, PART II. Priors for testing, gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

10 Jeffreys and testing: the essentials We focus here on his ideas on testing specially on the solutions he provided to testing problem, but also and very importantly, on his particular perception of the problem. This has motivated the design of this presentation: PART I. Looking at testing problem through Jeffreys eyes, PART II. Priors for testing, PART III. Generalizations of Jeffreys ideas, gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

11 Jeffreys and testing: the essentials We focus here on his ideas on testing specially on the solutions he provided to testing problem, but also and very importantly, on his particular perception of the problem. This has motivated the design of this presentation: PART I. Looking at testing problem through Jeffreys eyes, PART II. Priors for testing, PART III. Generalizations of Jeffreys ideas, PART IV. Divergence based priors. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

12 PART I. Looking at testing problem through Jeffreys eyes PART I. Looking at testing problem through Jeffreys eyes gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

13 PART I. Looking at testing problem through Jeffreys eyes Testing a point null hypothesis Let y = (y 1, y 2,..., y n ) be a random sample from f (y θ, ν). We are interested in testing the hypotheses H 1 : θ = θ 0 (null), vs H 2 : θ θ 0 (alternative). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

14 PART I. Looking at testing problem through Jeffreys eyes Testing a point null hypothesis Let y = (y 1, y 2,..., y n ) be a random sample from f (y θ, ν). We are interested in testing the hypotheses H 1 : θ = θ 0 (null), vs H 2 : θ θ 0 (alternative). This is equivalent to selecting between the competing models M 1 : f 1 (y ν) = f (y θ 0, ν) vs M 2 : f 2 (y θ, ν ) = f (y θ, ν ). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

15 PART I. Looking at testing problem through Jeffreys eyes Testing a point null hypothesis Let y = (y 1, y 2,..., y n ) be a random sample from f (y θ, ν). We are interested in testing the hypotheses H 1 : θ = θ 0 (null), vs H 2 : θ θ 0 (alternative). This is equivalent to selecting between the competing models M 1 : f 1 (y ν) = f (y θ 0, ν) vs M 2 : f 2 (y θ, ν ) = f (y θ, ν ). Jeffreys (1961) calls θ the new parameter; gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

16 PART I. Looking at testing problem through Jeffreys eyes Testing a point null hypothesis Let y = (y 1, y 2,..., y n ) be a random sample from f (y θ, ν). We are interested in testing the hypotheses H 1 : θ = θ 0 (null), vs H 2 : θ θ 0 (alternative). This is equivalent to selecting between the competing models M 1 : f 1 (y ν) = f (y θ 0, ν) vs M 2 : f 2 (y θ, ν ) = f (y θ, ν ). Jeffreys (1961) calls θ the new parameter; ν, ν are the old (or common) parameters. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

17 PART I. Looking at testing problem through Jeffreys eyes Initial considerations Due to the statement in H 1, this problem has received, in the literature, the name of punctual, precise, sharp hypothesis testing or significance tests (Jeffreys 1961). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

18 PART I. Looking at testing problem through Jeffreys eyes Initial considerations Due to the statement in H 1, this problem has received, in the literature, the name of punctual, precise, sharp hypothesis testing or significance tests (Jeffreys 1961). A precise hypothesis is (in many situations) a convenient approximation of the more realistic hypothesis H : θ = θ 0 H : θ θ 0 < ɛ, ɛ <<. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

19 PART I. Looking at testing problem through Jeffreys eyes Initial considerations Due to the statement in H 1, this problem has received, in the literature, the name of punctual, precise, sharp hypothesis testing or significance tests (Jeffreys 1961). A precise hypothesis is (in many situations) a convenient approximation of the more realistic hypothesis H : θ = θ 0 H : θ θ 0 < ɛ, ɛ <<. See Berger and Delampady (1987) for a quantification of this correspondence. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

20 PART I. Looking at testing problem through Jeffreys eyes Initial considerations Due to the statement in H 1, this problem has received, in the literature, the name of punctual, precise, sharp hypothesis testing or significance tests (Jeffreys 1961). A precise hypothesis is (in many situations) a convenient approximation of the more realistic hypothesis H : θ = θ 0 H : θ θ 0 < ɛ, ɛ <<. See Berger and Delampady (1987) for a quantification of this correspondence. Common parameters ν in M 1 and ν in M 2 do not (necessarily) identify the same magnitude. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

21 PART I. Looking at testing problem through Jeffreys eyes Testing precise hypothesis is not an estimation problem gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

22 PART I. Looking at testing problem through Jeffreys eyes Testing precise hypothesis is not an estimation problem Many authors, including Jeffreys, have repeatedly argued that estimation and testing problem are statistical problems of different nature. Jeffreys, in his book Theory of probability right at the beginning of the chapter devoted to testing problems says gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

23 PART I. Looking at testing problem through Jeffreys eyes Testing precise hypothesis is not an estimation problem Many authors, including Jeffreys, have repeatedly argued that estimation and testing problem are statistical problems of different nature. Jeffreys, in his book Theory of probability right at the beginning of the chapter devoted to testing problems says In the last chapters we were concerned with the estimation of the parameters in a law, the form of the law itself being given. We are now concerned with the more difficult question: in what circumstances do observations support a change of the form of the law itself? gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

24 PART I. Looking at testing problem through Jeffreys eyes (genuine) Testing Testing (genuine) problems The proposal θ = θ 0, is (a priori) plausible and represents, in a broad sense, a theory that needs to be tested. As opposed to gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

25 PART I. Looking at testing problem through Jeffreys eyes (genuine) Testing versus estimation Testing (genuine) problems The proposal θ = θ 0, is (a priori) plausible and represents, in a broad sense, a theory that needs to be tested. As opposed to Estimation problems There is no uncertainty about what is the right model and there is no a special value of θ. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

26 PART I. Looking at testing problem through Jeffreys eyes Different problems According to these ideas, it is not that surprising that the tools used for estimation problems do not work for testing problems: gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

27 PART I. Looking at testing problem through Jeffreys eyes Different problems According to these ideas, it is not that surprising that the tools used for estimation problems do not work for testing problems: Improper priors produce arbitrary answers, gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

28 PART I. Looking at testing problem through Jeffreys eyes Different problems According to these ideas, it is not that surprising that the tools used for estimation problems do not work for testing problems: Improper priors produce arbitrary answers, same (or worst) about proper flat priors, Gamma( , ) gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

29 PART I. Looking at testing problem through Jeffreys eyes Different problems According to these ideas, it is not that surprising that the tools used for estimation problems do not work for testing problems: Improper priors produce arbitrary answers, same (or worst) about proper flat priors, Gamma( , ) Credible intervals to solve hypothesis testing [Example]. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

30 PART I. Looking at testing problem through Jeffreys eyes Different problems According to these ideas, it is not that surprising that the tools used for estimation problems do not work for testing problems: Improper priors produce arbitrary answers, same (or worst) about proper flat priors, Gamma( , ) Credible intervals to solve hypothesis testing [Example]. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

31 PART I. Looking at testing problem through Jeffreys eyes Jeffreys tools Jeffreys based the solution of the problem on what we call now Bayes factors: M 1 : f 1 (y ν) vs M 2 : f 2 (y θ, ν ), B 12 = m 1(y) m 2 (y), where, the prior marginals are m 1 (y) = f 1 (y ν)π 1 (ν)dν, m 2 (y) = f 2 (y θ, ν )π 2 (θ, ν )dν dθ. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

32 PART I. Looking at testing problem through Jeffreys eyes Other approaches There are many other Bayesian measures used in the problem of hypothesis testing, like the BIC, DIC, etc. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

33 PART I. Looking at testing problem through Jeffreys eyes Other approaches There are many other Bayesian measures used in the problem of hypothesis testing, like the BIC, DIC, etc. Principle (Berger y Pericchi 2001) Testing and model selection methods should correspond, in some sense, to actual Bayes factors, arising from reasonable default prior distributions. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

34 PART I. Looking at testing problem through Jeffreys eyes Other approaches There are many other Bayesian measures used in the problem of hypothesis testing, like the BIC, DIC, etc. Principle (Berger y Pericchi 2001) Testing and model selection methods should correspond, in some sense, to actual Bayes factors, arising from reasonable default prior distributions.? The question of what priors should be used is an open question, of which only partial answers are known. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

35 PART II. Objective testing priors PART II Objective testing priors gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

36 PART II. Objective testing priors The normal scenario Y i (iid)n(µ, σ 2 ), i = 1, 2,..., n, σ unknown. H 1 : µ = 0 (M 1 : N(0, σ 2 )), H 2 : µ 0 (M 2 : N(µ, τ 2 )). B 12 = p1 (y σ)π 1 (σ)dσ p2 (y µ, τ)π 2.1 (τ)π(µ τ)dτdµ. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

37 PART II. Objective testing priors Jeffreys proposal π 1 (σ) = h(σ), π 2 (µ, τ) = h(τ)π(µ τ). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

38 PART II. Objective testing priors Jeffreys proposal π 1 (σ) = h(σ), π 2 (µ, τ) = h(τ)π(µ τ). Arguing, with respect to the common parameters µ and τ are orthogonal so τ and σ have the same meaning, justifying the assignment of the same prior, h in this case. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

39 PART II. Objective testing priors Jeffreys proposal π 1 (σ) = cσ 1, π 2 (µ, τ) = cτ 1 π(µ τ). Arguing, with respect to the common parameters µ and τ are orthogonal so τ and σ have the same meaning, justifying the assignment of the same prior, h in this case. Bayes factor is quite robust to the election of h, and hence we can use h(σ) = σ 1 (constants cancel out). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

40 PART II. Objective testing priors Jeffreys proposal π 1 (σ) = cσ 1, π 2 (µ, τ) = cτ 1 π(µ τ). With respect to the new parameter: π(µ τ) has to be proper,centered at zero (the null model),scaled by τ, symmetric around zero, gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

41 PART II. Objective testing priors Jeffreys proposal π 1 (σ) = cσ 1, π 2 (µ, τ) = cτ 1 π(µ τ). With respect to the new parameter: π(µ τ) has to be proper,centered at zero (the null model),scaled by τ, symmetric around zero, with heavy tails. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

42 PART II. Objective testing priors Jeffreys proposal π 1 (σ) = cσ 1, π 2 (µ, τ) = cτ 1 Ca(µ 0, τ). With respect to the new parameter: π(µ τ) has to be proper,centered at zero (the null model),scaled by τ, symmetric around zero, with heavy tails. The simpler function with these properties is the Ca(0, τ). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

43 PART II. Objective testing priors Jeffreys proposal have resisted the passage of time. It has been used as a benchmark for other priors. Why? It provides reasonable responses. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

44 PART II. Objective testing priors Jeffreys proposal have resisted the passage of time. It has been used as a benchmark for other priors. Why? It provides reasonable responses. It is information consistent: as the information against the null becomes overwhelming the Bayes factor against the null tends to infinity. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

45 PART II. Objective testing priors Jeffreys proposal have resisted the passage of time. It has been used as a benchmark for other priors. Why? It provides reasonable responses. It is information consistent: as the information against the null becomes overwhelming the Bayes factor against the null tends to infinity. Surprisingly the Ca(0, τ) has arisen, approximately, using other modern methodologies: the intrinsic priors (Berger and Pericchi, 1996) and the expected posterior priors (Pérez and Berger, 2002). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

46 PART II. Objective testing priors Diving a little bit deeper: a key slide gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

47 PART II. Objective testing priors Diving a little bit deeper: a key slide Implicitly recognizes the null model as a source of objective information: M 0 gives the center and the scale. This should be interpreted as of unitary size, gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

48 PART II. Objective testing priors Diving a little bit deeper: a key slide Implicitly recognizes the null model as a source of objective information: M 0 gives the center and the scale. This should be interpreted as of unitary size,...the mere fact that it it has been suggested that the parameter is zero corresponds to some presumption that it is fairly small. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

49 PART II. Objective testing priors Diving a little bit deeper: a key slide Implicitly recognizes the null model as a source of objective information: M 0 gives the center and the scale. This should be interpreted as of unitary size,...the mere fact that it it has been suggested that the parameter is zero corresponds to some presumption that it is fairly small. The importance of the tails (information consistency). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

50 PART III. Generalizations of Jeffreys ideas PART III Generalizations of Jeffreys ideas gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

51 PART III. Generalizations of Jeffreys ideas Jeffreys proposals are very specific to the problem at hand (a normal mean). They do not arise through a rule which can be reasonably applied to other testing scenarios. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

52 PART III. Generalizations of Jeffreys ideas Jeffreys proposals are very specific to the problem at hand (a normal mean). They do not arise through a rule which can be reasonably applied to other testing scenarios. Is it always a Cauchy a sensible prior? gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

53 PART III. Generalizations of Jeffreys ideas Jeffreys proposals are very specific to the problem at hand (a normal mean). They do not arise through a rule which can be reasonably applied to other testing scenarios. Is it always a Cauchy a sensible prior? What is a good scale? gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

54 PART III. Generalizations of Jeffreys ideas Jeffreys proposals are very specific to the problem at hand (a normal mean). They do not arise through a rule which can be reasonably applied to other testing scenarios. Is it always a Cauchy a sensible prior? What is a good scale? What about problems with multivariate parameters or problems with some design? gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

55 PART III. Generalizations of Jeffreys ideas Jeffreys proposals are very specific to the problem at hand (a normal mean). They do not arise through a rule which can be reasonably applied to other testing scenarios. Is it always a Cauchy a sensible prior? What is a good scale? What about problems with multivariate parameters or problems with some design? Although not too many, there are some interesting ideas in the literature about the extension of Jeffreys ideas. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

56 PART III. Generalizations of Jeffreys ideas Jeffreys proposals are very specific to the problem at hand (a normal mean). They do not arise through a rule which can be reasonably applied to other testing scenarios. Is it always a Cauchy a sensible prior? What is a good scale? What about problems with multivariate parameters or problems with some design? Although not too many, there are some interesting ideas in the literature about the extension of Jeffreys ideas. We start focusing on priors for new parameters. These ideas are the basis for the full problem with nuisance parameters (more on this at the end) gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

57 PART III. Generalizations of Jeffreys ideas Idea 1: Jeffreys extending Jeffreys gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

58 PART III. Generalizations of Jeffreys ideas Idea 1: Jeffreys extending Jeffreys An omnipresent quantity in Jeffreys work is the divergence: D = log f 1 d ( ) f 1 f 2. f 2 gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

59 PART III. Generalizations of Jeffreys ideas Idea 1: Jeffreys extending Jeffreys An omnipresent quantity in Jeffreys work is the divergence: D = log f 1 d ( ) f 1 f 2. f 2 For the general problem: H 1 : y iid f (y θ 0 ),, H 2 : y iid f (y θ), θ R, Jeffreys proposed the rule gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

60 PART III. Generalizations of Jeffreys ideas Idea 1: Jeffreys extending Jeffreys An omnipresent quantity in Jeffreys work is the divergence: D = log f 1 d ( ) f 1 f 2. f 2 For the general problem: H 1 : y iid f (y θ 0 ),, H 2 : y iid f (y θ), θ R, Jeffreys proposed the rule π J (θ) = 1 π d dθ tan 1 (D[θ, θ 0 ]) 1/2, where D[θ, θ 0 ] is the divergence between f (y θ 0 ) and f (y θ). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

61 PART III. Generalizations of Jeffreys ideas Jeffreys extending Jeffreys (cont ) π J (θ) reduces to Jeffreys Cauchy proposal when θ is a normal mean. Noticeable, when θ θ 0 is small, π J (θ) 1 π ( 1 + D[θ, θ0 ] ) 1 π NJ (θ), where π NJ (θ) is Jeffreys (estimation) prior (i.e. the squared root of the expected Fisher information). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

62 PART III. Generalizations of Jeffreys ideas Jeffreys extending Jeffreys (cont ) π J (θ) reduces to Jeffreys Cauchy proposal when θ is a normal mean. Noticeable, when θ θ 0 is small, π J (θ) 1 π ( 1 + D[θ, θ0 ] ) 1 π NJ (θ), where π NJ (θ) is Jeffreys (estimation) prior (i.e. the squared root of the expected Fisher information). Unfortunately Note that π J can lead to improper priors and at least in principle can not be applied for multivariate parameters. However, the approximation was a main inspiration for the definition of DB priors, with clear similarities between them. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

63 PART III. Generalizations of Jeffreys ideas Idea 2: Zellner and Siow (1980, 1984) For the linear model y N(αX 1 + βx, σ 2 I), X t 1X = 0, with hypotheses they propose H 1 : β = 0, H 2 : β 0, β Ca k (0, nσ 2 (X t X) 1 ), the matrix (X t X) 1 suggested by the information matrix. This is a proposal followed by many others: Liang et al (2008), Bayarri and García-Donato (2007). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

64 PART III. Generalizations of Jeffreys ideas Idea 3: Kass and Wasserman (1995) Along these same lines, for a general multidimensional parameter β, Kass and Wasserman (1995) propose β Ca k (0, I β (α, β = 0) 1 ), where I β is the block of the Fisher information matrix corresponding to β (one observation). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

65 PART III. Generalizations of Jeffreys ideas Notice These generalizations focus on the Fisher information, a second order approximation of the Kullback-Leibler divergence. Jeffreys did a similar exercise. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

66 PART III. Generalizations of Jeffreys ideas Notice These generalizations focus on the Fisher information, a second order approximation of the Kullback-Leibler divergence. Jeffreys did a similar exercise. They are primarily concerned with location-type parameters. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

67 PART III. Generalizations of Jeffreys ideas Notice These generalizations focus on the Fisher information, a second order approximation of the Kullback-Leibler divergence. Jeffreys did a similar exercise. They are primarily concerned with location-type parameters. Our work retakes the essence of the Jeffreys ideas using directly divergence measures as a central quantity. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

68 PART IV. Divergence based priors PART IV Divergence based priors gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

69 PART IV. Divergence based priors Motivation: θ is a location parameter (in R) H 1 : θ = θ 0 vs. H 2 : θ θ 0 gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

70 PART IV. Divergence based priors Motivation: θ is a location parameter (in R) H 1 : θ = θ 0 vs. H 2 : θ θ 0 M 1 : f 1 (y) = f (y θ 0 ) vs. M 2 : f 2 (y θ) = f (y θ), where f is a given model. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

71 PART IV. Divergence based priors Motivation: θ is a location parameter (in R) H 1 : θ = θ 0 vs. H 2 : θ θ 0 M 1 : f 1 (y) = f (y θ 0 ) vs. M 2 : f 2 (y θ) = f (y θ), where f is a given model. Let D[θ, θ 0 ] be a measure of unitary distance between M 1 and M 2. We prefer use the Kullback-Leibler divergence divided by n. n D[θ, θ 0 ] = log f (y θ 0) ( f (y θ) f (y θ0 ) ) dy. f (y θ) gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

72 PART IV. Divergence based priors Motivation: θ is a location parameter (in R) H 1 : θ = θ 0 vs. H 2 : θ θ 0 M 1 : f 1 (y) = f (y θ 0 ) vs. M 2 : f 2 (y θ) = f (y θ), where f is a given model. Let D[θ, θ 0 ] be a measure of unitary distance between M 1 and M 2. We prefer use the Kullback-Leibler divergence divided by n. n D[θ, θ 0 ] = log f (y θ 0) ( f (y θ) f (y θ0 ) ) dy. f (y θ) Let h q ( ) be a real valued decreasing function, possibly indexed by a real argument q > 0. Simple possibilities: h(x) = e q x, h q (x) = (1 + x) q, h q (x) = (1 + x/q) q. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

73 PART IV. Divergence based priors Definition gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

74 PART IV. Divergence based priors Definition Define (θ location) being q such that π D is proper. π D (θ) h q (D[θ, θ 0 ]), gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

75 PART IV. Divergence based priors Definition Define (θ location) being q such that π D is proper. π D (θ) h q (D[θ, θ 0 ]), Interpretation π D is a robust prior which uses M 1 as a guess for M 2 : those values of θ such that f 1 (y) and f 2 (y θ) are close, have the greatest probability. This idea goes a step beyond other proposals, under which the prior is (an inverse) function of (θ θ 0 ) 2 or I (θ 0 ) (θ θ 0 ) 2. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

76 PART IV. Divergence based priors Basic property π D is a unimodal density, with mode at θ 0, symmetric around θ 0. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

77 PART IV. Divergence based priors Testing a normal mean H 1 : µ = 0, revisited (σ known) We have D[µ, 0] = µ 2 /σ 2, and hence π D (µ) = Γ(q) Γ(q.5)σ ( µ 2 ) q, 1 + q > 0.5 π σ 2 gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

78 PART IV. Divergence based priors Testing a normal mean H 1 : µ = 0, revisited (σ known) We have D[µ, 0] = µ 2 /σ 2, and hence π D (µ) = Γ(q) Γ(q.5)σ ( µ 2 ) q, 1 + q > 0.5 π σ 2 gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

79 PART IV. Divergence based priors Testing a normal mean H 1 : µ = 0, revisited (σ known) We have D[µ, 0] = µ 2 /σ 2, and hence π D (µ) = Γ(q) Γ(q.5)σ ( µ 2 ) q, 1 + q > 0.5 π σ 2 gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

80 PART IV. Divergence based priors Testing a normal mean H 1 : µ = 0, revisited (σ known) We have D[µ, 0] = µ 2 /σ 2, and hence π D (µ) = Γ(q) Γ(q.5)σ ( µ 2 ) q, 1 + q > 0.5 π σ 2 gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

81 PART IV. Divergence based priors Testing a normal mean H 1 : µ = 0, revisited (σ known) We have D[µ, 0] = µ 2 /σ 2, and hence π D (µ) = Γ(q) Γ(q.5)σ ( µ 2 ) q, 1 + q > 0.5 π σ 2 gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

82 PART IV. Divergence based priors Scale parameters: introducing the general rule The above proposal can be easily applied to scale parameters, σ: M 1 : f 1 (y) = f (y σ 0 ) vs. M 2 : f 2 (y σ) = f (y σ). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

83 PART IV. Divergence based priors Scale parameters: introducing the general rule The above proposal can be easily applied to scale parameters, σ: M 1 : f 1 (y) = f (y σ 0 ) vs. M 2 : f 2 (y σ) = f (y σ). Reparameterize in terms of θ = log σ, a location parameter, derive the DB prior π D (θ) h q (D [θ, log σ 0 ]), and transform it to the original parameterization. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

84 PART IV. Divergence based priors Scale parameters: introducing the general rule The above proposal can be easily applied to scale parameters, σ: M 1 : f 1 (y) = f (y σ 0 ) vs. M 2 : f 2 (y σ) = f (y σ). Reparameterize in terms of θ = log σ, a location parameter, derive the DB prior π D (θ) h q (D [θ, log σ 0 ]), and transform it to the original parameterization.one obtains π D (σ) h q (D [log σ, log σ 0 ]) 1 σ. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

85 PART IV. Divergence based priors Scale parameters: introducing the general rule The above proposal can be easily applied to scale parameters, σ: M 1 : f 1 (y) = f (y σ 0 ) vs. M 2 : f 2 (y σ) = f (y σ). Reparameterize in terms of θ = log σ, a location parameter, derive the DB prior π D (θ) h q (D [θ, log σ 0 ]), and transform it to the original parameterization.one obtains Note, if D is invariant π D (σ) h q (D [log σ, log σ 0 ]) 1 σ. π D (σ) h q (D[σ, σ 0 ])π N (σ). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

86 PART IV. Divergence based priors The general case (without nuisance par.): construction gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

87 PART IV. Divergence based priors The general case (without nuisance par.): construction M 1 : f 1 (y) = f (y θ 0 ) vs. M 2 : f 2 (y θ) = f (y θ), with θ a k-dimensional parameter. Let π N (θ) the reference prior. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

88 PART IV. Divergence based priors The general case (without nuisance par.): construction M 1 : f 1 (y) = f (y θ 0 ) vs. M 2 : f 2 (y θ) = f (y θ), with θ a k-dimensional parameter. Let π N (θ) the reference prior. Parameterize in terms of ξ = g(θ) such that π N (ξ) = 1. Remarkably ξ behaves, asymptotically, as location parameter (Bernardo 2005). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

89 PART IV. Divergence based priors The general case (without nuisance par.): construction M 1 : f 1 (y) = f (y θ 0 ) vs. M 2 : f 2 (y θ) = f (y θ), with θ a k-dimensional parameter. Let π N (θ) the reference prior. Parameterize in terms of ξ = g(θ) such that π N (ξ) = 1. Remarkably ξ behaves, asymptotically, as location parameter (Bernardo 2005). Apply the definition of DB priors for location parameters. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

90 PART IV. Divergence based priors The general case (without nuisance par.): construction M 1 : f 1 (y) = f (y θ 0 ) vs. M 2 : f 2 (y θ) = f (y θ), with θ a k-dimensional parameter. Let π N (θ) the reference prior. Parameterize in terms of ξ = g(θ) such that π N (ξ) = 1. Remarkably ξ behaves, asymptotically, as location parameter (Bernardo 2005). Apply the definition of DB priors for location parameters. Obtain the prior in the original parameterization (apply the transformation of variables formula). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

91 PART IV. Divergence based priors The general case (without nuisance par.): construction M 1 : f 1 (y) = f (y θ 0 ) vs. M 2 : f 2 (y θ) = f (y θ), with θ a k-dimensional parameter. Let π N (θ) the reference prior. Parameterize in terms of ξ = g(θ) such that π N (ξ) = 1. Remarkably ξ behaves, asymptotically, as location parameter (Bernardo 2005). Apply the definition of DB priors for location parameters. Obtain the prior in the original parameterization (apply the transformation of variables formula). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

92 PART IV. Divergence based priors General definition gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

93 PART IV. Divergence based priors General definition Of course, all these steps are not at all needed. By construction, we have arrived at the general definition: DB prior π D (θ) h q (D[θ, θ 0 ])π N (θ). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

94 PART IV. Divergence based priors Properties: The DB priors in words They capture the information contained into the problem, making use of the distance between the models. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

95 PART IV. Divergence based priors Properties: The DB priors in words They capture the information contained into the problem, making use of the distance between the models. The information contained is equivalent to the information in a sample of unitary size. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

96 PART IV. Divergence based priors Properties: The DB priors in words They capture the information contained into the problem, making use of the distance between the models. The information contained is equivalent to the information in a sample of unitary size. DB priors generalize the work of Jeffreys (1961) and Zellner and Siow (1980). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

97 PART IV. Divergence based priors Properties: More theoretically If θ is a location parameter then π D approximately behaves as a Student centered at θ 0 and scaled by the Fisher information matrix evaluated at θ 0. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

98 PART IV. Divergence based priors Properties: More theoretically If θ is a location parameter then π D approximately behaves as a Student centered at θ 0 and scaled by the Fisher information matrix evaluated at θ 0. π D is invariant under reparameterizations. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

99 PART IV. Divergence based priors Properties: More theoretically If θ is a location parameter then π D approximately behaves as a Student centered at θ 0 and scaled by the Fisher information matrix evaluated at θ 0. π D is invariant under reparameterizations. π D satisfies the sufficiency principle (e.g. do not change if the problem is reduced via sufficient statistics). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

100 PART IV. Divergence based priors Properties: More theoretically If θ is a location parameter then π D approximately behaves as a Student centered at θ 0 and scaled by the Fisher information matrix evaluated at θ 0. π D is invariant under reparameterizations. π D satisfies the sufficiency principle (e.g. do not change if the problem is reduced via sufficient statistics). For the example analyzed it is consistent in information gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

101 PART IV. Divergence based priors Properties: More theoretically If θ is a location parameter then π D approximately behaves as a Student centered at θ 0 and scaled by the Fisher information matrix evaluated at θ 0. π D is invariant under reparameterizations. π D satisfies the sufficiency principle (e.g. do not change if the problem is reduced via sufficient statistics). For the example analyzed it is consistent in information The (DB) Bayes factor can be expressed as B D 21 = B N 21 Correction factor (FC), where FC is a expectation with respect to the posterior distribution. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

102 PART IV. Divergence based priors The choice of q Clearly, π D depends on the election of q, a parameter which needs to be assigned. This parameter regulates the heaviness of the tails. Our proposal is q = q + 0.5, with q = inf{q > 0 : π D (θ) is proper. This reproduces the Jeffreys-Zellner-Siow priors in the normal scenario. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

103 PART IV. Divergence based priors DB priors in practice Leaving aside formal criteria... gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

104 PART IV. Divergence based priors DB priors in practice Leaving aside formal criteria... the best way of knowing if a proposal is reasonable is studying its behaviour in scenarios, representative of the practice of statistics. Scenario Standard type Improper likelihood Irregular models gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

105 PART IV. Divergence based priors DB priors in practice Leaving aside formal criteria... the best way of knowing if a proposal is reasonable is studying its behaviour in scenarios, representative of the practice of statistics. Scenario Standard type Improper likelihood Irregular models gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

106 PART IV. Divergence based priors The problem with nuisance parameters: some few words M 1 : f 1 (y ν 1 ) = f (y θ 0, ν) vs. M 2 : f 2 (y θ, ν ) = f (y θ, ν ), gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

107 PART IV. Divergence based priors The problem with nuisance parameters: some few words M 1 : f 1 (y ν 1 ) = f (y θ 0, ν) vs. M 2 : f 2 (y θ, ν ) = f (y θ, ν ), Assume that θ and ν are orthogonal. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

108 PART IV. Divergence based priors The problem with nuisance parameters: some few words M 1 : f 1 (y ν 1 ) = f (y θ 0, ν) vs. M 2 : f 2 (y θ, ν ) = f (y θ, ν ), Assume that θ and ν are orthogonal. Use π 1 (ν) = π1 N (ν) and, gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

109 PART IV. Divergence based priors The problem with nuisance parameters: some few words M 1 : f 1 (y ν 1 ) = f (y θ 0, ν) vs. M 2 : f 2 (y θ, ν ) = f (y θ, ν ), Assume that θ and ν are orthogonal. Use π 1 (ν) = π1 N (ν) and, π 2 (θ, ν ) = π1 N(ν )π D (θ ν ), where ) π D (θ ν) h q (D[(θ, θ 0 ) ν] π N (θ ν). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

110 PART IV. Divergence based priors To be highlighted... DB priors have a simple form and are easy to explain! gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

111 PART IV. Divergence based priors To be highlighted... DB priors have a simple form and are easy to explain! They have appealing theoretical properties and in standard scenarios DB priors produce similar answers to other well established methods. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

112 PART IV. Divergence based priors To be highlighted... DB priors have a simple form and are easy to explain! They have appealing theoretical properties and in standard scenarios DB priors produce similar answers to other well established methods. In complex scenarios, where the other proposals fail, DB priors are well defined and produce quite reasonable results. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

113 PART IV. Divergence based priors To be highlighted... DB priors have a simple form and are easy to explain! They have appealing theoretical properties and in standard scenarios DB priors produce similar answers to other well established methods. In complex scenarios, where the other proposals fail, DB priors are well defined and produce quite reasonable results. DB priors are quite general and, we hope, can be considered a reasonable extension of Jeffreys ideas on testing. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

114 PART IV. Divergence based priors Full technical details available at: gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

115 PART IV. Divergence based priors Full technical details available at: THANKS! gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

116 PART IV. Divergence based priors Simple example: credible intervals and testing Suppose: From a sample y i N(µ, 1), of size n = 50 I get ȳ = 0.25 from which I derive the 95% credible interval using π(µ) = 1. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

117 PART IV. Divergence based priors Simple example: credible intervals and testing Suppose: From a sample y i N(µ, 1), of size n = 50 I get ȳ = 0.25 from which I derive the 95% credible interval using π(µ) = 1. I obtain CI=[0.11,0.39]. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

118 PART IV. Divergence based priors Simple example: credible intervals and testing Suppose: From a sample y i N(µ, 1), of size n = 50 I get ȳ = 0.25 from which I derive the 95% credible interval using π(µ) = 1. I obtain CI=[0.11,0.39]. I am interested in knowing if H : µ = 0 = no effect (although I did not explicitly say). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

119 PART IV. Divergence based priors Simple example: credible intervals and testing Suppose: From a sample y i N(µ, 1), of size n = 50 I get ȳ = 0.25 from which I derive the 95% credible interval using π(µ) = 1. I obtain CI=[0.11,0.39]. I am interested in knowing if H : µ = 0 = no effect (although I did not explicitly say). From the interval above, I would decide Bayesianly, to reject H (say has a probability which is less than 0.05). But gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

120 PART IV. Divergence based priors Simple example: credible intervals and testing Suppose: From a sample y i N(µ, 1), of size n = 50 I get ȳ = 0.25 from which I derive the 95% credible interval using π(µ) = 1. I obtain CI=[0.11,0.39]. I am interested in knowing if H : µ = 0 = no effect (although I did not explicitly say). From the interval above, I would decide Bayesianly, to reject H (say has a probability which is less than 0.05). But Prob(H y) = 0.66 Come back! gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

121 PART IV. Divergence based priors Standard scenarios In linear models, the DB priors coincide with the proposals of Zellner and Siow (1980, 1984). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

122 PART IV. Divergence based priors Standard scenarios In linear models, the DB priors coincide with the proposals of Zellner and Siow (1980, 1984). In other traditional problems (Bernoulli, exponential, Poisson or Normal), DB priors produce Bayes factors very similar to the intrinsic. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

123 PART IV. Divergence based priors Standard scenarios In linear models, the DB priors coincide with the proposals of Zellner and Siow (1980, 1984). In other traditional problems (Bernoulli, exponential, Poisson or Normal), DB priors produce Bayes factors very similar to the intrinsic. H 1 : (µ, σ) = (0, 1) vs.h 2 : (µ, σ) (0, 1) Come back! gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

124 PART IV. Divergence based priors Scenario with improper likelihood Let f (y θ, p) = p N(y 0, 1) + (1 p) N(y θ, 1), (p known) The problem: H 1 : θ = 0 vs. H 2 : θ 0. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

125 PART IV. Divergence based priors Scenario with improper likelihood Let f (y θ, p) = p N(y 0, 1) + (1 p) N(y θ, 1), (p known) The problem: H 1 : θ = 0 vs. H 2 : θ 0. Intrinsic Bayes factors do not exist. Berger and Pericchi (2001) recommend using π2 BP (θ) = Ca(θ 0, 1). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

126 PART IV. Divergence based priors Scenario with improper likelihood Let f (y θ, p) = p N(y 0, 1) + (1 p) N(y θ, 1), (p known) The problem: H 1 : θ = 0 vs. H 2 : θ 0. Intrinsic Bayes factors do not exist. Berger and Pericchi (2001) recommend using π2 BP (θ) = Ca(θ 0, 1). The DB prior does not have a closed form, although a good approximation is π D 2 (θ) Ca(θ 0, 1/(1 p)). gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

127 PART IV. Divergence based priors Comparisons π D (solid line), Ca(0, 1/1 p) (dashed) and π2 BP (µ) = Ca(µ 0, 1) (dots) mu mu p=0.50 Come back! p=0.75 gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

128 PART IV. Divergence based priors Scenario with irregular models Sea f (y θ) = exp{ (y θ)}, y > θ, and the problem H 1 : θ = θ 0 vs. H 2 : θ > θ 0. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

129 PART IV. Divergence based priors Scenario with irregular models Sea f (y θ) = exp{ (y θ)}, y > θ, and the problem H 1 : θ = θ 0 vs. H 2 : θ > θ 0. These models pose in serious problems to the intrinsic priors. They conjecture that it is π AI 2 (θ) = ( e θ θ 0 log(1 e θ 0 θ ) 1 ), θ > θ 0. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

130 PART IV. Divergence based priors Scenario with irregular models Sea f (y θ) = exp{ (y θ)}, y > θ, and the problem H 1 : θ = θ 0 vs. H 2 : θ > θ 0. These models pose in serious problems to the intrinsic priors. They conjecture that it is π AI 2 (θ) = ( e θ θ 0 log(1 e θ 0 θ ) 1 ), θ > θ 0. DB prior is π D 2 (θ) = ( 1 + 2(θ θ 0 ) ) 3/2, θ > θ0. gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

131 PART IV. Divergence based priors Comparisons (θ 0 = 0) For, π D 2 (solid line) y πai 2 (dashed line) Come back! gonzalo garcía-donato and susie Bayarri () DB priors May 22, / 46

Divergence Based Priors for Bayesian Hypothesis testing

Divergence Based Priors for Bayesian Hypothesis testing Divergence Based Priors for Bayesian Hypothesis testing M.J. Bayarri University of Valencia G. García-Donato University of Castilla-La Mancha November, 2006 Abstract Maybe the main difficulty for objective

More information

Lecture 5: Conventional Model Selection Priors

Lecture 5: Conventional Model Selection Priors Lecture 5: Conventional Model Selection Priors Susie Bayarri University of Valencia CBMS Conference on Model Uncertainty and Multiplicity July 23-28, 2012 Outline The general linear model and Orthogonalization

More information

1 Hypothesis Testing and Model Selection

1 Hypothesis Testing and Model Selection A Short Course on Bayesian Inference (based on An Introduction to Bayesian Analysis: Theory and Methods by Ghosh, Delampady and Samanta) Module 6: From Chapter 6 of GDS 1 Hypothesis Testing and Model Selection

More information

Invariant HPD credible sets and MAP estimators

Invariant HPD credible sets and MAP estimators Bayesian Analysis (007), Number 4, pp. 681 69 Invariant HPD credible sets and MAP estimators Pierre Druilhet and Jean-Michel Marin Abstract. MAP estimators and HPD credible sets are often criticized in

More information

7. Estimation and hypothesis testing. Objective. Recommended reading

7. Estimation and hypothesis testing. Objective. Recommended reading 7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing

More information

Hypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33

Hypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33 Hypothesis Testing Econ 690 Purdue University Justin L. Tobias (Purdue) Testing 1 / 33 Outline 1 Basic Testing Framework 2 Testing with HPD intervals 3 Example 4 Savage Dickey Density Ratio 5 Bartlett

More information

Harrison B. Prosper. CMS Statistics Committee

Harrison B. Prosper. CMS Statistics Committee Harrison B. Prosper Florida State University CMS Statistics Committee 08-08-08 Bayesian Methods: Theory & Practice. Harrison B. Prosper 1 h Lecture 3 Applications h Hypothesis Testing Recap h A Single

More information

Extending Conventional priors for Testing General Hypotheses in Linear Models

Extending Conventional priors for Testing General Hypotheses in Linear Models Extending Conventional priors for Testing General Hypotheses in Linear Models María Jesús Bayarri University of Valencia 46100 Valencia, Spain Gonzalo García-Donato University of Castilla-La Mancha 0071

More information

7. Estimation and hypothesis testing. Objective. Recommended reading

7. Estimation and hypothesis testing. Objective. Recommended reading 7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing

More information

An Extended BIC for Model Selection

An Extended BIC for Model Selection An Extended BIC for Model Selection at the JSM meeting 2007 - Salt Lake City Surajit Ray Boston University (Dept of Mathematics and Statistics) Joint work with James Berger, Duke University; Susie Bayarri,

More information

An Overview of Objective Bayesian Analysis

An Overview of Objective Bayesian Analysis An Overview of Objective Bayesian Analysis James O. Berger Duke University visiting the University of Chicago Department of Statistics Spring Quarter, 2011 1 Lectures Lecture 1. Objective Bayesian Analysis:

More information

Bayesian Assessment of Hypotheses and Models

Bayesian Assessment of Hypotheses and Models 8 Bayesian Assessment of Hypotheses and Models This is page 399 Printer: Opaque this 8. Introduction The three preceding chapters gave an overview of how Bayesian probability models are constructed. Once

More information

Bayesian Inference for Normal Mean

Bayesian Inference for Normal Mean Al Nosedal. University of Toronto. November 18, 2015 Likelihood of Single Observation The conditional observation distribution of y µ is Normal with mean µ and variance σ 2, which is known. Its density

More information

A BAYESIAN MATHEMATICAL STATISTICS PRIMER. José M. Bernardo Universitat de València, Spain

A BAYESIAN MATHEMATICAL STATISTICS PRIMER. José M. Bernardo Universitat de València, Spain A BAYESIAN MATHEMATICAL STATISTICS PRIMER José M. Bernardo Universitat de València, Spain jose.m.bernardo@uv.es Bayesian Statistics is typically taught, if at all, after a prior exposure to frequentist

More information

Introduction to Bayesian Methods

Introduction to Bayesian Methods Introduction to Bayesian Methods Jessi Cisewski Department of Statistics Yale University Sagan Summer Workshop 2016 Our goal: introduction to Bayesian methods Likelihoods Priors: conjugate priors, non-informative

More information

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models Matthew S. Johnson New York ASA Chapter Workshop CUNY Graduate Center New York, NY hspace1in December 17, 2009 December

More information

Introduction to Bayesian Statistics

Introduction to Bayesian Statistics Bayesian Parameter Estimation Introduction to Bayesian Statistics Harvey Thornburg Center for Computer Research in Music and Acoustics (CCRMA) Department of Music, Stanford University Stanford, California

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Checking for Prior-Data Conflict

Checking for Prior-Data Conflict Bayesian Analysis (2006) 1, Number 4, pp. 893 914 Checking for Prior-Data Conflict Michael Evans and Hadas Moshonov Abstract. Inference proceeds from ingredients chosen by the analyst and data. To validate

More information

1 Priors, Contd. ISyE8843A, Brani Vidakovic Handout Nuisance Parameters

1 Priors, Contd. ISyE8843A, Brani Vidakovic Handout Nuisance Parameters ISyE8843A, Brani Vidakovic Handout 6 Priors, Contd.. Nuisance Parameters Assume that unknown parameter is θ, θ ) but we are interested only in θ. The parameter θ is called nuisance parameter. How do we

More information

Some Curiosities Arising in Objective Bayesian Analysis

Some Curiosities Arising in Objective Bayesian Analysis . Some Curiosities Arising in Objective Bayesian Analysis Jim Berger Duke University Statistical and Applied Mathematical Institute Yale University May 15, 2009 1 Three vignettes related to John s work

More information

STAT 425: Introduction to Bayesian Analysis

STAT 425: Introduction to Bayesian Analysis STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 1) Fall 2017 1 / 10 Lecture 7: Prior Types Subjective

More information

Noninformative Priors for the Ratio of the Scale Parameters in the Inverted Exponential Distributions

Noninformative Priors for the Ratio of the Scale Parameters in the Inverted Exponential Distributions Communications for Statistical Applications and Methods 03, Vol. 0, No. 5, 387 394 DOI: http://dx.doi.org/0.535/csam.03.0.5.387 Noninformative Priors for the Ratio of the Scale Parameters in the Inverted

More information

Overall Objective Priors

Overall Objective Priors Overall Objective Priors Jim Berger, Jose Bernardo and Dongchu Sun Duke University, University of Valencia and University of Missouri Recent advances in statistical inference: theory and case studies University

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 Suggested Projects: www.cs.ubc.ca/~arnaud/projects.html First assignement on the web: capture/recapture.

More information

Eco517 Fall 2004 C. Sims MIDTERM EXAM

Eco517 Fall 2004 C. Sims MIDTERM EXAM Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering

More information

Module 22: Bayesian Methods Lecture 9 A: Default prior selection

Module 22: Bayesian Methods Lecture 9 A: Default prior selection Module 22: Bayesian Methods Lecture 9 A: Default prior selection Peter Hoff Departments of Statistics and Biostatistics University of Washington Outline Jeffreys prior Unit information priors Empirical

More information

Remarks on Improper Ignorance Priors

Remarks on Improper Ignorance Priors As a limit of proper priors Remarks on Improper Ignorance Priors Two caveats relating to computations with improper priors, based on their relationship with finitely-additive, but not countably-additive

More information

Stat260: Bayesian Modeling and Inference Lecture Date: February 10th, Jeffreys priors. exp 1 ) p 2

Stat260: Bayesian Modeling and Inference Lecture Date: February 10th, Jeffreys priors. exp 1 ) p 2 Stat260: Bayesian Modeling and Inference Lecture Date: February 10th, 2010 Jeffreys priors Lecturer: Michael I. Jordan Scribe: Timothy Hunter 1 Priors for the multivariate Gaussian Consider a multivariate

More information

On the use of non-local prior densities in Bayesian hypothesis tests

On the use of non-local prior densities in Bayesian hypothesis tests J. R. Statist. Soc. B (2010) 72, Part 2, pp. 143 170 On the use of non-local prior densities in Bayesian hypothesis tests Valen E. Johnson M. D. Anderson Cancer Center, Houston, USA and David Rossell Institute

More information

Unified Frequentist and Bayesian Testing of a Precise Hypothesis

Unified Frequentist and Bayesian Testing of a Precise Hypothesis Statistical Science 1997, Vol. 12, No. 3, 133 160 Unified Frequentist and Bayesian Testing of a Precise Hypothesis J. O. Berger, B. Boukai and Y. Wang Abstract. In this paper, we show that the conditional

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

STAT 740: Testing & Model Selection

STAT 740: Testing & Model Selection STAT 740: Testing & Model Selection Timothy Hanson Department of Statistics, University of South Carolina Stat 740: Statistical Computing 1 / 34 Testing & model choice, likelihood-based A common way to

More information

Comparing Non-informative Priors for Estimation and. Prediction in Spatial Models

Comparing Non-informative Priors for Estimation and. Prediction in Spatial Models Comparing Non-informative Priors for Estimation and Prediction in Spatial Models Vigre Semester Report by: Regina Wu Advisor: Cari Kaufman January 31, 2010 1 Introduction Gaussian random fields with specified

More information

Bayesian variable selection in high dimensional problems without assumptions on prior model probabilities

Bayesian variable selection in high dimensional problems without assumptions on prior model probabilities arxiv:1607.02993v1 [stat.me] 11 Jul 2016 Bayesian variable selection in high dimensional problems without assumptions on prior model probabilities J. O. Berger 1, G. García-Donato 2, M. A. Martínez-Beneito

More information

Objective Prior for the Number of Degrees of Freedom of a t Distribution

Objective Prior for the Number of Degrees of Freedom of a t Distribution Bayesian Analysis (4) 9, Number, pp. 97 Objective Prior for the Number of Degrees of Freedom of a t Distribution Cristiano Villa and Stephen G. Walker Abstract. In this paper, we construct an objective

More information

The Jeffreys Prior. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) The Jeffreys Prior MATH / 13

The Jeffreys Prior. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) The Jeffreys Prior MATH / 13 The Jeffreys Prior Yingbo Li Clemson University MATH 9810 Yingbo Li (Clemson) The Jeffreys Prior MATH 9810 1 / 13 Sir Harold Jeffreys English mathematician, statistician, geophysicist, and astronomer His

More information

ST 740: Model Selection

ST 740: Model Selection ST 740: Model Selection Alyson Wilson Department of Statistics North Carolina State University November 25, 2013 A. Wilson (NCSU Statistics) Model Selection November 25, 2013 1 / 29 Formal Bayesian Model

More information

arxiv: v1 [math.st] 10 Aug 2007

arxiv: v1 [math.st] 10 Aug 2007 The Marginalization Paradox and the Formal Bayes Law arxiv:0708.1350v1 [math.st] 10 Aug 2007 Timothy C. Wallstrom Theoretical Division, MS B213 Los Alamos National Laboratory Los Alamos, NM 87545 tcw@lanl.gov

More information

1 Introduction. P (n = 1 red ball drawn) =

1 Introduction. P (n = 1 red ball drawn) = Introduction Exercises and outline solutions. Y has a pack of 4 cards (Ace and Queen of clubs, Ace and Queen of Hearts) from which he deals a random of selection 2 to player X. What is the probability

More information

Multinomial Data. f(y θ) θ y i. where θ i is the probability that a given trial results in category i, i = 1,..., k. The parameter space is

Multinomial Data. f(y θ) θ y i. where θ i is the probability that a given trial results in category i, i = 1,..., k. The parameter space is Multinomial Data The multinomial distribution is a generalization of the binomial for the situation in which each trial results in one and only one of several categories, as opposed to just two, as in

More information

Statistical Theory MT 2006 Problems 4: Solution sketches

Statistical Theory MT 2006 Problems 4: Solution sketches Statistical Theory MT 006 Problems 4: Solution sketches 1. Suppose that X has a Poisson distribution with unknown mean θ. Determine the conjugate prior, and associate posterior distribution, for θ. Determine

More information

Chapter 8: Sampling distributions of estimators Sections

Chapter 8: Sampling distributions of estimators Sections Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample variance Skip: p.

More information

Statistical Theory MT 2007 Problems 4: Solution sketches

Statistical Theory MT 2007 Problems 4: Solution sketches Statistical Theory MT 007 Problems 4: Solution sketches 1. Consider a 1-parameter exponential family model with density f(x θ) = f(x)g(θ)exp{cφ(θ)h(x)}, x X. Suppose that the prior distribution has the

More information

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics A short review of the principles of mathematical statistics (or, what you should have learned in EC 151).

More information

Bayesian Regression (1/31/13)

Bayesian Regression (1/31/13) STA613/CBB540: Statistical methods in computational biology Bayesian Regression (1/31/13) Lecturer: Barbara Engelhardt Scribe: Amanda Lea 1 Bayesian Paradigm Bayesian methods ask: given that I have observed

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 CS students: don t forget to re-register in CS-535D. Even if you just audit this course, please do register.

More information

A Bayesian perspective on GMM and IV

A Bayesian perspective on GMM and IV A Bayesian perspective on GMM and IV Christopher A. Sims Princeton University sims@princeton.edu November 26, 2013 What is a Bayesian perspective? A Bayesian perspective on scientific reporting views all

More information

Unified Conditional Frequentist and Bayesian Testing of Composite Hypotheses

Unified Conditional Frequentist and Bayesian Testing of Composite Hypotheses Unified Conditional Frequentist and Bayesian Testing of Composite Hypotheses Sarat C. Dass University of Michigan James O. Berger Duke University October 26, 2000 Abstract Testing of a composite null hypothesis

More information

I. Bayesian econometrics

I. Bayesian econometrics I. Bayesian econometrics A. Introduction B. Bayesian inference in the univariate regression model C. Statistical decision theory Question: once we ve calculated the posterior distribution, what do we do

More information

Bayesian model selection for computer model validation via mixture model estimation

Bayesian model selection for computer model validation via mixture model estimation Bayesian model selection for computer model validation via mixture model estimation Kaniav Kamary ATER, CNAM Joint work with É. Parent, P. Barbillon, M. Keller and N. Bousquet Outline Computer model validation

More information

Mathematical Statistics

Mathematical Statistics Mathematical Statistics MAS 713 Chapter 8 Previous lecture: 1 Bayesian Inference 2 Decision theory 3 Bayesian Vs. Frequentist 4 Loss functions 5 Conjugate priors Any questions? Mathematical Statistics

More information

A simple two-sample Bayesian t-test for hypothesis testing

A simple two-sample Bayesian t-test for hypothesis testing A simple two-sample Bayesian t-test for hypothesis testing arxiv:159.2568v1 [stat.me] 8 Sep 215 Min Wang Department of Mathematical Sciences, Michigan Technological University, Houghton, MI, USA and Guangying

More information

STAT 830 Bayesian Estimation

STAT 830 Bayesian Estimation STAT 830 Bayesian Estimation Richard Lockhart Simon Fraser University STAT 830 Fall 2011 Richard Lockhart (Simon Fraser University) STAT 830 Bayesian Estimation STAT 830 Fall 2011 1 / 23 Purposes of These

More information

General Bayesian Inference I

General Bayesian Inference I General Bayesian Inference I Outline: Basic concepts, One-parameter models, Noninformative priors. Reading: Chapters 10 and 11 in Kay-I. (Occasional) Simplified Notation. When there is no potential for

More information

Seminar über Statistik FS2008: Model Selection

Seminar über Statistik FS2008: Model Selection Seminar über Statistik FS2008: Model Selection Alessia Fenaroli, Ghazale Jazayeri Monday, April 2, 2008 Introduction Model Choice deals with the comparison of models and the selection of a model. It can

More information

F79SM STATISTICAL METHODS

F79SM STATISTICAL METHODS F79SM STATISTICAL METHODS SUMMARY NOTES 9 Hypothesis testing 9.1 Introduction As before we have a random sample x of size n of a population r.v. X with pdf/pf f(x;θ). The distribution we assign to X is

More information

Statistical Data Analysis Stat 3: p-values, parameter estimation

Statistical Data Analysis Stat 3: p-values, parameter estimation Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,

More information

Predictive Distributions

Predictive Distributions Predictive Distributions October 6, 2010 Hoff Chapter 4 5 October 5, 2010 Prior Predictive Distribution Before we observe the data, what do we expect the distribution of observations to be? p(y i ) = p(y

More information

Chapter 5. Bayesian Statistics

Chapter 5. Bayesian Statistics Chapter 5. Bayesian Statistics Principles of Bayesian Statistics Anything unknown is given a probability distribution, representing degrees of belief [subjective probability]. Degrees of belief [subjective

More information

Bayesian Hypothesis Testing: Redux

Bayesian Hypothesis Testing: Redux Bayesian Hypothesis Testing: Redux Hedibert F. Lopes and Nicholas G. Polson Insper and Chicago Booth arxiv:1808.08491v1 [math.st] 26 Aug 2018 First draft: March 2018 This draft: August 2018 Abstract Bayesian

More information

Foundations of Statistical Inference

Foundations of Statistical Inference Foundations of Statistical Inference Julien Berestycki Department of Statistics University of Oxford MT 2016 Julien Berestycki (University of Oxford) SB2a MT 2016 1 / 20 Lecture 6 : Bayesian Inference

More information

Integrated Objective Bayesian Estimation and Hypothesis Testing

Integrated Objective Bayesian Estimation and Hypothesis Testing Integrated Objective Bayesian Estimation and Hypothesis Testing José M. Bernardo Universitat de València, Spain jose.m.bernardo@uv.es 9th Valencia International Meeting on Bayesian Statistics Benidorm

More information

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007 Bayesian inference Fredrik Ronquist and Peter Beerli October 3, 2007 1 Introduction The last few decades has seen a growing interest in Bayesian inference, an alternative approach to statistical inference.

More information

Harrison B. Prosper. CMS Statistics Committee

Harrison B. Prosper. CMS Statistics Committee Harrison B. Prosper Florida State University CMS Statistics Committee 7 August, 2008 Bayesian Methods: Theory & Practice. Harrison B. Prosper 1 h Lecture 2 Foundations & Applications h The Bayesian Approach

More information

Checking for Prior-Data Conßict. by Michael Evans Department of Statistics University of Toronto. and

Checking for Prior-Data Conßict. by Michael Evans Department of Statistics University of Toronto. and Checking for Prior-Data Conßict by Michael Evans Department of Statistics University of Toronto and Hadas Moshonov Department of Statistics University of Toronto Technical Report No. 413, February 24,

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Introduction to Applied Bayesian Modeling. ICPSR Day 4

Introduction to Applied Bayesian Modeling. ICPSR Day 4 Introduction to Applied Bayesian Modeling ICPSR Day 4 Simple Priors Remember Bayes Law: Where P(A) is the prior probability of A Simple prior Recall the test for disease example where we specified the

More information

Part 4: Multi-parameter and normal models

Part 4: Multi-parameter and normal models Part 4: Multi-parameter and normal models 1 The normal model Perhaps the most useful (or utilized) probability model for data analysis is the normal distribution There are several reasons for this, e.g.,

More information

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Partitioning the Parameter Space. Topic 18 Composite Hypotheses Topic 18 Composite Hypotheses Partitioning the Parameter Space 1 / 10 Outline Partitioning the Parameter Space 2 / 10 Partitioning the Parameter Space Simple hypotheses limit us to a decision between one

More information

A union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling

A union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling A union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling Min-ge Xie Department of Statistics, Rutgers University Workshop on Higher-Order Asymptotics

More information

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability

More information

Bayesian model selection: methodology, computation and applications

Bayesian model selection: methodology, computation and applications Bayesian model selection: methodology, computation and applications David Nott Department of Statistics and Applied Probability National University of Singapore Statistical Genomics Summer School Program

More information

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017 Bayesian inference Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 10, 2017 1 / 22 Outline for today A genetic example Bayes theorem Examples Priors Posterior summaries

More information

A Discussion of the Bayesian Approach

A Discussion of the Bayesian Approach A Discussion of the Bayesian Approach Reference: Chapter 10 of Theoretical Statistics, Cox and Hinkley, 1974 and Sujit Ghosh s lecture notes David Madigan Statistics The subject of statistics concerns

More information

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1 Chapter 4 HOMEWORK ASSIGNMENTS These homeworks may be modified as the semester progresses. It is your responsibility to keep up to date with the correctly assigned homeworks. There may be some errors in

More information

Bias Variance Trade-off

Bias Variance Trade-off Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]

More information

P Values and Nuisance Parameters

P Values and Nuisance Parameters P Values and Nuisance Parameters Luc Demortier The Rockefeller University PHYSTAT-LHC Workshop on Statistical Issues for LHC Physics CERN, Geneva, June 27 29, 2007 Definition and interpretation of p values;

More information

Foundations of Statistical Inference

Foundations of Statistical Inference Foundations of Statistical Inference Julien Berestycki Department of Statistics University of Oxford MT 2016 Julien Berestycki (University of Oxford) SB2a MT 2016 1 / 32 Lecture 14 : Variational Bayes

More information

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference 1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE

More information

Bayesian Model Averaging

Bayesian Model Averaging Bayesian Model Averaging Hoff Chapter 9, Hoeting et al 1999, Clyde & George 2004, Liang et al 2008 October 24, 2017 Bayesian Model Choice Models for the variable selection problem are based on a subset

More information

Bayesian estimation of the discrepancy with misspecified parametric models

Bayesian estimation of the discrepancy with misspecified parametric models Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012

More information

Introduction to Bayesian Inference: Supplemental Topics

Introduction to Bayesian Inference: Supplemental Topics Introduction to Bayesian Inference: Supplemental Topics Tom Loredo Cornell Center for Astrophysics and Planetary Science http://www.astro.cornell.edu/staff/loredo/bayes/ CASt Summer School 7 8 June 2017

More information

A Very Brief Summary of Bayesian Inference, and Examples

A Very Brief Summary of Bayesian Inference, and Examples A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X

More information

Search Procedures in High Energy Physics

Search Procedures in High Energy Physics Version 2.20 July 1, 2009 Search Procedures in High Energy Physics Luc Demortier 1 Laboratory of Experimental High-Energy Physics The Rockefeller University Abstract The usual procedure for searching for

More information

Bayesian Inference: Concept and Practice

Bayesian Inference: Concept and Practice Inference: Concept and Practice fundamentals Johan A. Elkink School of Politics & International Relations University College Dublin 5 June 2017 1 2 3 Bayes theorem In order to estimate the parameters of

More information

BAYESIAN MODEL CRITICISM

BAYESIAN MODEL CRITICISM Monte via Chib s BAYESIAN MODEL CRITICM Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes

More information

Objective Bayesian Upper Limits for Poisson Processes

Objective Bayesian Upper Limits for Poisson Processes CDF/MEMO/STATISTICS/PUBLIC/5928 Version 2.1 June 24, 25 Objective Bayesian Upper Limits for Poisson Processes Luc Demortier Laboratory of Experimental High-Energy Physics The Rockefeller University...all

More information

Confidence Distribution

Confidence Distribution Confidence Distribution Xie and Singh (2013): Confidence distribution, the frequentist distribution estimator of a parameter: A Review Céline Cunen, 15/09/2014 Outline of Article Introduction The concept

More information

Bayesian Statistics Adrian Raftery and Jeff Gill One-day course for the American Sociological Association August 15, 2002

Bayesian Statistics Adrian Raftery and Jeff Gill One-day course for the American Sociological Association August 15, 2002 Bayesian Statistics Adrian Raftery and Jeff Gill One-day course for the American Sociological Association August 15, 2002 Bayes Course, ASA Meeting, August 2002 c Adrian E. Raftery 2002 1 Outline 1. Bayes

More information

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero 1999 32 Statistic used Meaning in plain english Reduction ratio T (X) [X 1,..., X n ] T, entire data sample RR 1 T (X) [X (1),..., X (n) ] T, rank

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Chapter 8 Maximum Likelihood Estimation 8. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function

More information

Mixtures of Prior Distributions

Mixtures of Prior Distributions Mixtures of Prior Distributions Hoff Chapter 9, Liang et al 2007, Hoeting et al (1999), Clyde & George (2004) November 10, 2016 Bartlett s Paradox The Bayes factor for comparing M γ to the null model:

More information

Bayesian Inference. Chapter 9. Linear models and regression

Bayesian Inference. Chapter 9. Linear models and regression Bayesian Inference Chapter 9. Linear models and regression M. Concepcion Ausin Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering

More information

Unobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior:

Unobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior: Pi Priors Unobservable Parameter population proportion, p prior: π ( p) Conjugate prior π ( p) ~ Beta( a, b) same PDF family exponential family only Posterior π ( p y) ~ Beta( a + y, b + n y) Observed

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

A Bayesian Treatment of Linear Gaussian Regression

A Bayesian Treatment of Linear Gaussian Regression A Bayesian Treatment of Linear Gaussian Regression Frank Wood December 3, 2009 Bayesian Approach to Classical Linear Regression In classical linear regression we have the following model y β, σ 2, X N(Xβ,

More information

Bayesian Inference. Chapter 2: Conjugate models

Bayesian Inference. Chapter 2: Conjugate models Bayesian Inference Chapter 2: Conjugate models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in

More information

Bayesian Loss-based Approach to Change Point Analysis

Bayesian Loss-based Approach to Change Point Analysis Bayesian Loss-based Approach to Change Point Analysis Laurentiu Hinoveanu 1, Fabrizio Leisen 1, and Cristiano Villa 1 arxiv:1702.05462v2 stat.me] 7 Jan 2018 1 School of Mathematics, Statistics and Actuarial

More information