Likelihood Inference in the Presence of Nuisance Parameters

Size: px
Start display at page:

Download "Likelihood Inference in the Presence of Nuisance Parameters"

Transcription

1 Likelihood Inference in the Presence of Nuance Parameters N Reid, DAS Fraser Department of Stattics, University of Toronto, Toronto Canada M5S 3G3 We describe some recent approaches to likelihood based inference in the presence of nuance parameters Our approach based on plotting the likelihood function and the p-value function, using recently developed third order approximations Orthogonal parameters and adjustments to profile likelihood are also dcussed Connections to classical approaches of conditional and marginal inference are outlined 1 INTRODUCTION We take the view that the most effective form of inference provided by the observed likelihood function along with the associated p-value function In the case of a scalar parameter the likelihood function simply proportional to the density function The p-value function can be obtained exactly if there a one-dimensional stattic that measures the parameter If not, the p-value can be obtained to a high order of approximation using recently developed methods of likelihood asymptotics In the presence of nuance parameters, the likelihood function for a (one-dimensional parameter of interest obtained via an adjustment to the profile likelihood function The p-value function obtained from quantities computed from the likelihood function using a canonical parametrization ϕ = ϕ(θ, which computed locally at the data point Th generalizes the method of eliminating nuance parameters by conditioning or marginalizing to more general contexts In Section 2 we give some background notation and introduce the notion of orthogonal parameters In Section 3 we illustrate the p-value function approach in a simple model with no nuance parameters Profile likelihood and adjustments to profile likelihood are described in Section 4 Third order p-values for problems with nuance parameters are described in Section 5 Section 6 describes the classical conditional and marginal likelihood approach 2 NOTATION AND ORTHOGONAL PARAMETERS We assume our measurement(s y can be modelled as coming from a probability dtribution with density or mass function f(y; θ, where θ = (ψ, λ takes values in R d We assume ψ a one-dimensional parameter of interest, and λ a vector of nuance parameters If there interest in more than one component of θ, the methods described here can be applied to each component of interest in turn The likelihood function L(θ = L(θ; y = c(yf(y; θ; (1 it defined only up to arbitrary multiples which may depend on y but not on θ Th ensures in particular that the likelihood function invariant to one-toone transformations of the measurement(s y In the context of independent, identically dtributed sampling, where y = (y 1,, y n and each y i follows the model f(y; θ the likelihood function proportional to Πf(y i ; θ and the log-likelihood function becomes a sum of independent and identically dtributed components: l(θ = l(θ; y = Σ log f(y i ; θ + a(y (2 The maximum likelihood estimate ˆθ the value of θ at which the likelihood takes its maximum, and in regular models defined by the score equation l (ˆθ; y = 0 (3 The observed Fher information function j(θ the curvature of the log-likelihood: j(θ = l (θ (4 and the expected Fher information the model quantity i(θ = E{ l (θ} = l (θ; yf(y; θdy (5 If y a sample of size n then i(θ = O(n In accord with the partitioning of θ we partition the observed and expected information matrices and use the notation ( iψψ i i(θ = ψλ (6 i λψ i λλ and ( i 1 i (θ = ψψ i ψλ i λψ i λλ (7 We say ψ orthogonal to λ (with respect to expected Fher information if i ψλ (θ = 0 When ψ scalar a transformation from (ψ, λ to (ψ, η(ψ, λ such that ψ orthogonal to η can always be found (Cox and Reid [1] The most directly interpreted consequence 265

2 of parameter orthogonality that the maximum likelihood estimates of orthogonal components are asymptotically independent Example 1: ratio of Poson means Suppose y 1 and y 2 are independent counts modelled as Poson with mean λ and ψλ, respectively Then the likelihood function L(ψ, λ; y 1, y 2 = e λ(1+ψ ψ y2 λ y1+y2 and ψ orthogonal to η(ψ, λ = λ(ψ + 1 In fact in th example the likelihood function factors as L 1 (ψl 2 (η, which a stronger property than parameter orthogonality The first factor the likelihood for a binomial dtribution with index y 1 + y 2 and probability of success ψ/(1 + ψ, and the second that for a Poson dtribution with mean η Example 2: exponential regression Suppose y i, i = 1,, n are independent observations, each from an exponential dtribution with mean λ exp( ψx i, where x i known The log-likelihood function l(ψ, λ; y = n log λ + ψσx i λ 1 Σy i exp(ψx i (8 and i ψλ (θ = 0 if and only if Σx i = 0 The stronger property of factorization of the likelihood does not hold 3 LIKELIHOOD INFERENCE WITH NO NUISANCE PARAMETERS We assume now that θ one-dimensional A plot of the log-likelihood function as a function of θ can quickly reveal irregularities in the model, such as a non-unique maximum, or a maximum on the boundary, and can also provide a vual guide to deviance from normality, as the log-likelihood function for a normal dtribution a parabola and hence symmetric about the maximum In order to calibrate the log-likelihood function we can use the approximation r(θ = sign(ˆθ θ[2{l(ˆθ l(θ}] 1/2 N(0, 1, (9 which equivalent to the result that twice the log likelihood ratio approximately χ 2 1 Th will typically provide a better approximation than the asymptotically equivalent result that ˆθ θ N(0, i 1 (θ (10 as it partially accommodates the potential asymmetry in the log-likelihood function These two approximations are sometimes called first order approximations because in the context where the log-likelihood O(n, we have (under regularity conditions results such as Pr{r(θ; y r(θ; y 0 } = Pr{Z r(θ; y 0 } (11 {1 + O(n 1/2 } Table I The p-values for testing µ = 0, ie that the number of observed events constent with the background upper p-value lower p-value mid p-value Φ(r Φ(r Φ{(ˆθ θĵ 1/2 } where Z follows a standard normal dtribution It relatively simple to improve the approximation to third order, ie with relative error O(n 3/2, using the so-called r approximation r (θ = r(θ + {1/r(θ} log{q(θ/r(θ} N(0, 1 (12 where q(θ a likelihood-based stattic and a generalization of the Wald stattic (ˆθ θj 1/2 (ˆθ; see Fraser [2] Example 3: truncated Poson Suppose that y follows a Poson dtribution with mean θ = b + µ, where b a background rate that assumed known In th model the p-value function can be computed exactly simply by summing the Poson probabilities Because the Poson dtribution dcrete, the p-value could reasonably be defined as either or Pr(y y 0 ; θ (13 Pr(y < y 0 ; θ, (14 sometimes called the upper and lower p-values, respectively For the values y 0 = 17, b = 67, Figure 1 shows the likelihood function as a function of µ and the p- value function p(µ computed using both the upper and lower p-values In Figure 2 we plot the mid p- value, which Pr(y < y 0 + (1/2Pr(y = y 0 (15 The approximation based on r nearly identical to the mid-p-value; the difference cannot be seen on Figure 2 Table 1 compares the p-values at µ = 0 Th example taken from Fraser, Reid and Wong [3] 4 PROFILE AND ADJUSTED PROFILE LIKELIHOOD FUNCTIONS We now assume θ = (ψ, λ and denote by ˆλ ψ the restricted maximum likelihood estimate obtained by 266

3 also sometimes called the concentrated likelihood or the peak likelihood The approximations of the previous section generalize to likelihood p-value mu Figure 1: The likelihood function (top and p-value function (bottom for the Poson model, with b = 67 and y 0 = 17 For µ = 0 the p-value interval (099940, pvalue mu Figure 2: The upper and lower p-value functions and the mid-p-value function for the Poson model, with b = 67 and y 0 = 17 The approximation based on Φ(r identical to the mid-p-value function to the drawing accuracy maximizing the likelihood function over the nuance parameter λ with ψ fixed The profile likelihood function µ L p (ψ = L(ψ, ˆλ ψ ; (16 r(ψ = sign( ˆψ ψ[2{l p ( ˆψ l p (ψ}] 1/2 N(0, 1, (17 and ˆψ ψ N(0, {i ψψ (θ} 1 (18 These approximations, like the ones in Section 3, are derived from asymptotic results which assume that n, that we have a vector y of independent, identically dtributed observations, and that the dimension of the nuance parameter does not increase with n Further regularity conditions are required on the model, such as are outlined in textbook treatments of the asymptotic theory of maximum likelihood In finite samples these approximations can be mleading: profile likelihood too concentrated, and can be maximized at the wrong value Example 4: normal theory regression Suppose y i = x i β + ɛ i, where x i = (x i1,, x ip a vector of known covariate values, β an unknown parameter of length p, and ɛ i assumed to follow a N(0, ψ dtribution The maximum likelihood estimate of ψ ˆψ = 1 n Σ(y i x i ˆβ 2 (19 which tends to be too small, as it does not allow for the fact that p unknown parameters (the components of β have been estimated In th example there a simple improvement, based on the result that the likelihood function for (β, ψ factors into L 1 (β, ψ; ȳl 2 {ψ; Σ(y i x i ˆβ 2 } (20 where L 2 (ψ proportional to the marginal dtribution of Σ(y i x ˆβ i 2 Figure 3 shows the profile likelihood and the marginal likelihood; it easy to verify that the latter maximized at ˆψ m = 1 n p Σ(y i x i ˆβ 2 (21 which in fact an unbiased estimate of ψ Example 5: product of exponential means Suppose we have independent pairs of observations y 1i, y 2i, where y 1i Exp(ψλ i y 2i Exp(ψ/λ i, i = 1,, n The limiting normal theory for profile likelihood does not apply in th context, as the dimension of the parameter not fixed but increasing with the sample size, and it can be shown that ˆψ π 4 ψ (22 as n (Cox and Reid [4] The theory of higher order approximations can be used to derive a general improvement to the profile 267

4 PHYSTAT2003, SLAC, Stanford, California, September 8-11, 2003 likelihood σ profile marginal Figure 3: Profile likelihood and marginal likelihood for the variance parameter in a normal theory regression with 21 observations and three covariates (the Stack Loss data included in the Splus dtribution The profile likelihood maximized at a smaller value of ψ, and narrower; in th case both the estimate and its estimated standard error are too small likelihood or log-likelihood function, which takes the form l a (ψ = l p (ψ log j λλ(ψ, ˆλ ψ + B(ψ (23 where j λλ defined by the partitioning of the observed information function, and B(ψ a further adjustment function that O p (1 Several versions of B(ψ have been suggested in the stattical literature: we use the one defined in Fraser [5] given by B(ψ = 1 2 log ϕ λ(ψ, ˆλ ψ j ϕϕ ( ˆψ, ˆλϕ λ(ψ, ˆλ ψ (24 Th depends on a so-called canonical parametrization ϕ = ϕ(θ = l ;V (θ; y 0 which dcussed in Fraser, Reid and Wu [6] and Reid [7] In the special case that ψ orthogonal to the nuance parameter λ a simplification of l a (ψ available as l CR (ψ = l p (ψ 1 2 log j λλ(ψ, ˆλ ψ (25 which was first introduced in Cox and Reid (1987 The change of sign on log j comes from the orthogonality equations In iid sampling, l p (ψ O p (n, ie the sum of n bounded random variables, whereas log j O p (1 A drawback of l CR that it not invariant to one-to-one reparametrizations of λ, all of which are orthogonal to ψ In contrast l a (ψ invariant to transformations θ = (ψ, λ to θ = (ψ, η(ψ, λ, sometimes called interest-respecting transformations Example 5 continued In th example ψ orthogonal to λ = (λ 1,, λ n, and l CR (ψ = (3n/2 log ψ (2/ψΣ (y 1i y 2i (26 The value that maximizes l CR more nearly constent than the maximum likelihood estimate as ˆψ CR (π/3ψ 5 P -VALUES FROM PROFILE LIKELIHOOD The limiting theory for profile likelihood gives first order approximations to p-values, such as and p(ψ = Φ(r p (27 p(ψ = Φ{( ˆψ ψj 1/2 p ( ˆψ} (28 although the dcussion in the previous section suggests these may not provide very accurate approximations As in the scalar parameter case, though, a much better approximation available using Φ(r where r (ψ = r p (ψ + 1/{r p (ψ} log{q(ψ/r p (ψ} (29 where Q can also be derived from the likelihood function and a function ϕ(θ, y 0 as where Q = (ˆν ˆν ψ ˆσ 1/2 ν ν(θ = e T ψϕ(θ, e ψ = ψ ϕ (ˆθ ψ / ψ ϕ (ˆθ ψ, ˆσ 2 ν = j (λλ (ˆθ ψ / j (θθ (ˆθ, j (θθ (ˆθ = j θθ (ˆθ ϕ θ (ˆθ 2, j (λλ (ˆθ ψ = j λλ (ˆθ ψ ϕ λ (ˆθ ψ 2 The derivation described in Fraser, Reid and Wu [6] and Reid [7] The key ingredients are the log-likelihood function l(θ and a reparametrization ϕ(θ = ϕ(θ; y 0, which defined by using an approximating model at the observed data point y 0 ; th approximation in turn based on a conditioning argument A closely related approach due to Barndorff- Nielsen; see Barndorff-Nielsen and Cox [8, Ch 7], and the two approaches are compared in [7] Example 6: comparing two binomials Table 2 shows the employment htory of men and women at the Space Telescope Science Institute, as reported in Science Feb We denote by y 1 the number of males who left and model th as a Binomial with sample size 19 and probability p 1 ; similarly the number of females who left, y 2, modelled as Binomial with sample size 7 and probability p 2 We write the parameter of interest ψ = log p 1(1 p 2 p 2 (1 p 1 (30 The hypothes of interest p 1 = p 2, or ψ = 0 The p- value function for ψ plotted in Figure 4 The p-value at ψ = using the normal approximation to r p, and using the normal approximation 268

5 Table II Employment of men and women at the Space Telescope Science Institute, (from Science magazine, Volume 299, page 993, 14 February 2003 pvalue Left Stayed Total Men Women Total Figure 4: The p-value function for the log-odds ratio, ψ, for the data of Table II The value ψ = 0 corresponds to the hypothes that the probabilities of leaving are equal for men and women to r Using Fher s exact test gives a mid p-value of , so the approximations are anticonservative in th case Example 7: Poson with estimated background Suppose in the context of Example 3 that we allow for imprecion in the background, replacing b by an unknown parameter β with estimated value ˆβ We assume that the background estimate obtained from a Poson count x, which has mean kβ, and the signal measurement an independent Poson count, y, with mean β+µ We have ˆβ = x/k and var ˆβ = β/k, so the estimated precion of the background gives us a value for k For example, if the background estimated to be 67 ± 21 th implies a value for k of 67/(21 2 = 15 Uncertainty in the standard error of the background ignored here We now outline the steps in the computation of the r approximation (29 The log-likelihood function based on the two independent observations x and y l(β, µ = x log(kβ kβ + y log(β + µ β µ (31 with canonical parameter ϕ = (log β, log(β + µ Then ( ϕ θ (θ = ϕ(θ 0 1/β θ =, (32 1/(β + µ 1/(β + µ ψ from which Then we have ϕ 1 θ = ( β β + µ β 0 (33 ψ ϕ = ( β, β + µ (34 χ(ˆθ = ˆβ µ log( ˆβ + ( ˆβ µ + µ log( ˆβ + ˆµ { ˆβ2 µ + ( ˆβ µ + µ 2 } (35 χ(ˆθ ψ = ˆβ µ log( ˆβ µ + ( ˆβ µ + µ log( ˆβ µ + µ { ˆβ2 µ + ( ˆβ,(36 µ + µ 2 } and finally { Q = j (θθ (ˆθ = y 1 y 2 = k/ ˆβ( ˆβ + ˆµ (37 j (λλ (ˆθ ψ = y 1( ˆβ µ + µ 2 + y 2 ˆβ2 µ ( ˆβ µ + µ 2 + ˆβ 2 µ ( ˆβ µ + µ log The likelihood root ( ˆβ + ˆµ ˆβ µ + µ ˆβ µ log ˆβ } ˆβ µ (38 {k ˆβ( ˆβ + ˆµ} 1/2 {k ˆβ( ˆβ µ + µ 2 + ( ˆβ + ˆµ ˆβ (39 µ} 2 1/2 r = sign(q [2{l( ˆβ, ˆµ l( ˆβ µ, µ}] (40 = sign(q (2[k ˆβ log{ ˆβ/ ˆβ µ } + ( ˆβ + ˆµ log{( ˆβ + ˆµ/( ˆβ µ + µ} k( ˆβ ˆβ µ { ˆβ + ˆµ ( ˆβ µ + µ}] (41 The third order approximation to the p-value function 1 Φ(r, where r = r + (1/r log(q/r (42 Figure 5 shows the p-value function for µ using the mid-p-value function from the Poson with no adjustment for the error in the background, and the p-value function from 1 Φ(r The p-value for testing µ = , allowing for the uncertainty in the background, whereas it ignoring th uncertainty The hypothes Ey = β could also be tested by modelling the mean of y as νβ, say, and testing the value ν = 1 In th formulation we can eliminate the nuance parameter exactly by using the binomial dtribution of y conditioned on the total x + y, as described in example 1 Th gives a mid-p-value of The computation much easier than that outlined above, and seems quite appropriate for testing the equality of the two means However if inference about the mean of the signal needed, in the form of a point estimate or confidence bounds, then the formulation as a ratio seems less natural at least in the context of HEP experiments A more complete comparon of methods for th problem given in Linnemann [8] 269

6 pvalue estimated background known background mu Figure 5: Comparon of the p-value functions computed assuming the background known and using the mid-p-value with the third order approximation allowing a background error of ±175 6 CONDITIONAL AND MARGINAL LIKELIHOOD In special model classes, it possible to eliminate nuance parameters by either conditioning or marginalizing The conditional or marginal likelihood then gives essentially exact inference for the parameter of interest, if th likelihood can itself be computed exactly In Example 1 above, L 1 the density for y 2 conditional on y 1 + y 2, so a conditional likelihood for ψ Th an example of the more general class of linear exponential families: f(y; ψ, λ = exp{ψs(y+λ t(y c(ψ, λ d(y}; (43 in which f cond (s t; ψ = exp{ψs C t (ψ D t (s} (44 defines the conditional likelihood The comparon of two binomials in Example 6 in th class, with ψ as defined at (30 and λ = log{p 2 /(1 p 2 } The difference of two Poson means, in Example 7, cannot be formulated th way, however, even though the Poson dtribution an exponential family, because the parameter of interest ψ not a component of the canonical parameter It can be shown that in models of the form (43 the log-likelihood l a (ψ = l p (ψ + (1/2 log j λλ approximates the conditional log-likelihood l cond (ψ = log f cond (s t; ψ, and that where p(ψ = Φ(r (45 r = r a + 1 r a log( Q r a r a = ±[2{l a ( ˆψ a l a (ψ}] 1/2 Q = ( ˆψ 1/2 a ψ{j a ( ˆψ} approximates the p-value function with relative error O(n 3/2 in iid sampling An asymptotically equivalent approximation based on the profile log-likelihood where r = r p + 1 r p log( Q r p p(ψ = Φ(r (46 r p = ±[2{l p ( ˆψ l p (ψ}] 1/2 Q = ( ˆψ ψ{j p ( ˆψ} 1/2 j λλ(ψ, ˆλ ψ 1/2 j λλ ( ˆψ, ˆλ 1/2 In the latter approximation an adjustment for nuance parameters made to Q, whereas in the former the adjustment built into the likelihood function Approximation (46 was used in Figure 3 270

7 A similar dcussion applies to the class of transformation models, using marginal approximations Both classes are reviewed in Reid [9] Acknowledgments The authors wh to thank Anthony Davon and Augustine Wong for helpful dcussion Th research was partially supported by the Natural Sciences and Engineering Research Council References [1] DR Cox and N Reid, Parameter Orthogonality and Approximate Conditional Inference, J R Statt Soc B, 47, 1, 1987 [2] DAS Fraser, Stattical Inference: Likelihood to Significance, J Am Statt Assoc , 1991 [3] DAS Fraser, N Reid and A Wong, On Inference for Bounded Parameters, arxiv: physics/ , v1, 27 Mar 2003 to appear in Phys Rev D [4] DR Cox and N Reid, A Note on the Difference between Profile and Modified Profile Likelihood, Biometrika 79, 408, 1992 [5] DAS Fraser, Likelihood for Component Parameters, Biometrika 90, 327, (2003 [6] DAS Fraser, N Reid and J Wu, A Simple General Formula for Tail Probabilities for Frequentt and Bayesian Inference, Biometrika 86, 246, 1999 [7] N Reid, Asymptotics and the Theory of Inference, Ann Statt, to appear, 2004 [8] J T Linnemann, Measures of Significance in HEP and Astrophysics, available in these Proceedings on page 35 [9] N Reid, Likelihood and Higher-Order Approximations to Tail Areas: a Review and Annotated Bibliography, Canad J Statt 24, 141,

Likelihood Inference in the Presence of Nuisance Parameters

Likelihood Inference in the Presence of Nuisance Parameters PHYSTAT2003, SLAC, September 8-11, 2003 1 Likelihood Inference in the Presence of Nuance Parameters N. Reid, D.A.S. Fraser Department of Stattics, University of Toronto, Toronto Canada M5S 3G3 We describe

More information

Likelihood inference in the presence of nuisance parameters

Likelihood inference in the presence of nuisance parameters Likelihood inference in the presence of nuisance parameters Nancy Reid, University of Toronto www.utstat.utoronto.ca/reid/research 1. Notation, Fisher information, orthogonal parameters 2. Likelihood inference

More information

Last week. posterior marginal density. exact conditional density. LTCC Likelihood Theory Week 3 November 19, /36

Last week. posterior marginal density. exact conditional density. LTCC Likelihood Theory Week 3 November 19, /36 Last week Nuisance parameters f (y; ψ, λ), l(ψ, λ) posterior marginal density π m (ψ) =. c (2π) q el P(ψ) l P ( ˆψ) j P ( ˆψ) 1/2 π(ψ, ˆλ ψ ) j λλ ( ˆψ, ˆλ) 1/2 π( ˆψ, ˆλ) j λλ (ψ, ˆλ ψ ) 1/2 l p (ψ) =

More information

ASSESSING A VECTOR PARAMETER

ASSESSING A VECTOR PARAMETER SUMMARY ASSESSING A VECTOR PARAMETER By D.A.S. Fraser and N. Reid Department of Statistics, University of Toronto St. George Street, Toronto, Canada M5S 3G3 dfraser@utstat.toronto.edu Some key words. Ancillary;

More information

Nancy Reid SS 6002A Office Hours by appointment

Nancy Reid SS 6002A Office Hours by appointment Nancy Reid SS 6002A reid@utstat.utoronto.ca Office Hours by appointment Problems assigned weekly, due the following week http://www.utstat.toronto.edu/reid/4508s16.html Various types of likelihood 1. likelihood,

More information

Default priors and model parametrization

Default priors and model parametrization 1 / 16 Default priors and model parametrization Nancy Reid O-Bayes09, June 6, 2009 Don Fraser, Elisabeta Marras, Grace Yun-Yi 2 / 16 Well-calibrated priors model f (y; θ), F(y; θ); log-likelihood l(θ)

More information

Nancy Reid SS 6002A Office Hours by appointment

Nancy Reid SS 6002A Office Hours by appointment Nancy Reid SS 6002A reid@utstat.utoronto.ca Office Hours by appointment Light touch assessment One or two problems assigned weekly graded during Reading Week http://www.utstat.toronto.edu/reid/4508s14.html

More information

DEFNITIVE TESTING OF AN INTEREST PARAMETER: USING PARAMETER CONTINUITY

DEFNITIVE TESTING OF AN INTEREST PARAMETER: USING PARAMETER CONTINUITY Journal of Statistical Research 200x, Vol. xx, No. xx, pp. xx-xx ISSN 0256-422 X DEFNITIVE TESTING OF AN INTEREST PARAMETER: USING PARAMETER CONTINUITY D. A. S. FRASER Department of Statistical Sciences,

More information

Bayesian and frequentist inference

Bayesian and frequentist inference Bayesian and frequentist inference Nancy Reid March 26, 2007 Don Fraser, Ana-Maria Staicu Overview Methods of inference Asymptotic theory Approximate posteriors matching priors Examples Logistic regression

More information

Accurate directional inference for vector parameters

Accurate directional inference for vector parameters Accurate directional inference for vector parameters Nancy Reid February 26, 2016 with Don Fraser, Nicola Sartori, Anthony Davison Nancy Reid Accurate directional inference for vector parameters York University

More information

Approximating models. Nancy Reid, University of Toronto. Oxford, February 6.

Approximating models. Nancy Reid, University of Toronto. Oxford, February 6. Approximating models Nancy Reid, University of Toronto Oxford, February 6 www.utstat.utoronto.reid/research 1 1. Context Likelihood based inference model f(y; θ), log likelihood function l(θ; y) y = (y

More information

Applied Asymptotics Case studies in higher order inference

Applied Asymptotics Case studies in higher order inference Applied Asymptotics Case studies in higher order inference Nancy Reid May 18, 2006 A.C. Davison, A. R. Brazzale, A. M. Staicu Introduction likelihood-based inference in parametric models higher order approximations

More information

Accurate directional inference for vector parameters

Accurate directional inference for vector parameters Accurate directional inference for vector parameters Nancy Reid October 28, 2016 with Don Fraser, Nicola Sartori, Anthony Davison Parametric models and likelihood model f (y; θ), θ R p data y = (y 1,...,

More information

ASYMPTOTICS AND THE THEORY OF INFERENCE

ASYMPTOTICS AND THE THEORY OF INFERENCE ASYMPTOTICS AND THE THEORY OF INFERENCE N. Reid University of Toronto Abstract Asymptotic analysis has always been very useful for deriving distributions in statistics in cases where the exact distribution

More information

Staicu, A-M., & Reid, N. (2007). On the uniqueness of probability matching priors.

Staicu, A-M., & Reid, N. (2007). On the uniqueness of probability matching priors. Staicu, A-M., & Reid, N. (2007). On the uniqueness of probability matching priors. Early version, also known as pre-print Link to publication record in Explore Bristol Research PDF-document University

More information

Approximate Inference for the Multinomial Logit Model

Approximate Inference for the Multinomial Logit Model Approximate Inference for the Multinomial Logit Model M.Rekkas Abstract Higher order asymptotic theory is used to derive p-values that achieve superior accuracy compared to the p-values obtained from traditional

More information

Likelihood and Asymptotic Theory for Statistical Inference

Likelihood and Asymptotic Theory for Statistical Inference Likelihood and Asymptotic Theory for Statistical Inference Nancy Reid 020 7679 1863 reid@utstat.utoronto.ca n.reid@ucl.ac.uk http://www.utstat.toronto.edu/reid/ltccf12.html LTCC Likelihood Theory Week

More information

Likelihood and Asymptotic Theory for Statistical Inference

Likelihood and Asymptotic Theory for Statistical Inference Likelihood and Asymptotic Theory for Statistical Inference Nancy Reid 020 7679 1863 reid@utstat.utoronto.ca n.reid@ucl.ac.uk http://www.utstat.toronto.edu/reid/ltccf12.html LTCC Likelihood Theory Week

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Improved Inference for First Order Autocorrelation using Likelihood Analysis

Improved Inference for First Order Autocorrelation using Likelihood Analysis Improved Inference for First Order Autocorrelation using Likelihood Analysis M. Rekkas Y. Sun A. Wong Abstract Testing for first-order autocorrelation in small samples using the standard asymptotic test

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

COMBINING p-values: A DEFINITIVE PROCESS. Galley

COMBINING p-values: A DEFINITIVE PROCESS. Galley 0 Journal of Statistical Research ISSN 0 - X 00, Vol., No., pp. - Bangladesh COMBINING p-values: A DEFINITIVE PROCESS D.A.S. Fraser Department of Statistics, University of Toronto, Toronto, Canada MS G

More information

ANCILLARY STATISTICS: A REVIEW

ANCILLARY STATISTICS: A REVIEW 1 ANCILLARY STATISTICS: A REVIEW M. Ghosh, N. Reid and D.A.S. Fraser University of Florida and University of Toronto Abstract: In a parametric statistical model, a function of the data is said to be ancillary

More information

ANCILLARY STATISTICS: A REVIEW

ANCILLARY STATISTICS: A REVIEW Statistica Sinica 20 (2010), 1309-1332 ANCILLARY STATISTICS: A REVIEW M. Ghosh 1, N. Reid 2 and D. A. S. Fraser 2 1 University of Florida and 2 University of Toronto Abstract: In a parametric statistical

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

Marginal Posterior Simulation via Higher-order Tail Area Approximations

Marginal Posterior Simulation via Higher-order Tail Area Approximations Bayesian Analysis (2014) 9, Number 1, pp. 129 146 Marginal Posterior Simulation via Higher-order Tail Area Approximations Erlis Ruli, Nicola Sartori and Laura Ventura Abstract. A new method for posterior

More information

Measuring nuisance parameter effects in Bayesian inference

Measuring nuisance parameter effects in Bayesian inference Measuring nuisance parameter effects in Bayesian inference Alastair Young Imperial College London WHOA-PSI-2017 1 / 31 Acknowledgements: Tom DiCiccio, Cornell University; Daniel Garcia Rasines, Imperial

More information

Third-order inference for autocorrelation in nonlinear regression models

Third-order inference for autocorrelation in nonlinear regression models Third-order inference for autocorrelation in nonlinear regression models P. E. Nguimkeu M. Rekkas Abstract We propose third-order likelihood-based methods to derive highly accurate p-value approximations

More information

CONVERTING OBSERVED LIKELIHOOD FUNCTIONS TO TAIL PROBABILITIES. D.A.S. Fraser Mathematics Department York University North York, Ontario M3J 1P3

CONVERTING OBSERVED LIKELIHOOD FUNCTIONS TO TAIL PROBABILITIES. D.A.S. Fraser Mathematics Department York University North York, Ontario M3J 1P3 CONVERTING OBSERVED LIKELIHOOD FUNCTIONS TO TAIL PROBABILITIES D.A.S. Fraser Mathematics Department York University North York, Ontario M3J 1P3 N. Reid Department of Statistics University of Toronto Toronto,

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

Likelihood based Statistical Inference. Dottorato in Economia e Finanza Dipartimento di Scienze Economiche Univ. di Verona

Likelihood based Statistical Inference. Dottorato in Economia e Finanza Dipartimento di Scienze Economiche Univ. di Verona Likelihood based Statistical Inference Dottorato in Economia e Finanza Dipartimento di Scienze Economiche Univ. di Verona L. Pace, A. Salvan, N. Sartori Udine, April 2008 Likelihood: observed quantities,

More information

hoa: An R Package Bundle for Higher Order Likelihood Inference

hoa: An R Package Bundle for Higher Order Likelihood Inference hoa: An R Package Bundle for Higher Order Likelihood Inference by Alessandra R. Brazzale Rnews, 5/1 May 2005, pp. 20 27 Introduction The likelihood function represents the basic ingredient of many commonly

More information

Default priors for Bayesian and frequentist inference

Default priors for Bayesian and frequentist inference Default priors for Bayesian and frequentist inference D.A.S. Fraser and N. Reid University of Toronto, Canada E. Marras Centre for Advanced Studies and Development, Sardinia University of Rome La Sapienza,

More information

An Improved Specification Test for AR(1) versus MA(1) Disturbances in Linear Regression Models

An Improved Specification Test for AR(1) versus MA(1) Disturbances in Linear Regression Models An Improved Specification Test for AR(1) versus MA(1) Disturbances in Linear Regression Models Pierre Nguimkeu Georgia State University Abstract This paper proposes an improved likelihood-based method

More information

Advanced Quantitative Methods: maximum likelihood

Advanced Quantitative Methods: maximum likelihood Advanced Quantitative Methods: Maximum Likelihood University College Dublin 4 March 2014 1 2 3 4 5 6 Outline 1 2 3 4 5 6 of straight lines y = 1 2 x + 2 dy dx = 1 2 of curves y = x 2 4x + 5 of curves y

More information

Improved Inference for Moving Average Disturbances in Nonlinear Regression Models

Improved Inference for Moving Average Disturbances in Nonlinear Regression Models Improved Inference for Moving Average Disturbances in Nonlinear Regression Models Pierre Nguimkeu Georgia State University November 22, 2013 Abstract This paper proposes an improved likelihood-based method

More information

Conditional Inference by Estimation of a Marginal Distribution

Conditional Inference by Estimation of a Marginal Distribution Conditional Inference by Estimation of a Marginal Distribution Thomas J. DiCiccio and G. Alastair Young 1 Introduction Conditional inference has been, since the seminal work of Fisher (1934), a fundamental

More information

Various types of likelihood

Various types of likelihood Various types of likelihood 1. likelihood, marginal likelihood, conditional likelihood, profile likelihood, adjusted profile likelihood 2. semi-parametric likelihood, partial likelihood 3. empirical likelihood,

More information

Aspects of Likelihood Inference

Aspects of Likelihood Inference Submitted to the Bernoulli Aspects of Likelihood Inference NANCY REID 1 1 Department of Statistics University of Toronto 100 St. George St. Toronto, Canada M5S 3G3 E-mail: reid@utstat.utoronto.ca, url:

More information

Integrated likelihoods in models with stratum nuisance parameters

Integrated likelihoods in models with stratum nuisance parameters Riccardo De Bin, Nicola Sartori, Thomas A. Severini Integrated likelihoods in models with stratum nuisance parameters Technical Report Number 157, 2014 Department of Statistics University of Munich http://www.stat.uni-muenchen.de

More information

Likelihood and p-value functions in the composite likelihood context

Likelihood and p-value functions in the composite likelihood context Likelihood and p-value functions in the composite likelihood context D.A.S. Fraser and N. Reid Department of Statistical Sciences University of Toronto November 19, 2016 Abstract The need for combining

More information

Outline of GLMs. Definitions

Outline of GLMs. Definitions Outline of GLMs Definitions This is a short outline of GLM details, adapted from the book Nonparametric Regression and Generalized Linear Models, by Green and Silverman. The responses Y i have density

More information

Bootstrap and Parametric Inference: Successes and Challenges

Bootstrap and Parametric Inference: Successes and Challenges Bootstrap and Parametric Inference: Successes and Challenges G. Alastair Young Department of Mathematics Imperial College London Newton Institute, January 2008 Overview Overview Review key aspects of frequentist

More information

Modern likelihood inference for the parameter of skewness: An application to monozygotic

Modern likelihood inference for the parameter of skewness: An application to monozygotic Working Paper Series, N. 10, December 2013 Modern likelihood inference for the parameter of skewness: An application to monozygotic twin studies Mameli Valentina Department of Mathematics and Computer

More information

Integrated likelihoods in survival models for highlystratified

Integrated likelihoods in survival models for highlystratified Working Paper Series, N. 1, January 2014 Integrated likelihoods in survival models for highlystratified censored data Giuliana Cortese Department of Statistical Sciences University of Padua Italy Nicola

More information

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Theory of Maximum Likelihood Estimation. Konstantin Kashin Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical

More information

PARAMETER CURVATURE REVISITED AND THE BAYES-FREQUENTIST DIVERGENCE.

PARAMETER CURVATURE REVISITED AND THE BAYES-FREQUENTIST DIVERGENCE. Journal of Statistical Research 200x, Vol. xx, No. xx, pp. xx-xx Bangladesh ISSN 0256-422 X PARAMETER CURVATURE REVISITED AND THE BAYES-FREQUENTIST DIVERGENCE. A.M. FRASER Department of Mathematics, University

More information

Stat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota

Stat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota Stat 5102 Lecture Slides Deck 3 Charles J. Geyer School of Statistics University of Minnesota 1 Likelihood Inference We have learned one very general method of estimation: method of moments. the Now we

More information

Principles of Statistical Inference

Principles of Statistical Inference Principles of Statistical Inference Nancy Reid and David Cox August 30, 2013 Introduction Statistics needs a healthy interplay between theory and applications theory meaning Foundations, rather than theoretical

More information

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability

More information

Principles of Statistical Inference

Principles of Statistical Inference Principles of Statistical Inference Nancy Reid and David Cox August 30, 2013 Introduction Statistics needs a healthy interplay between theory and applications theory meaning Foundations, rather than theoretical

More information

Recap. Vector observation: Y f (y; θ), Y Y R m, θ R d. sample of independent vectors y 1,..., y n. pairwise log-likelihood n m. weights are often 1

Recap. Vector observation: Y f (y; θ), Y Y R m, θ R d. sample of independent vectors y 1,..., y n. pairwise log-likelihood n m. weights are often 1 Recap Vector observation: Y f (y; θ), Y Y R m, θ R d sample of independent vectors y 1,..., y n pairwise log-likelihood n m i=1 r=1 s>r w rs log f 2 (y ir, y is ; θ) weights are often 1 more generally,

More information

Marginal posterior simulation via higher-order tail area approximations

Marginal posterior simulation via higher-order tail area approximations Bayesian Analysis (2004) 1, Number 1, pp. 1 13 Marginal posterior simulation via higher-order tail area approximations Erlis Ruli, Nicola Sartori and Laura Ventura Abstract. A new method for posterior

More information

Research Article Inference for the Sharpe Ratio Using a Likelihood-Based Approach

Research Article Inference for the Sharpe Ratio Using a Likelihood-Based Approach Journal of Probability and Statistics Volume 202 Article ID 87856 24 pages doi:0.55/202/87856 Research Article Inference for the Sharpe Ratio Using a Likelihood-Based Approach Ying Liu Marie Rekkas 2 and

More information

Flat and multimodal likelihoods and model lack of fit in curved exponential families

Flat and multimodal likelihoods and model lack of fit in curved exponential families Mathematical Statistics Stockholm University Flat and multimodal likelihoods and model lack of fit in curved exponential families Rolf Sundberg Research Report 2009:1 ISSN 1650-0377 Postal address: Mathematical

More information

F & B Approaches to a simple model

F & B Approaches to a simple model A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys

More information

Generalized Linear Models Introduction

Generalized Linear Models Introduction Generalized Linear Models Introduction Statistics 135 Autumn 2005 Copyright c 2005 by Mark E. Irwin Generalized Linear Models For many problems, standard linear regression approaches don t work. Sometimes,

More information

More on nuisance parameters

More on nuisance parameters BS2 Statistical Inference, Lecture 3, Hilary Term 2009 January 30, 2009 Suppose that there is a minimal sufficient statistic T = t(x ) partitioned as T = (S, C) = (s(x ), c(x )) where: C1: the distribution

More information

Linear Models A linear model is defined by the expression

Linear Models A linear model is defined by the expression Linear Models A linear model is defined by the expression x = F β + ɛ. where x = (x 1, x 2,..., x n ) is vector of size n usually known as the response vector. β = (β 1, β 2,..., β p ) is the transpose

More information

The formal relationship between analytic and bootstrap approaches to parametric inference

The formal relationship between analytic and bootstrap approaches to parametric inference The formal relationship between analytic and bootstrap approaches to parametric inference T.J. DiCiccio Cornell University, Ithaca, NY 14853, U.S.A. T.A. Kuffner Washington University in St. Louis, St.

More information

Nuisance parameters and their treatment

Nuisance parameters and their treatment BS2 Statistical Inference, Lecture 2, Hilary Term 2008 April 2, 2008 Ancillarity Inference principles Completeness A statistic A = a(x ) is said to be ancillary if (i) The distribution of A does not depend

More information

Sample size determination for logistic regression: A simulation study

Sample size determination for logistic regression: A simulation study Sample size determination for logistic regression: A simulation study Stephen Bush School of Mathematical Sciences, University of Technology Sydney, PO Box 123 Broadway NSW 2007, Australia Abstract This

More information

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ ) Setting RHS to be zero, 0= (θ )+ 2 L(θ ) (θ θ ), θ θ = 2 L(θ ) 1 (θ )= H θθ (θ ) 1 d θ (θ ) O =0 θ 1 θ 3 θ 2 θ Figure 1: The Newton-Raphson Algorithm where H is the Hessian matrix, d θ is the derivative

More information

Statistics and econometrics

Statistics and econometrics 1 / 36 Slides for the course Statistics and econometrics Part 10: Asymptotic hypothesis testing European University Institute Andrea Ichino September 8, 2014 2 / 36 Outline Why do we need large sample

More information

Parametric fractional imputation for missing data analysis

Parametric fractional imputation for missing data analysis 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 Biometrika (????),??,?, pp. 1 15 C???? Biometrika Trust Printed in

More information

For more information about how to cite these materials visit

For more information about how to cite these materials visit Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/

More information

REGRESSION WITH SPATIALLY MISALIGNED DATA. Lisa Madsen Oregon State University David Ruppert Cornell University

REGRESSION WITH SPATIALLY MISALIGNED DATA. Lisa Madsen Oregon State University David Ruppert Cornell University REGRESSION ITH SPATIALL MISALIGNED DATA Lisa Madsen Oregon State University David Ruppert Cornell University SPATIALL MISALIGNED DATA 10 X X X X X X X X 5 X X X X X 0 X 0 5 10 OUTLINE 1. Introduction 2.

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

Model comparison and selection

Model comparison and selection BS2 Statistical Inference, Lectures 9 and 10, Hilary Term 2008 March 2, 2008 Hypothesis testing Consider two alternative models M 1 = {f (x; θ), θ Θ 1 } and M 2 = {f (x; θ), θ Θ 2 } for a sample (X = x)

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information

Linear Models and Estimation by Least Squares

Linear Models and Estimation by Least Squares Linear Models and Estimation by Least Squares Jin-Lung Lin 1 Introduction Causal relation investigation lies in the heart of economics. Effect (Dependent variable) cause (Independent variable) Example:

More information

Topic 12 Overview of Estimation

Topic 12 Overview of Estimation Topic 12 Overview of Estimation Classical Statistics 1 / 9 Outline Introduction Parameter Estimation Classical Statistics Densities and Likelihoods 2 / 9 Introduction In the simplest possible terms, the

More information

Advanced Quantitative Methods: maximum likelihood

Advanced Quantitative Methods: maximum likelihood Advanced Quantitative Methods: Maximum Likelihood University College Dublin March 23, 2011 1 Introduction 2 3 4 5 Outline Introduction 1 Introduction 2 3 4 5 Preliminaries Introduction Ordinary least squares

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

1. Fisher Information

1. Fisher Information 1. Fisher Information Let f(x θ) be a density function with the property that log f(x θ) is differentiable in θ throughout the open p-dimensional parameter set Θ R p ; then the score statistic (or score

More information

JEREMY TAYLOR S CONTRIBUTIONS TO TRANSFORMATION MODEL

JEREMY TAYLOR S CONTRIBUTIONS TO TRANSFORMATION MODEL 1 / 25 JEREMY TAYLOR S CONTRIBUTIONS TO TRANSFORMATION MODELS DEPT. OF STATISTICS, UNIV. WISCONSIN, MADISON BIOMEDICAL STATISTICAL MODELING. CELEBRATION OF JEREMY TAYLOR S OF 60TH BIRTHDAY. UNIVERSITY

More information

Physics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester Physics 403 Propagation of Uncertainties Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Maximum Likelihood and Minimum Least Squares Uncertainty Intervals

More information

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j Standard Errors & Confidence Intervals β β asy N(0, I( β) 1 ), where I( β) = [ 2 l(β, φ; y) ] β i β β= β j We can obtain asymptotic 100(1 α)% confidence intervals for β j using: β j ± Z 1 α/2 se( β j )

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Loglikelihood and Confidence Intervals

Loglikelihood and Confidence Intervals Stat 504, Lecture 2 1 Loglikelihood and Confidence Intervals The loglikelihood function is defined to be the natural logarithm of the likelihood function, l(θ ; x) = log L(θ ; x). For a variety of reasons,

More information

Single-level Models for Binary Responses

Single-level Models for Binary Responses Single-level Models for Binary Responses Distribution of Binary Data y i response for individual i (i = 1,..., n), coded 0 or 1 Denote by r the number in the sample with y = 1 Mean and variance E(y) =

More information

8. Hypothesis Testing

8. Hypothesis Testing FE661 - Statistical Methods for Financial Engineering 8. Hypothesis Testing Jitkomut Songsiri introduction Wald test likelihood-based tests significance test for linear regression 8-1 Introduction elements

More information

Reparametrization of COM-Poisson Regression Models with Applications in the Analysis of Experimental Count Data

Reparametrization of COM-Poisson Regression Models with Applications in the Analysis of Experimental Count Data Reparametrization of COM-Poisson Regression Models with Applications in the Analysis of Experimental Count Data Eduardo Elias Ribeiro Junior 1 2 Walmes Marques Zeviani 1 Wagner Hugo Bonat 1 Clarice Garcia

More information

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Jae-Kwang Kim 1 Iowa State University June 26, 2013 1 Joint work with Shu Yang Introduction 1 Introduction

More information

Logistic regression. 11 Nov Logistic regression (EPFL) Applied Statistics 11 Nov / 20

Logistic regression. 11 Nov Logistic regression (EPFL) Applied Statistics 11 Nov / 20 Logistic regression 11 Nov 2010 Logistic regression (EPFL) Applied Statistics 11 Nov 2010 1 / 20 Modeling overview Want to capture important features of the relationship between a (set of) variable(s)

More information

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown.

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown. Weighting We have seen that if E(Y) = Xβ and V (Y) = σ 2 G, where G is known, the model can be rewritten as a linear model. This is known as generalized least squares or, if G is diagonal, with trace(g)

More information

The Maximum Likelihood Estimator

The Maximum Likelihood Estimator Chapter 4 he Maximum Likelihood Estimator 4. he Maximum likelihood estimator As illustrated in the exponential family of distributions, discussed above, the maximum likelihood estimator of θ 0 the true

More information

Posterior Simulation via the Signed Root Log-Likelihood Ratio

Posterior Simulation via the Signed Root Log-Likelihood Ratio Bayesian Analysis (2010) 5, Number 4, pp. 787 816 Posterior Simulation via the Signed Root Log-Likelihood Ratio S. A. Kharroubi and T. J. Sweeting Abstract. We explore the use of importance sampling based

More information

Interval Estimation for the Ratio and Difference of Two Lognormal Means

Interval Estimation for the Ratio and Difference of Two Lognormal Means UW Biostatistics Working Paper Series 12-7-2005 Interval Estimation for the Ratio and Difference of Two Lognormal Means Yea-Hung Chen University of Washington, yeahung@u.washington.edu Xiao-Hua Zhou University

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

A Very Brief Summary of Bayesian Inference, and Examples

A Very Brief Summary of Bayesian Inference, and Examples A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X

More information

Chapter 7. Hypothesis Testing

Chapter 7. Hypothesis Testing Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Lecture 3. Hypothesis testing. Goodness of Fit. Model diagnostics GLM (Spring, 2018) Lecture 3 1 / 34 Models Let M(X r ) be a model with design matrix X r (with r columns) r n

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Advanced Methods for Data Analysis (36-402/36-608 Spring 2014 1 Generalized linear models 1.1 Introduction: two regressions So far we ve seen two canonical settings for regression.

More information

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter

More information

Notes on the Multivariate Normal and Related Topics

Notes on the Multivariate Normal and Related Topics Version: July 10, 2013 Notes on the Multivariate Normal and Related Topics Let me refresh your memory about the distinctions between population and sample; parameters and statistics; population distributions

More information

2018 2019 1 9 sei@mistiu-tokyoacjp http://wwwstattu-tokyoacjp/~sei/lec-jhtml 11 552 3 0 1 2 3 4 5 6 7 13 14 33 4 1 4 4 2 1 1 2 2 1 1 12 13 R?boxplot boxplotstats which does the computation?boxplotstats

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

arxiv: v2 [math.st] 13 Apr 2012

arxiv: v2 [math.st] 13 Apr 2012 The Asymptotic Covariance Matrix of the Odds Ratio Parameter Estimator in Semiparametric Log-bilinear Odds Ratio Models Angelika Franke, Gerhard Osius Faculty 3 / Mathematics / Computer Science University

More information

DA Freedman Notes on the MLE Fall 2003

DA Freedman Notes on the MLE Fall 2003 DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar

More information