An Information Criteria for Order-restricted Inference

Size: px
Start display at page:

Download "An Information Criteria for Order-restricted Inference"

Transcription

1 An Information Criteria for Order-restricted Inference Nan Lin a, Tianqing Liu 2,b, and Baoxue Zhang,2,b a Department of Mathematics, Washington University in Saint Louis, Saint Louis, MO 633, U.S.A. b Key Laboratory for Applied Statistics of MOE and School of Mathematics and Statistics, Northeast Normal University, Changchun 324 Jilin Province, P. R. China Abstract: A general information criterion with a general penalty which depends on the size of samples is developed for nested and non-nested models in the context of inequality constraints. The true parameters may be defined by a specified parametric model, or a set of specified estimating functions. When the true parameters are defined by estimating functions, we use the empirical likelihood approach to construct information criterion. The consistency of the relevant detection procedures are also established. A Monte Carlo study indicates that our new criterion is effective, compared to Anraku (999) information criterion, for detecting the configuration of normal means satisfying the simple order restriction. Keywords: AIC; BIC; Empirical likelihood; Changepoint; No-observed-adverse-effect level; Simple order restriction; Umbrella Order restriction; Consistency. Introduction Many types of problems are concerned with identifying meaningful structure in real world situations. Structure involving orderings and inequalities is often useful since it is easy to interpret, understand, and explain. When experiments conditions have an inherent ordering, making use of ordering information can improve inference. In bioassays of dose-effect experiments, the no-observed-adverse-effect level L is used to decide the highest dose level at which the adverse effect is acceptably small. In order to determine L, Kikuchi, Yanagawa & Nishiyama (993) and Yanagawa, Kikuchi & Brown (994) suggested several multiple comparison procedures, and also derived bxzhang@nenu.edu.cn 2 The authors are partly supported by NSF of China (No.8737).

2 a method based on Akaike s information criterion, AIC (Akaike, 973). In Kikuchi et al. (993), independent random samples corresponding to k dose levels are supposed to come from k normal populations with an unknown common variance σ 2 and means θ θ 2 θ k. In their paper, L may be defined as a changepoint of mean parameters θ m ( m k) such that θ = = θ m < θ m + θ k. Kikuchi et al. (993) and Yanagawa et al. (994) also examined the probabilities of correct decision of the AIC method and other testing procedures by simulation. However, as far as we know, in classical statistics, hardly any model selection procedure is available for nested or non-nested models in the context of inequality constraints. The problem is that most model selection criterion such AIC and BIC have a penalty for the number of parameters, and in the inequality constrained models, this number is not clear. Recently, Anraku (999) proposed a better theoretical information criterion for parameters under a simple order restriction by employing a more favorable penalty number. Peng et al. (22) proposed an information criterion for detecting the multiplicity of the largest parameter as well as the configuration of true parameters with simple order restriction. They also established strong consistency of their procedures. In this paper, a general information criterion with a general penalty which depends on the size of samples is developed for nested or non-nested models in the context of inequality constraints. The true parameters may be defined by a specified parametric model, or a set of specified estimating functions. When the true parameters are defined by estimating functions, we use the empirical likelihood approach to construct information criterion. The consistency of the relevant detection procedures are also established. 2 An Information Criteria for Order-restricted Inference Suppose we have independent random samples from each of k populations specified by scalar-valued, unknown parameters θ,, θ k satisfying some unknown order restriction. Our concern is to seek distinct parameters among θ,, θ k based on the data. The following definition help us in constructing information criterion. Definition: Let M(θ) be a saturated model with parameter θ Θ R k, where Θ is the parameter space. Let Θ = {θ : θ = = θ k } and Θ t be a measurable set such 2

3 that Θ Θ t Θ. Furthermore, let M(Θ t ) =: M(θ), θ Θ t, which means M(Θ t ) is a sub-model with parameter space Θ t. Define model family as follows: F = {M(Θ t ) : Θ Θ Θ 2 Θ T } where T is finite. It is easily seen that M(Θ t ) is nested in M(Θ) for all t and when t < t 2, M(Θ t ) is nested in M(Θ t2 ). For example, the classic information criteria consider model M(θ), θ = (θ,, θ T ) τ, θ R T with (θ,, θ t ) τ Θ t = R t, that is (θ t+,, θ T ) τ are fixed, it is easy to have F = {M(Θ t )} T t= is a model family. Under order restriction, Let M : θ = θ 2 = θ 3, M 2 : θ = θ 2 θ 3, M 3 : θ θ 2 θ 3, M 4 : θ θ 2 θ 3, where means the independence of two parameters. Then, F = {M t } 4 t= is a model family. Suppose we have s model families F s = {M(Θ st ) : Θ Θ s Θ s2 Θ sts }, s =, 2,, s. (2.) For convenience, let M(s, t) denote the model M(Θ st ). information criteria as follows: In this paper, we consider c n (s, t) = 2 log(l n (x, s, t))/n + p(s, t) ϕ(n)/n, (2.2) where p(s, t) is the dimension of model M(s, t). L n (x, s, t) is the maximum likelihood based on the model M(s, t), the model is selected that corresponds to (ŝ, ˆt) = arg min c n (s, t). (2.3) s s,t t s Theorem 2. Let s and t s be some finite positive integers. Suppose model families {F s } s s= subject to F i Fj =, i j. Let (s, t ) be the correct model family and the correct number of parameters with s s and t t s, which means model M(s, t ) is the correct model. Assume that < p(s, t) < p for s =,, s, t =,, t, where p is a finite constant, and for fixed s, p(s, t ) < p(s, t 2 ) if t < t 2. Furthermore, assume F s is the largest model family that contains true model M(s, t ). Let L n (s, t) be the maximum likelihood based on model M(s, t). For t > t and M(s, t) F s, assume that 2[log(L n (s, t)) log(l n (s, t ))] d W (s, t, t) (2.4) where d denotes the convergence in distribution, W (s, t, t) is a non-degenerate distribution. Furthermore, we assume Then, lim ϕ(n)/n =, lim ϕ(n) =. (2.5) lim P [ŝ = s, ˆt = t ] =. (2.6) 3

4 3 Order-restricted Information Criteria Based on Parametric Likelihood In this section, suppose we specify a probability model p(x θ) for data x. Let L n (x, θ) be the likelihood under the model M(θ), where θ Θ. Denote ˆθ λ as the MLE of θ under model M(Θ λ ) that is ˆθ λ (x) = arg sup θ Θ λ L n (x, θ), (3.) where Θ = {θ : θ = = θ k }. Furthermore, let { A n (y, λ) = 2 log(l n (y, ˆθ λ (y))) log(l n (y, ˆθ } (x))) (3.2) Then, the dimension for the model M(Θ λ ) is defined as p w (λ) = A n (y, λ)p(y ˆθ (x))dy (3.3) It is easy to verify that the model dimension p w (λ) satisfy the conditions of Theorem 2.. Then, the information criteria (2.2) reduces to ORIC w (λ) = 2l(λ) + p w (λ) ϕ(n), (3.4) where l(λ) is the maximum log-likelihood under the λ-th candidate model. According to Theorem 2., (3.4) is a consistent criterion. Let ν (λ) and ν 2 (λ) are the number of and {, } specified in the λ th candidate model. Then, we define the model dimension as follows p(λ) = ν (λ) + ν 2 (λ) i= + i. (3.5) It is easy to verify that the model dimension p(λ) satisfy the conditions of Theorem 2.. Then, the information criteria (2.2) reduces to ORIC(λ) = 2l(λ) + p(λ) ϕ(n), (3.6) where l(λ) is the maximum log-likelihood under the λ-th candidate model. According to Theorem 2., (3.6) is a consistent criterion. Under order restrictions, we choose ϕ (n) = 2 ϕ 2 (n) = a log(log(n)) ϕ 3 (n) = b log(n) ϕ 4 (n) = c n where, a, b and c is constants. Generally, we can take a = 2/ log(log(k)), b = 2/ log(k), and c = 2/ k. 4

5 4 Order-restricted Information Criteria Based on Empirical Likelihood Suppose we have k independent random samples. For the ith sample, let x i,, x ini be i.i.d. observations from a d i -variate distribution F i, that there is a -dimensional parameter θ i associated with F i and information about θ i and F i is available in the form of Eg i (x i, θ i ) =. Let n = k i= n i, then the log-empirical likelihood is given by l n (F,, F k ) = k n i log p ij (4.) i= j= Obviously, we should maximize l(f,, F k ) under constraints given by n i j= p ij = n i j= p ij g i (x ij, θ i ) = for i =,, k. Then, it is easy to have p ij (θ i ) = n i { + λ τ i g i(x ij, θ i )} and for fixed θ i, λ i is a Lagrange multiplier satisfying n i j= p ij (θ i )g i (x ij, θ i ) = for i =,, k. The corresponding profile log-empirical likelihood is given by l E (θ) = k n i log{p ij (θ i )} (4.2) i= j= where θ = (θ τ,, θ τ k )τ. (MELE) of θ under model M(Θ λ ) that is Denote ˆθ λ as the maximum empirical likelihood estimate ˆθ λ = arg sup θ Θ λ l E (θ). (4.3) Then, we define the empirical-likelihood-based information criteria as ORIC EL (λ) = 2l E (ˆθ λ ) + p(λ) ϕ(n), (4.4) where p(λ) is defined in (3.5). According to Theorem 2., (4.4) is a consistent criterion. 5

6 5 A simulation study To examine the performance of the order-restricted information criterion, we executed Monte Carlo simulations for k = 4. We generated normal variates with means µ, µ 2, µ 3, µ 4 and variance. The following configurations of parameters were selected for µ = (µ, µ 2, µ 3, µ 4 ): H : µ = µ2 = µ3 = µ4, µ = (,,, ); K : µ = µ2 = µ3 µ4, K 2 : µ = µ2 µ3 = µ4, K 3 : µ µ2 = µ3 = µ4, K 4 : µ = µ2 µ3 µ4, µ = (,,,.5); µ = (,,.5,.5); µ = (,.5,.5,.5); µ = (,,.5, ); K 5 : µ µ2 = µ3 µ4, µ = (,.5,.5, ); K 6 : µ µ2 µ3 = µ4, µ = (,.5,, ); K 7 : µ µ2 µ3 µ4, µ = (,.5,,.5); M : µ = µ2 = µ3 µ4, µ = (,,,.5); M 2 : µ = µ2 µ3 = µ4, µ = (,,.5,.5); M 3 : µ µ2 = µ3 = µ4, µ = (,.5,.5,.5); M 4 : µ = µ2 µ3 µ4, µ = (,,.5, ); M 5 : µ µ2 = µ3 µ4, µ = (,.5,.5, ); M 6 : µ µ2 µ3 = µ4, µ = (,.5,, ); M 7 : µ µ2 µ3 µ4, µ = (,.5,,.5). We examined two types of comparison. First we compared the performance of our order-restricted criterion with that of Anraku(999) for detecting the true configuration of the parameters. In this simulation study, models H, K, K 2,, K 7, M, M 2,, M 7 are jointly considered. Here we assume the variance is known to be. First, the penalty numbers corresponding to H, K, K 2,, K 7, M, M 2,, M 7 are given by (3.3) for Figure and Figure 2. Second, the penalty numbers corresponding to H, K, K 2,, K 7, M, M 2,, M 7 are given by (3.5) for Figure 3 and Figure 4. Furthermore, in order to provide some clues for how to select the ϕ(n) under moderate-size sample, or even small sample, we choose ϕ(n) = (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n for our three case. The sample sizes for each population are taken equally, ranging form to 2 with step. Figure and Figure 2 show the frequency of error detection by these four methods based on simulations for true models H, K, K 2,, K 7, M, M 2,, M 7, respectively. We observe that our order information criterion work much better than Anraku(999) s order information criterion when ϕ(n) = (2/ log(log(4))) log(log(n)) or ϕ(n) = (2/ log(4)) log(n). And generally, under the restriction that lim ϕ(n)/n = 6

7 , frequencies of error detection for criteria with ϕ(n) increasing to more rapidly converge to more quickly than those with slower ϕ(n). But the slower ϕ(n) is more robust than the faster ϕ(n), see K 6 and K 7 of Figure 2. When we used the penalty numbers (3.5), from Figure 3 and Figure 4, we can draw the same conclusions. Secondly we compared four procedures for detecting the first changepoint (see (6) of Anraku(999)). Models H, M, M 2,, M 7 are considered. Here we assumed the variances of normal variates are common and unknown, and the penalty numbers are given by (3.3). And ϕ(n) is choose as above. The sample sizes for each population are taken equally, ranging form to 2 with step. Figure 5 shows the frequency of error detection by these four methods based on simulations for true models H, M, M 2,, M 7, respectively. We also observe that our order information criterion work much better than Anraku(999) s order information criterion when ϕ(n) = (2/ log(log(4))) log(log(n)) or ϕ(n) = (2/ log(4)) log(n). And the slower ϕ(n) is more robust than the faster ϕ(n), see M 3, M 6 and M 7 of Figure 5. For empirical likelihood inference, we generated X N(µ, ), X 2 N(µ 2, 2), X 3 t(6)+µ 3, X 4 χ 2 () +µ 4 Figure 6 and Figure 7 shows the frequency of error detection by these four methods based on simulations for true models H, K, K 2,, K 7, M, M 2,, M 7, respectively. We observe that our order information criterion works well when ϕ(n) = (2/ log(log(4))) log(log(n)) or ϕ(n) = (2/ log(4)) log(n). And generally, under the restriction that lim ϕ(n)/n =, frequencies of error detection for criteria with ϕ(n) increasing to more rapidly converge to more quickly than those with slower ϕ(n). But the slower ϕ(n) is more robust than the faster ϕ(n), see K 4, K 5, K 6 and K 7 of Figure 6 and Figure 7. 6 Comments and Conclusions As is well known, the calculation of P (i, k, w) may be difficult except for special weights under various order restrictions. Our information criterion avoids this computation problem. 7 Appendix Proof of Theorem 2.. When s s, t t s then the model M(s, t) with p(s, t) parameters is misspecified, so that plim log(l n (s, t))/n < plim log(l n (s, t ))/n, (A.) 7

8 where plim denotes convergence in probability. lim ϕ(n)/n = that Hence, it follows from (A.) and lim n(s, t ) c n (s, t)] = lim P [ 2 log(l n (s, t )) + p(s, t )ϕ(n)/n 2 log(l n (s, t)) + p(s, t)ϕ(n)/n] = lim P [log(l n (s, t ))/n log(l n (s, t))/n.5(p(s, t ) p(s, t))ϕ(n)/n] lim P [log(l n (s, t ))/n log(l n (s, t))/n pϕ(n)/n] = (A.2) so that = lim P [ˆt t m, ŝ s] lim P [c n (s, t ) c n (s, t) for some t t s, s s ] lim P [c n (s, t ) c n (s, t)] s s t<t (A.3) When s = s, if t < t then the model with p(s, t) parameters is misspecified, so that plim log(l n (s, t))/n < plim log(l n (s, t ))/n (A.4) Hence, it follows from (A.4) and lim ϕ(n)/n = that lim n(s, t ) c n (s, t)] = lim P [ 2 log(l n (s, t ))/n + p(s, t )ϕ(n)/n 2 log(l n (s, t))/n + p(s, t)ϕ(n)/n] = lim P [log(l n (s, t ))/n log(l n (s, t))/n.5(p(s, t ) p(s, t))ϕ(n)/n] = (A.5) so that lim P [ˆt < t, ŝ = s ] = lim P [ˆt < t ŝ = s ]P [ŝ = s ] lim P [c n (s, t ) c n (s, t) for some t < t ŝ = s ]P [ŝ = s ] t<t lim P [c n (s, t ) c n (s, t) ŝ = s ]P [ŝ = s ] = (A.6) For t > t it is follows form the assumption that 2[log(L n (s, t)) log(l n (s, t ))] d W (A.7) 8

9 where d denotes the convergence in distribution. Therefore, by lim ϕ(n) = and (A.7) implies that plim 2[log(L n (s, t)) log(l n (s, t ))]/ϕ(n) = hence plim n(c n (s, t ) c n (s, t))/ϕ(n) = plim 2[log(L n (s, t)) log(l n (s, t ))]/ϕ(n) + p(s, t ) p(s, t) = p(s, t ) p(s, t) < so that This implies that lim P [c n(s, t ) c n (s, t))] = lim P [ˆt > t, ŝ = s ] = lim P [ˆt > t ŝ = s ]P [ŝ = s ] =. Thus, we have References lim P [ŝ = s, ˆt = t ] = (A.8) [] Anraku, K. (999), An information criterion for parameters under a simple order restriction. Biometrika, 86:4-52. [2] Barlow, R.E., Bartholomew, D.J., Bremner, J.M. and Brunk, H.D. (972), Statistical Inference under Order Restrictions. New York: Wiley. [3] Barmi H.E. (996), Empirical likelihood ratio test for or against a set of inequality constraints Journal of Statistical Planning and Inference, 55:9-24. [4] Bartholomew, D.J. (959), A test of homogeneity for ordered alternatives. Biometrika, 46: [5] Bohrer, R. and Chow, W. (978), Weights for one-sided multivariate inference. Appl. Statist, 27:-4. [6] Childs, D.R. (967), Reduction of the multivariate normal integral to characteristic form. Biometrika, 54: [7] Konishi, S. and Kitagawa, G. (996), Generalised information criteria in model selection Biometrika, 83:

10 [8] Kullback, S. and Leibler, R.A. (95), On information and sufficiency. Ann.Math. Statist, 22: [9] LEE, C.I.C., Robertson, T. and Wright, F.T. (993), Bounds on distributions arising in order restricted inferences with restricted weights. Biometrika, 8: [] Perlman, M. D. (969), One-sided problems in multivariate analysis. Ann.Math. Statist, 4: [] Qin, J. and J. Lawless (955), empirical likelihood and tests of parameters subject to constraints. Canad. J. Statist, 23: [2] Rao, C.R. (973), Linear Statistical Inference and its Applications, 2nd ed. New York: Wiley. [3] Robertson, T. and Wright, F.T. (982), Bounds on mixtures of distributions arising in order restricted inference. Ann. Statist, : [4] Robertson, T. and Wright, F.T. and Dykstra, R.L. (988), Order Restricted Statistical Inference. New York: Wiley. [5] Williams, D.A. (97), A test for differences between treatment means when several dose levels are compared with a zero dose control Biometrics, 27:3-7. [6] Zhao, L. and Peng, L. (22), Model selection under order restriction. Statistics & Probability Letters, 57:3-36.

11 .7 H.7 K.7 K K 3 K 4 K Figure : The curve denote the curve for Anraku(999) s order information criterion. Those for our restricted criteria with the penalty numbers (3.3), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively.

12 K 6 K 7.8 M M 2 M M M M 4.5 M 7 Figure 2: The curve denote the curve for Anraku(999) s order information criterion.those for our restricted criteria with the penalty numbers (3.3), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 2

13 .7 H.7 K.7 K K 3 K 4 K Figure 3: The curve denote the curve for Anraku(999) s order information criterion. Those for our restricted criteria with the penalty numbers (3.5), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 3

14 K 6 K 7.8 M M 2 M M M M 4.5 M 7 Figure 4: The curve denote the curve for Anraku(999) s order information criterion. Those for our restricted criteria with the penalty numbers (3.5), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 4

15 .4 H.7 M.7 M 2.5 M M 4.5 M 5.4 M 6.4 M Figure 5: The curve denote the curve for Anraku(999) s order information criterion. Those for our restricted criteria with the penalty numbers (3.3), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 5

16 .7 H.8 K.8 K K 3 K 4 K Figure 6: Those for our empirical-likelihood-based information criteria (4.4), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 6

17 K 6 K 7.8 M M 2 M M 3.5 M 6.5 M 4.5 M 7 Figure 7: Those for our empirical-likelihood-based information criteria (4.4), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 7

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction Sankhyā : The Indian Journal of Statistics 2007, Volume 69, Part 4, pp. 700-716 c 2007, Indian Statistical Institute More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order

More information

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30 MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)

More information

A Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when the Covariance Matrices are Unknown but Common

A Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when the Covariance Matrices are Unknown but Common Journal of Statistical Theory and Applications Volume 11, Number 1, 2012, pp. 23-45 ISSN 1538-7887 A Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when

More information

Akaike Information Criterion

Akaike Information Criterion Akaike Information Criterion Shuhua Hu Center for Research in Scientific Computation North Carolina State University Raleigh, NC February 7, 2012-1- background Background Model statistical model: Y j =

More information

Econometrics I, Estimation

Econometrics I, Estimation Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the

More information

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,

More information

Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems

Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems Jeremy S. Conner and Dale E. Seborg Department of Chemical Engineering University of California, Santa Barbara, CA

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsa

1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsa On the Use of Marginal Likelihood in Model Selection Peide Shi Department of Probability and Statistics Peking University, Beijing 100871 P. R. China Chih-Ling Tsai Graduate School of Management University

More information

Model comparison and selection

Model comparison and selection BS2 Statistical Inference, Lectures 9 and 10, Hilary Term 2008 March 2, 2008 Hypothesis testing Consider two alternative models M 1 = {f (x; θ), θ Θ 1 } and M 2 = {f (x; θ), θ Θ 2 } for a sample (X = x)

More information

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)

More information

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data Song Xi CHEN Guanghua School of Management and Center for Statistical Science, Peking University Department

More information

Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition

Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Hirokazu Yanagihara 1, Tetsuji Tonda 2 and Chieko Matsumoto 3 1 Department of Social Systems

More information

The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models

The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population Health

More information

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population

More information

MIT Spring 2016

MIT Spring 2016 MIT 18.655 Dr. Kempthorne Spring 2016 1 MIT 18.655 Outline 1 2 MIT 18.655 Decision Problem: Basic Components P = {P θ : θ Θ} : parametric model. Θ = {θ}: Parameter space. A{a} : Action space. L(θ, a) :

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Lecture 3. Hypothesis testing. Goodness of Fit. Model diagnostics GLM (Spring, 2018) Lecture 3 1 / 34 Models Let M(X r ) be a model with design matrix X r (with r columns) r n

More information

Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models

Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models Junfeng Shang Bowling Green State University, USA Abstract In the mixed modeling framework, Monte Carlo simulation

More information

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM University of Illinois Department of Economics Spring 2017 Econ 574 Roger Koenker Lecture 14 GEE-GMM Throughout the course we have emphasized methods of estimation and inference based on the principle

More information

Likelihood-based inference with missing data under missing-at-random

Likelihood-based inference with missing data under missing-at-random Likelihood-based inference with missing data under missing-at-random Jae-kwang Kim Joint work with Shu Yang Department of Statistics, Iowa State University May 4, 014 Outline 1. Introduction. Parametric

More information

Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis

Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis Tetsuro Sakurai and Yasunori Fujikoshi Center of General Education, Tokyo University

More information

Modification and Improvement of Empirical Likelihood for Missing Response Problem

Modification and Improvement of Empirical Likelihood for Missing Response Problem UW Biostatistics Working Paper Series 12-30-2010 Modification and Improvement of Empirical Likelihood for Missing Response Problem Kwun Chuen Gary Chan University of Washington - Seattle Campus, kcgchan@u.washington.edu

More information

Research Article Optimal Portfolio Estimation for Dependent Financial Returns with Generalized Empirical Likelihood

Research Article Optimal Portfolio Estimation for Dependent Financial Returns with Generalized Empirical Likelihood Advances in Decision Sciences Volume 2012, Article ID 973173, 8 pages doi:10.1155/2012/973173 Research Article Optimal Portfolio Estimation for Dependent Financial Returns with Generalized Empirical Likelihood

More information

DA Freedman Notes on the MLE Fall 2003

DA Freedman Notes on the MLE Fall 2003 DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar

More information

All models are wrong but some are useful. George Box (1979)

All models are wrong but some are useful. George Box (1979) All models are wrong but some are useful. George Box (1979) The problem of model selection is overrun by a serious difficulty: even if a criterion could be settled on to determine optimality, it is hard

More information

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures FE661 - Statistical Methods for Financial Engineering 9. Model Selection Jitkomut Songsiri statistical models overview of model selection information criteria goodness-of-fit measures 9-1 Statistical models

More information

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS020) p.3863 Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Jinfang Wang and

More information

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j Standard Errors & Confidence Intervals β β asy N(0, I( β) 1 ), where I( β) = [ 2 l(β, φ; y) ] β i β β= β j We can obtain asymptotic 100(1 α)% confidence intervals for β j using: β j ± Z 1 α/2 se( β j )

More information

An Akaike Criterion based on Kullback Symmetric Divergence in the Presence of Incomplete-Data

An Akaike Criterion based on Kullback Symmetric Divergence in the Presence of Incomplete-Data An Akaike Criterion based on Kullback Symmetric Divergence Bezza Hafidi a and Abdallah Mkhadri a a University Cadi-Ayyad, Faculty of sciences Semlalia, Department of Mathematics, PB.2390 Marrakech, Moroco

More information

Long-Run Covariability

Long-Run Covariability Long-Run Covariability Ulrich K. Müller and Mark W. Watson Princeton University October 2016 Motivation Study the long-run covariability/relationship between economic variables great ratios, long-run Phillips

More information

Graduate Econometrics I: Maximum Likelihood I

Graduate Econometrics I: Maximum Likelihood I Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY Econometrics Working Paper EWP0401 ISSN 1485-6441 Department of Economics AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY Lauren Bin Dong & David E. A. Giles Department of Economics, University of Victoria

More information

1 Hypothesis Testing and Model Selection

1 Hypothesis Testing and Model Selection A Short Course on Bayesian Inference (based on An Introduction to Bayesian Analysis: Theory and Methods by Ghosh, Delampady and Samanta) Module 6: From Chapter 6 of GDS 1 Hypothesis Testing and Model Selection

More information

7. Estimation and hypothesis testing. Objective. Recommended reading

7. Estimation and hypothesis testing. Objective. Recommended reading 7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing

More information

Conjugate Predictive Distributions and Generalized Entropies

Conjugate Predictive Distributions and Generalized Entropies Conjugate Predictive Distributions and Generalized Entropies Eduardo Gutiérrez-Peña Department of Probability and Statistics IIMAS-UNAM, Mexico Padova, Italy. 21-23 March, 2013 Menu 1 Antipasto/Appetizer

More information

10-704: Information Processing and Learning Fall Lecture 24: Dec 7

10-704: Information Processing and Learning Fall Lecture 24: Dec 7 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 24: Dec 7 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy of

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Yasunori Fujikoshi and Tetsuro Sakurai Department of Mathematics, Graduate School of Science,

More information

Comment about AR spectral estimation Usually an estimate is produced by computing the AR theoretical spectrum at (ˆφ, ˆσ 2 ). With our Monte Carlo

Comment about AR spectral estimation Usually an estimate is produced by computing the AR theoretical spectrum at (ˆφ, ˆσ 2 ). With our Monte Carlo Comment aout AR spectral estimation Usually an estimate is produced y computing the AR theoretical spectrum at (ˆφ, ˆσ 2 ). With our Monte Carlo simulation approach, for every draw (φ,σ 2 ), we can compute

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

Model Fitting, Bootstrap, & Model Selection

Model Fitting, Bootstrap, & Model Selection Model Fitting, Bootstrap, & Model Selection G. Jogesh Babu Penn State University http://www.stat.psu.edu/ babu http://astrostatistics.psu.edu Model Fitting Non-linear regression Density (shape) estimation

More information

COS513 LECTURE 8 STATISTICAL CONCEPTS

COS513 LECTURE 8 STATISTICAL CONCEPTS COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions

More information

An Extended BIC for Model Selection

An Extended BIC for Model Selection An Extended BIC for Model Selection at the JSM meeting 2007 - Salt Lake City Surajit Ray Boston University (Dept of Mathematics and Statistics) Joint work with James Berger, Duke University; Susie Bayarri,

More information

Model selection in penalized Gaussian graphical models

Model selection in penalized Gaussian graphical models University of Groningen e.c.wit@rug.nl http://www.math.rug.nl/ ernst January 2014 Penalized likelihood generates a PATH of solutions Consider an experiment: Γ genes measured across T time points. Assume

More information

Modelling the Covariance

Modelling the Covariance Modelling the Covariance Jamie Monogan Washington University in St Louis February 9, 2010 Jamie Monogan (WUStL) Modelling the Covariance February 9, 2010 1 / 13 Objectives By the end of this meeting, participants

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

over the parameters θ. In both cases, consequently, we select the minimizing

over the parameters θ. In both cases, consequently, we select the minimizing MONOTONE REGRESSION JAN DE LEEUW Abstract. This is an entry for The Encyclopedia of Statistics in Behavioral Science, to be published by Wiley in 200. In linear regression we fit a linear function y =

More information

Statistical Inference

Statistical Inference Statistical Inference Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Week 12. Testing and Kullback-Leibler Divergence 1. Likelihood Ratios Let 1, 2, 2,...

More information

Label Switching and Its Simple Solutions for Frequentist Mixture Models

Label Switching and Its Simple Solutions for Frequentist Mixture Models Label Switching and Its Simple Solutions for Frequentist Mixture Models Weixin Yao Department of Statistics, Kansas State University, Manhattan, Kansas 66506, U.S.A. wxyao@ksu.edu Abstract The label switching

More information

Information Theory Based Estimator of the Number of Sources in a Sparse Linear Mixing Model

Information Theory Based Estimator of the Number of Sources in a Sparse Linear Mixing Model Information heory Based Estimator of the Number of Sources in a Sparse Linear Mixing Model Radu Balan University of Maryland Department of Mathematics, Center for Scientific Computation And Mathematical

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information

COMPARING TRANSFORMATIONS USING TESTS OF SEPARATE FAMILIES. Department of Biostatistics, University of North Carolina at Chapel Hill, NC.

COMPARING TRANSFORMATIONS USING TESTS OF SEPARATE FAMILIES. Department of Biostatistics, University of North Carolina at Chapel Hill, NC. COMPARING TRANSFORMATIONS USING TESTS OF SEPARATE FAMILIES by Lloyd J. Edwards and Ronald W. Helms Department of Biostatistics, University of North Carolina at Chapel Hill, NC. Institute of Statistics

More information

Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests

Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests Biometrika (2014),,, pp. 1 13 C 2014 Biometrika Trust Printed in Great Britain Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests BY M. ZHOU Department of Statistics, University

More information

Semiparametric Gaussian Copula Models: Progress and Problems

Semiparametric Gaussian Copula Models: Progress and Problems Semiparametric Gaussian Copula Models: Progress and Problems Jon A. Wellner University of Washington, Seattle 2015 IMS China, Kunming July 1-4, 2015 2015 IMS China Meeting, Kunming Based on joint work

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 2013-14 We know that X ~ B(n,p), but we do not know p. We get a random sample

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction

More information

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models Thomas Kneib Department of Mathematics Carl von Ossietzky University Oldenburg Sonja Greven Department of

More information

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis https://doi.org/10.1007/s42081-019-00032-4 ORIGINAL PAPER Consistency of test based method for selection of variables in high dimensional two group discriminant analysis Yasunori Fujikoshi 1 Tetsuro Sakurai

More information

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Ann Inst Stat Math (2009) 61:773 787 DOI 10.1007/s10463-008-0172-6 Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Taisuke Otsu Received: 1 June 2007 / Revised:

More information

Akaike criterion: Kullback-Leibler discrepancy

Akaike criterion: Kullback-Leibler discrepancy Model choice. Akaike s criterion Akaike criterion: Kullback-Leibler discrepancy Given a family of probability densities {f ( ; ψ), ψ Ψ}, Kullback-Leibler s index of f ( ; ψ) relative to f ( ; θ) is (ψ

More information

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications. Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable

More information

Empirical likelihood-based methods for the difference of two trimmed means

Empirical likelihood-based methods for the difference of two trimmed means Empirical likelihood-based methods for the difference of two trimmed means 24.09.2012. Latvijas Universitate Contents 1 Introduction 2 Trimmed mean 3 Empirical likelihood 4 Empirical likelihood for the

More information

Bayesian Analysis of Risk for Data Mining Based on Empirical Likelihood

Bayesian Analysis of Risk for Data Mining Based on Empirical Likelihood 1 / 29 Bayesian Analysis of Risk for Data Mining Based on Empirical Likelihood Yuan Liao Wenxin Jiang Northwestern University Presented at: Department of Statistics and Biostatistics Rutgers University

More information

Maximum likelihood: counterexamples, examples, and open problems. Jon A. Wellner. University of Washington. Maximum likelihood: p.

Maximum likelihood: counterexamples, examples, and open problems. Jon A. Wellner. University of Washington. Maximum likelihood: p. Maximum likelihood: counterexamples, examples, and open problems Jon A. Wellner University of Washington Maximum likelihood: p. 1/7 Talk at University of Idaho, Department of Mathematics, September 15,

More information

A MODIFIED LIKELIHOOD RATIO TEST FOR HOMOGENEITY IN FINITE MIXTURE MODELS

A MODIFIED LIKELIHOOD RATIO TEST FOR HOMOGENEITY IN FINITE MIXTURE MODELS A MODIFIED LIKELIHOOD RATIO TEST FOR HOMOGENEITY IN FINITE MIXTURE MODELS Hanfeng Chen, Jiahua Chen and John D. Kalbfleisch Bowling Green State University and University of Waterloo Abstract Testing for

More information

F & B Approaches to a simple model

F & B Approaches to a simple model A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys

More information

Discrepancy-Based Model Selection Criteria Using Cross Validation

Discrepancy-Based Model Selection Criteria Using Cross Validation 33 Discrepancy-Based Model Selection Criteria Using Cross Validation Joseph E. Cavanaugh, Simon L. Davies, and Andrew A. Neath Department of Biostatistics, The University of Iowa Pfizer Global Research

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

Simulation of truncated normal variables. Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris

Simulation of truncated normal variables. Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris Simulation of truncated normal variables Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris Abstract arxiv:0907.4010v1 [stat.co] 23 Jul 2009 We provide in this paper simulation algorithms

More information

Ch. 5 Hypothesis Testing

Ch. 5 Hypothesis Testing Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,

More information

Large Sample Properties of Estimators in the Classical Linear Regression Model

Large Sample Properties of Estimators in the Classical Linear Regression Model Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in

More information

Practice Problems Section Problems

Practice Problems Section Problems Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,

More information

Review and continuation from last week Properties of MLEs

Review and continuation from last week Properties of MLEs Review and continuation from last week Properties of MLEs As we have mentioned, MLEs have a nice intuitive property, and as we have seen, they have a certain equivariance property. We will see later that

More information

Lecture 4 September 15

Lecture 4 September 15 IFT 6269: Probabilistic Graphical Models Fall 2017 Lecture 4 September 15 Lecturer: Simon Lacoste-Julien Scribe: Philippe Brouillard & Tristan Deleu 4.1 Maximum Likelihood principle Given a parametric

More information

11 Survival Analysis and Empirical Likelihood

11 Survival Analysis and Empirical Likelihood 11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with

More information

Estimating GARCH models: when to use what?

Estimating GARCH models: when to use what? Econometrics Journal (2008), volume, pp. 27 38. doi: 0./j.368-423X.2008.00229.x Estimating GARCH models: when to use what? DA HUANG, HANSHENG WANG AND QIWEI YAO, Guanghua School of Management, Peking University,

More information

ICSA Applied Statistics Symposium 1. Balanced adjusted empirical likelihood

ICSA Applied Statistics Symposium 1. Balanced adjusted empirical likelihood ICSA Applied Statistics Symposium 1 Balanced adjusted empirical likelihood Art B. Owen Stanford University Sarah Emerson Oregon State University ICSA Applied Statistics Symposium 2 Empirical likelihood

More information

Information in Data. Sufficiency, Ancillarity, Minimality, and Completeness

Information in Data. Sufficiency, Ancillarity, Minimality, and Completeness Information in Data Sufficiency, Ancillarity, Minimality, and Completeness Important properties of statistics that determine the usefulness of those statistics in statistical inference. These general properties

More information

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section

More information

Quantifying Stochastic Model Errors via Robust Optimization

Quantifying Stochastic Model Errors via Robust Optimization Quantifying Stochastic Model Errors via Robust Optimization IPAM Workshop on Uncertainty Quantification for Multiscale Stochastic Systems and Applications Jan 19, 2016 Henry Lam Industrial & Operations

More information

A Bayesian Multiple Comparison Procedure for Order-Restricted Mixed Models

A Bayesian Multiple Comparison Procedure for Order-Restricted Mixed Models A Bayesian Multiple Comparison Procedure for Order-Restricted Mixed Models Junfeng Shang 1, Joseph E. Cavanaugh 2, Farroll T. Wright 3, 1 Department of Mathematics and Statistics, 450 Math Science Building,

More information

Model Selection in the Presence of Incidental Parameters

Model Selection in the Presence of Incidental Parameters Model Selection in the Presence of Incidental Parameters Yoonseok Lee University of Michigan February 2012 Abstract This paper considers model selection of nonlinear panel data models in the presence of

More information

Investigation of goodness-of-fit test statistic distributions by random censored samples

Investigation of goodness-of-fit test statistic distributions by random censored samples d samples Investigation of goodness-of-fit test statistic distributions by random censored samples Novosibirsk State Technical University November 22, 2010 d samples Outline 1 Nonparametric goodness-of-fit

More information

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk Ann Inst Stat Math (0) 64:359 37 DOI 0.007/s0463-00-036-3 Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk Paul Vos Qiang Wu Received: 3 June 009 / Revised:

More information

Augmented Likelihood Estimators for Mixture Models

Augmented Likelihood Estimators for Mixture Models Augmented Likelihood Estimators for Mixture Models Markus Haas Jochen Krause Marc S. Paolella Swiss Banking Institute, University of Zurich What is mixture degeneracy? mixtures under study are finite convex

More information

A union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling

A union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling A union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling Min-ge Xie Department of Statistics, Rutgers University Workshop on Higher-Order Asymptotics

More information

Introduction to graphical models: Lecture III

Introduction to graphical models: Lecture III Introduction to graphical models: Lecture III Martin Wainwright UC Berkeley Departments of Statistics, and EECS Martin Wainwright (UC Berkeley) Some introductory lectures January 2013 1 / 25 Introduction

More information

7. Estimation and hypothesis testing. Objective. Recommended reading

7. Estimation and hypothesis testing. Objective. Recommended reading 7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing

More information

STAT 535 Lecture 5 November, 2018 Brief overview of Model Selection and Regularization c Marina Meilă

STAT 535 Lecture 5 November, 2018 Brief overview of Model Selection and Regularization c Marina Meilă STAT 535 Lecture 5 November, 2018 Brief overview of Model Selection and Regularization c Marina Meilă mmp@stat.washington.edu Reading: Murphy: BIC, AIC 8.4.2 (pp 255), SRM 6.5 (pp 204) Hastie, Tibshirani

More information

Empirical Likelihood Methods for Sample Survey Data: An Overview

Empirical Likelihood Methods for Sample Survey Data: An Overview AUSTRIAN JOURNAL OF STATISTICS Volume 35 (2006), Number 2&3, 191 196 Empirical Likelihood Methods for Sample Survey Data: An Overview J. N. K. Rao Carleton University, Ottawa, Canada Abstract: The use

More information

Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools. Joan Llull. Microeconometrics IDEA PhD Program

Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools. Joan Llull. Microeconometrics IDEA PhD Program Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools Joan Llull Microeconometrics IDEA PhD Program Maximum Likelihood Chapter 1. A Brief Review of Maximum Likelihood, GMM, and Numerical

More information

Estimation in Generalized Linear Models with Heterogeneous Random Effects. Woncheol Jang Johan Lim. May 19, 2004

Estimation in Generalized Linear Models with Heterogeneous Random Effects. Woncheol Jang Johan Lim. May 19, 2004 Estimation in Generalized Linear Models with Heterogeneous Random Effects Woncheol Jang Johan Lim May 19, 2004 Abstract The penalized quasi-likelihood (PQL) approach is the most common estimation procedure

More information

Mixture Modeling. Identifying the Correct Number of Classes in a Growth Mixture Model. Davood Tofighi Craig Enders Arizona State University

Mixture Modeling. Identifying the Correct Number of Classes in a Growth Mixture Model. Davood Tofighi Craig Enders Arizona State University Identifying the Correct Number of Classes in a Growth Mixture Model Davood Tofighi Craig Enders Arizona State University Mixture Modeling Heterogeneity exists such that the data are comprised of two or

More information

Causal Model Selection Hypothesis Tests in Systems Genetics

Causal Model Selection Hypothesis Tests in Systems Genetics 1 Causal Model Selection Hypothesis Tests in Systems Genetics Elias Chaibub Neto and Brian S Yandell SISG 2012 July 13, 2012 2 Correlation and Causation The old view of cause and effect... could only fail;

More information

Model Selection in the Presence of Incidental Parameters

Model Selection in the Presence of Incidental Parameters Model Selection in the Presence of Incidental Parameters Yoonseok Lee University of Michigan July 2012 Abstract This paper considers model selection of nonlinear panel data models in the presence of incidental

More information

LTI Systems, Additive Noise, and Order Estimation

LTI Systems, Additive Noise, and Order Estimation LTI Systems, Additive oise, and Order Estimation Soosan Beheshti, Munther A. Dahleh Laboratory for Information and Decision Systems Department of Electrical Engineering and Computer Science Massachusetts

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information