An Information Criteria for Order-restricted Inference
|
|
- Regina Greer
- 5 years ago
- Views:
Transcription
1 An Information Criteria for Order-restricted Inference Nan Lin a, Tianqing Liu 2,b, and Baoxue Zhang,2,b a Department of Mathematics, Washington University in Saint Louis, Saint Louis, MO 633, U.S.A. b Key Laboratory for Applied Statistics of MOE and School of Mathematics and Statistics, Northeast Normal University, Changchun 324 Jilin Province, P. R. China Abstract: A general information criterion with a general penalty which depends on the size of samples is developed for nested and non-nested models in the context of inequality constraints. The true parameters may be defined by a specified parametric model, or a set of specified estimating functions. When the true parameters are defined by estimating functions, we use the empirical likelihood approach to construct information criterion. The consistency of the relevant detection procedures are also established. A Monte Carlo study indicates that our new criterion is effective, compared to Anraku (999) information criterion, for detecting the configuration of normal means satisfying the simple order restriction. Keywords: AIC; BIC; Empirical likelihood; Changepoint; No-observed-adverse-effect level; Simple order restriction; Umbrella Order restriction; Consistency. Introduction Many types of problems are concerned with identifying meaningful structure in real world situations. Structure involving orderings and inequalities is often useful since it is easy to interpret, understand, and explain. When experiments conditions have an inherent ordering, making use of ordering information can improve inference. In bioassays of dose-effect experiments, the no-observed-adverse-effect level L is used to decide the highest dose level at which the adverse effect is acceptably small. In order to determine L, Kikuchi, Yanagawa & Nishiyama (993) and Yanagawa, Kikuchi & Brown (994) suggested several multiple comparison procedures, and also derived bxzhang@nenu.edu.cn 2 The authors are partly supported by NSF of China (No.8737).
2 a method based on Akaike s information criterion, AIC (Akaike, 973). In Kikuchi et al. (993), independent random samples corresponding to k dose levels are supposed to come from k normal populations with an unknown common variance σ 2 and means θ θ 2 θ k. In their paper, L may be defined as a changepoint of mean parameters θ m ( m k) such that θ = = θ m < θ m + θ k. Kikuchi et al. (993) and Yanagawa et al. (994) also examined the probabilities of correct decision of the AIC method and other testing procedures by simulation. However, as far as we know, in classical statistics, hardly any model selection procedure is available for nested or non-nested models in the context of inequality constraints. The problem is that most model selection criterion such AIC and BIC have a penalty for the number of parameters, and in the inequality constrained models, this number is not clear. Recently, Anraku (999) proposed a better theoretical information criterion for parameters under a simple order restriction by employing a more favorable penalty number. Peng et al. (22) proposed an information criterion for detecting the multiplicity of the largest parameter as well as the configuration of true parameters with simple order restriction. They also established strong consistency of their procedures. In this paper, a general information criterion with a general penalty which depends on the size of samples is developed for nested or non-nested models in the context of inequality constraints. The true parameters may be defined by a specified parametric model, or a set of specified estimating functions. When the true parameters are defined by estimating functions, we use the empirical likelihood approach to construct information criterion. The consistency of the relevant detection procedures are also established. 2 An Information Criteria for Order-restricted Inference Suppose we have independent random samples from each of k populations specified by scalar-valued, unknown parameters θ,, θ k satisfying some unknown order restriction. Our concern is to seek distinct parameters among θ,, θ k based on the data. The following definition help us in constructing information criterion. Definition: Let M(θ) be a saturated model with parameter θ Θ R k, where Θ is the parameter space. Let Θ = {θ : θ = = θ k } and Θ t be a measurable set such 2
3 that Θ Θ t Θ. Furthermore, let M(Θ t ) =: M(θ), θ Θ t, which means M(Θ t ) is a sub-model with parameter space Θ t. Define model family as follows: F = {M(Θ t ) : Θ Θ Θ 2 Θ T } where T is finite. It is easily seen that M(Θ t ) is nested in M(Θ) for all t and when t < t 2, M(Θ t ) is nested in M(Θ t2 ). For example, the classic information criteria consider model M(θ), θ = (θ,, θ T ) τ, θ R T with (θ,, θ t ) τ Θ t = R t, that is (θ t+,, θ T ) τ are fixed, it is easy to have F = {M(Θ t )} T t= is a model family. Under order restriction, Let M : θ = θ 2 = θ 3, M 2 : θ = θ 2 θ 3, M 3 : θ θ 2 θ 3, M 4 : θ θ 2 θ 3, where means the independence of two parameters. Then, F = {M t } 4 t= is a model family. Suppose we have s model families F s = {M(Θ st ) : Θ Θ s Θ s2 Θ sts }, s =, 2,, s. (2.) For convenience, let M(s, t) denote the model M(Θ st ). information criteria as follows: In this paper, we consider c n (s, t) = 2 log(l n (x, s, t))/n + p(s, t) ϕ(n)/n, (2.2) where p(s, t) is the dimension of model M(s, t). L n (x, s, t) is the maximum likelihood based on the model M(s, t), the model is selected that corresponds to (ŝ, ˆt) = arg min c n (s, t). (2.3) s s,t t s Theorem 2. Let s and t s be some finite positive integers. Suppose model families {F s } s s= subject to F i Fj =, i j. Let (s, t ) be the correct model family and the correct number of parameters with s s and t t s, which means model M(s, t ) is the correct model. Assume that < p(s, t) < p for s =,, s, t =,, t, where p is a finite constant, and for fixed s, p(s, t ) < p(s, t 2 ) if t < t 2. Furthermore, assume F s is the largest model family that contains true model M(s, t ). Let L n (s, t) be the maximum likelihood based on model M(s, t). For t > t and M(s, t) F s, assume that 2[log(L n (s, t)) log(l n (s, t ))] d W (s, t, t) (2.4) where d denotes the convergence in distribution, W (s, t, t) is a non-degenerate distribution. Furthermore, we assume Then, lim ϕ(n)/n =, lim ϕ(n) =. (2.5) lim P [ŝ = s, ˆt = t ] =. (2.6) 3
4 3 Order-restricted Information Criteria Based on Parametric Likelihood In this section, suppose we specify a probability model p(x θ) for data x. Let L n (x, θ) be the likelihood under the model M(θ), where θ Θ. Denote ˆθ λ as the MLE of θ under model M(Θ λ ) that is ˆθ λ (x) = arg sup θ Θ λ L n (x, θ), (3.) where Θ = {θ : θ = = θ k }. Furthermore, let { A n (y, λ) = 2 log(l n (y, ˆθ λ (y))) log(l n (y, ˆθ } (x))) (3.2) Then, the dimension for the model M(Θ λ ) is defined as p w (λ) = A n (y, λ)p(y ˆθ (x))dy (3.3) It is easy to verify that the model dimension p w (λ) satisfy the conditions of Theorem 2.. Then, the information criteria (2.2) reduces to ORIC w (λ) = 2l(λ) + p w (λ) ϕ(n), (3.4) where l(λ) is the maximum log-likelihood under the λ-th candidate model. According to Theorem 2., (3.4) is a consistent criterion. Let ν (λ) and ν 2 (λ) are the number of and {, } specified in the λ th candidate model. Then, we define the model dimension as follows p(λ) = ν (λ) + ν 2 (λ) i= + i. (3.5) It is easy to verify that the model dimension p(λ) satisfy the conditions of Theorem 2.. Then, the information criteria (2.2) reduces to ORIC(λ) = 2l(λ) + p(λ) ϕ(n), (3.6) where l(λ) is the maximum log-likelihood under the λ-th candidate model. According to Theorem 2., (3.6) is a consistent criterion. Under order restrictions, we choose ϕ (n) = 2 ϕ 2 (n) = a log(log(n)) ϕ 3 (n) = b log(n) ϕ 4 (n) = c n where, a, b and c is constants. Generally, we can take a = 2/ log(log(k)), b = 2/ log(k), and c = 2/ k. 4
5 4 Order-restricted Information Criteria Based on Empirical Likelihood Suppose we have k independent random samples. For the ith sample, let x i,, x ini be i.i.d. observations from a d i -variate distribution F i, that there is a -dimensional parameter θ i associated with F i and information about θ i and F i is available in the form of Eg i (x i, θ i ) =. Let n = k i= n i, then the log-empirical likelihood is given by l n (F,, F k ) = k n i log p ij (4.) i= j= Obviously, we should maximize l(f,, F k ) under constraints given by n i j= p ij = n i j= p ij g i (x ij, θ i ) = for i =,, k. Then, it is easy to have p ij (θ i ) = n i { + λ τ i g i(x ij, θ i )} and for fixed θ i, λ i is a Lagrange multiplier satisfying n i j= p ij (θ i )g i (x ij, θ i ) = for i =,, k. The corresponding profile log-empirical likelihood is given by l E (θ) = k n i log{p ij (θ i )} (4.2) i= j= where θ = (θ τ,, θ τ k )τ. (MELE) of θ under model M(Θ λ ) that is Denote ˆθ λ as the maximum empirical likelihood estimate ˆθ λ = arg sup θ Θ λ l E (θ). (4.3) Then, we define the empirical-likelihood-based information criteria as ORIC EL (λ) = 2l E (ˆθ λ ) + p(λ) ϕ(n), (4.4) where p(λ) is defined in (3.5). According to Theorem 2., (4.4) is a consistent criterion. 5
6 5 A simulation study To examine the performance of the order-restricted information criterion, we executed Monte Carlo simulations for k = 4. We generated normal variates with means µ, µ 2, µ 3, µ 4 and variance. The following configurations of parameters were selected for µ = (µ, µ 2, µ 3, µ 4 ): H : µ = µ2 = µ3 = µ4, µ = (,,, ); K : µ = µ2 = µ3 µ4, K 2 : µ = µ2 µ3 = µ4, K 3 : µ µ2 = µ3 = µ4, K 4 : µ = µ2 µ3 µ4, µ = (,,,.5); µ = (,,.5,.5); µ = (,.5,.5,.5); µ = (,,.5, ); K 5 : µ µ2 = µ3 µ4, µ = (,.5,.5, ); K 6 : µ µ2 µ3 = µ4, µ = (,.5,, ); K 7 : µ µ2 µ3 µ4, µ = (,.5,,.5); M : µ = µ2 = µ3 µ4, µ = (,,,.5); M 2 : µ = µ2 µ3 = µ4, µ = (,,.5,.5); M 3 : µ µ2 = µ3 = µ4, µ = (,.5,.5,.5); M 4 : µ = µ2 µ3 µ4, µ = (,,.5, ); M 5 : µ µ2 = µ3 µ4, µ = (,.5,.5, ); M 6 : µ µ2 µ3 = µ4, µ = (,.5,, ); M 7 : µ µ2 µ3 µ4, µ = (,.5,,.5). We examined two types of comparison. First we compared the performance of our order-restricted criterion with that of Anraku(999) for detecting the true configuration of the parameters. In this simulation study, models H, K, K 2,, K 7, M, M 2,, M 7 are jointly considered. Here we assume the variance is known to be. First, the penalty numbers corresponding to H, K, K 2,, K 7, M, M 2,, M 7 are given by (3.3) for Figure and Figure 2. Second, the penalty numbers corresponding to H, K, K 2,, K 7, M, M 2,, M 7 are given by (3.5) for Figure 3 and Figure 4. Furthermore, in order to provide some clues for how to select the ϕ(n) under moderate-size sample, or even small sample, we choose ϕ(n) = (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n for our three case. The sample sizes for each population are taken equally, ranging form to 2 with step. Figure and Figure 2 show the frequency of error detection by these four methods based on simulations for true models H, K, K 2,, K 7, M, M 2,, M 7, respectively. We observe that our order information criterion work much better than Anraku(999) s order information criterion when ϕ(n) = (2/ log(log(4))) log(log(n)) or ϕ(n) = (2/ log(4)) log(n). And generally, under the restriction that lim ϕ(n)/n = 6
7 , frequencies of error detection for criteria with ϕ(n) increasing to more rapidly converge to more quickly than those with slower ϕ(n). But the slower ϕ(n) is more robust than the faster ϕ(n), see K 6 and K 7 of Figure 2. When we used the penalty numbers (3.5), from Figure 3 and Figure 4, we can draw the same conclusions. Secondly we compared four procedures for detecting the first changepoint (see (6) of Anraku(999)). Models H, M, M 2,, M 7 are considered. Here we assumed the variances of normal variates are common and unknown, and the penalty numbers are given by (3.3). And ϕ(n) is choose as above. The sample sizes for each population are taken equally, ranging form to 2 with step. Figure 5 shows the frequency of error detection by these four methods based on simulations for true models H, M, M 2,, M 7, respectively. We also observe that our order information criterion work much better than Anraku(999) s order information criterion when ϕ(n) = (2/ log(log(4))) log(log(n)) or ϕ(n) = (2/ log(4)) log(n). And the slower ϕ(n) is more robust than the faster ϕ(n), see M 3, M 6 and M 7 of Figure 5. For empirical likelihood inference, we generated X N(µ, ), X 2 N(µ 2, 2), X 3 t(6)+µ 3, X 4 χ 2 () +µ 4 Figure 6 and Figure 7 shows the frequency of error detection by these four methods based on simulations for true models H, K, K 2,, K 7, M, M 2,, M 7, respectively. We observe that our order information criterion works well when ϕ(n) = (2/ log(log(4))) log(log(n)) or ϕ(n) = (2/ log(4)) log(n). And generally, under the restriction that lim ϕ(n)/n =, frequencies of error detection for criteria with ϕ(n) increasing to more rapidly converge to more quickly than those with slower ϕ(n). But the slower ϕ(n) is more robust than the faster ϕ(n), see K 4, K 5, K 6 and K 7 of Figure 6 and Figure 7. 6 Comments and Conclusions As is well known, the calculation of P (i, k, w) may be difficult except for special weights under various order restrictions. Our information criterion avoids this computation problem. 7 Appendix Proof of Theorem 2.. When s s, t t s then the model M(s, t) with p(s, t) parameters is misspecified, so that plim log(l n (s, t))/n < plim log(l n (s, t ))/n, (A.) 7
8 where plim denotes convergence in probability. lim ϕ(n)/n = that Hence, it follows from (A.) and lim n(s, t ) c n (s, t)] = lim P [ 2 log(l n (s, t )) + p(s, t )ϕ(n)/n 2 log(l n (s, t)) + p(s, t)ϕ(n)/n] = lim P [log(l n (s, t ))/n log(l n (s, t))/n.5(p(s, t ) p(s, t))ϕ(n)/n] lim P [log(l n (s, t ))/n log(l n (s, t))/n pϕ(n)/n] = (A.2) so that = lim P [ˆt t m, ŝ s] lim P [c n (s, t ) c n (s, t) for some t t s, s s ] lim P [c n (s, t ) c n (s, t)] s s t<t (A.3) When s = s, if t < t then the model with p(s, t) parameters is misspecified, so that plim log(l n (s, t))/n < plim log(l n (s, t ))/n (A.4) Hence, it follows from (A.4) and lim ϕ(n)/n = that lim n(s, t ) c n (s, t)] = lim P [ 2 log(l n (s, t ))/n + p(s, t )ϕ(n)/n 2 log(l n (s, t))/n + p(s, t)ϕ(n)/n] = lim P [log(l n (s, t ))/n log(l n (s, t))/n.5(p(s, t ) p(s, t))ϕ(n)/n] = (A.5) so that lim P [ˆt < t, ŝ = s ] = lim P [ˆt < t ŝ = s ]P [ŝ = s ] lim P [c n (s, t ) c n (s, t) for some t < t ŝ = s ]P [ŝ = s ] t<t lim P [c n (s, t ) c n (s, t) ŝ = s ]P [ŝ = s ] = (A.6) For t > t it is follows form the assumption that 2[log(L n (s, t)) log(l n (s, t ))] d W (A.7) 8
9 where d denotes the convergence in distribution. Therefore, by lim ϕ(n) = and (A.7) implies that plim 2[log(L n (s, t)) log(l n (s, t ))]/ϕ(n) = hence plim n(c n (s, t ) c n (s, t))/ϕ(n) = plim 2[log(L n (s, t)) log(l n (s, t ))]/ϕ(n) + p(s, t ) p(s, t) = p(s, t ) p(s, t) < so that This implies that lim P [c n(s, t ) c n (s, t))] = lim P [ˆt > t, ŝ = s ] = lim P [ˆt > t ŝ = s ]P [ŝ = s ] =. Thus, we have References lim P [ŝ = s, ˆt = t ] = (A.8) [] Anraku, K. (999), An information criterion for parameters under a simple order restriction. Biometrika, 86:4-52. [2] Barlow, R.E., Bartholomew, D.J., Bremner, J.M. and Brunk, H.D. (972), Statistical Inference under Order Restrictions. New York: Wiley. [3] Barmi H.E. (996), Empirical likelihood ratio test for or against a set of inequality constraints Journal of Statistical Planning and Inference, 55:9-24. [4] Bartholomew, D.J. (959), A test of homogeneity for ordered alternatives. Biometrika, 46: [5] Bohrer, R. and Chow, W. (978), Weights for one-sided multivariate inference. Appl. Statist, 27:-4. [6] Childs, D.R. (967), Reduction of the multivariate normal integral to characteristic form. Biometrika, 54: [7] Konishi, S. and Kitagawa, G. (996), Generalised information criteria in model selection Biometrika, 83:
10 [8] Kullback, S. and Leibler, R.A. (95), On information and sufficiency. Ann.Math. Statist, 22: [9] LEE, C.I.C., Robertson, T. and Wright, F.T. (993), Bounds on distributions arising in order restricted inferences with restricted weights. Biometrika, 8: [] Perlman, M. D. (969), One-sided problems in multivariate analysis. Ann.Math. Statist, 4: [] Qin, J. and J. Lawless (955), empirical likelihood and tests of parameters subject to constraints. Canad. J. Statist, 23: [2] Rao, C.R. (973), Linear Statistical Inference and its Applications, 2nd ed. New York: Wiley. [3] Robertson, T. and Wright, F.T. (982), Bounds on mixtures of distributions arising in order restricted inference. Ann. Statist, : [4] Robertson, T. and Wright, F.T. and Dykstra, R.L. (988), Order Restricted Statistical Inference. New York: Wiley. [5] Williams, D.A. (97), A test for differences between treatment means when several dose levels are compared with a zero dose control Biometrics, 27:3-7. [6] Zhao, L. and Peng, L. (22), Model selection under order restriction. Statistics & Probability Letters, 57:3-36.
11 .7 H.7 K.7 K K 3 K 4 K Figure : The curve denote the curve for Anraku(999) s order information criterion. Those for our restricted criteria with the penalty numbers (3.3), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively.
12 K 6 K 7.8 M M 2 M M M M 4.5 M 7 Figure 2: The curve denote the curve for Anraku(999) s order information criterion.those for our restricted criteria with the penalty numbers (3.3), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 2
13 .7 H.7 K.7 K K 3 K 4 K Figure 3: The curve denote the curve for Anraku(999) s order information criterion. Those for our restricted criteria with the penalty numbers (3.5), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 3
14 K 6 K 7.8 M M 2 M M M M 4.5 M 7 Figure 4: The curve denote the curve for Anraku(999) s order information criterion. Those for our restricted criteria with the penalty numbers (3.5), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 4
15 .4 H.7 M.7 M 2.5 M M 4.5 M 5.4 M 6.4 M Figure 5: The curve denote the curve for Anraku(999) s order information criterion. Those for our restricted criteria with the penalty numbers (3.3), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 5
16 .7 H.8 K.8 K K 3 K 4 K Figure 6: Those for our empirical-likelihood-based information criteria (4.4), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 6
17 K 6 K 7.8 M M 2 M M 3.5 M 6.5 M 4.5 M 7 Figure 7: Those for our empirical-likelihood-based information criteria (4.4), and ϕ(n) being choose as (2/ log(log(4))) log(log(n)), (2/ log(4)) log(n), (2/ 4) n are showed by x, +, * curve, respectively. 7
More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction
Sankhyā : The Indian Journal of Statistics 2007, Volume 69, Part 4, pp. 700-716 c 2007, Indian Statistical Institute More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order
More informationMISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30
MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)
More informationA Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when the Covariance Matrices are Unknown but Common
Journal of Statistical Theory and Applications Volume 11, Number 1, 2012, pp. 23-45 ISSN 1538-7887 A Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when
More informationAkaike Information Criterion
Akaike Information Criterion Shuhua Hu Center for Research in Scientific Computation North Carolina State University Raleigh, NC February 7, 2012-1- background Background Model statistical model: Y j =
More informationEconometrics I, Estimation
Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the
More informationLasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices
Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,
More informationAnalysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems
Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems Jeremy S. Conner and Dale E. Seborg Department of Chemical Engineering University of California, Santa Barbara, CA
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More information1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsa
On the Use of Marginal Likelihood in Model Selection Peide Shi Department of Probability and Statistics Peking University, Beijing 100871 P. R. China Chih-Ling Tsai Graduate School of Management University
More informationModel comparison and selection
BS2 Statistical Inference, Lectures 9 and 10, Hilary Term 2008 March 2, 2008 Hypothesis testing Consider two alternative models M 1 = {f (x; θ), θ Θ 1 } and M 2 = {f (x; θ), θ Θ 2 } for a sample (X = x)
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationHigh Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data
High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data Song Xi CHEN Guanghua School of Management and Center for Statistical Science, Peking University Department
More informationBias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition
Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Hirokazu Yanagihara 1, Tetsuji Tonda 2 and Chieko Matsumoto 3 1 Department of Social Systems
More informationThe Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models
The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population Health
More informationModel Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model
Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population
More informationMIT Spring 2016
MIT 18.655 Dr. Kempthorne Spring 2016 1 MIT 18.655 Outline 1 2 MIT 18.655 Decision Problem: Basic Components P = {P θ : θ Θ} : parametric model. Θ = {θ}: Parameter space. A{a} : Action space. L(θ, a) :
More informationEM Algorithm II. September 11, 2018
EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data
More informationGeneralized Linear Models
Generalized Linear Models Lecture 3. Hypothesis testing. Goodness of Fit. Model diagnostics GLM (Spring, 2018) Lecture 3 1 / 34 Models Let M(X r ) be a model with design matrix X r (with r columns) r n
More informationSelection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models
Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models Junfeng Shang Bowling Green State University, USA Abstract In the mixed modeling framework, Monte Carlo simulation
More informationSpring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM
University of Illinois Department of Economics Spring 2017 Econ 574 Roger Koenker Lecture 14 GEE-GMM Throughout the course we have emphasized methods of estimation and inference based on the principle
More informationLikelihood-based inference with missing data under missing-at-random
Likelihood-based inference with missing data under missing-at-random Jae-kwang Kim Joint work with Shu Yang Department of Statistics, Iowa State University May 4, 014 Outline 1. Introduction. Parametric
More informationConsistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis
Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis Tetsuro Sakurai and Yasunori Fujikoshi Center of General Education, Tokyo University
More informationModification and Improvement of Empirical Likelihood for Missing Response Problem
UW Biostatistics Working Paper Series 12-30-2010 Modification and Improvement of Empirical Likelihood for Missing Response Problem Kwun Chuen Gary Chan University of Washington - Seattle Campus, kcgchan@u.washington.edu
More informationResearch Article Optimal Portfolio Estimation for Dependent Financial Returns with Generalized Empirical Likelihood
Advances in Decision Sciences Volume 2012, Article ID 973173, 8 pages doi:10.1155/2012/973173 Research Article Optimal Portfolio Estimation for Dependent Financial Returns with Generalized Empirical Likelihood
More informationDA Freedman Notes on the MLE Fall 2003
DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar
More informationAll models are wrong but some are useful. George Box (1979)
All models are wrong but some are useful. George Box (1979) The problem of model selection is overrun by a serious difficulty: even if a criterion could be settled on to determine optimality, it is hard
More information9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures
FE661 - Statistical Methods for Financial Engineering 9. Model Selection Jitkomut Songsiri statistical models overview of model selection information criteria goodness-of-fit measures 9-1 Statistical models
More informationModel Selection for Semiparametric Bayesian Models with Application to Overdispersion
Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS020) p.3863 Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Jinfang Wang and
More informationStandard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j
Standard Errors & Confidence Intervals β β asy N(0, I( β) 1 ), where I( β) = [ 2 l(β, φ; y) ] β i β β= β j We can obtain asymptotic 100(1 α)% confidence intervals for β j using: β j ± Z 1 α/2 se( β j )
More informationAn Akaike Criterion based on Kullback Symmetric Divergence in the Presence of Incomplete-Data
An Akaike Criterion based on Kullback Symmetric Divergence Bezza Hafidi a and Abdallah Mkhadri a a University Cadi-Ayyad, Faculty of sciences Semlalia, Department of Mathematics, PB.2390 Marrakech, Moroco
More informationLong-Run Covariability
Long-Run Covariability Ulrich K. Müller and Mark W. Watson Princeton University October 2016 Motivation Study the long-run covariability/relationship between economic variables great ratios, long-run Phillips
More informationGraduate Econometrics I: Maximum Likelihood I
Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood
More informationModel Selection and Geometry
Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model
More informationAN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY
Econometrics Working Paper EWP0401 ISSN 1485-6441 Department of Economics AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY Lauren Bin Dong & David E. A. Giles Department of Economics, University of Victoria
More information1 Hypothesis Testing and Model Selection
A Short Course on Bayesian Inference (based on An Introduction to Bayesian Analysis: Theory and Methods by Ghosh, Delampady and Samanta) Module 6: From Chapter 6 of GDS 1 Hypothesis Testing and Model Selection
More information7. Estimation and hypothesis testing. Objective. Recommended reading
7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing
More informationConjugate Predictive Distributions and Generalized Entropies
Conjugate Predictive Distributions and Generalized Entropies Eduardo Gutiérrez-Peña Department of Probability and Statistics IIMAS-UNAM, Mexico Padova, Italy. 21-23 March, 2013 Menu 1 Antipasto/Appetizer
More information10-704: Information Processing and Learning Fall Lecture 24: Dec 7
0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 24: Dec 7 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy of
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationConsistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis
Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Yasunori Fujikoshi and Tetsuro Sakurai Department of Mathematics, Graduate School of Science,
More informationComment about AR spectral estimation Usually an estimate is produced by computing the AR theoretical spectrum at (ˆφ, ˆσ 2 ). With our Monte Carlo
Comment aout AR spectral estimation Usually an estimate is produced y computing the AR theoretical spectrum at (ˆφ, ˆσ 2 ). With our Monte Carlo simulation approach, for every draw (φ,σ 2 ), we can compute
More informationStatistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach
Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score
More informationModel Fitting, Bootstrap, & Model Selection
Model Fitting, Bootstrap, & Model Selection G. Jogesh Babu Penn State University http://www.stat.psu.edu/ babu http://astrostatistics.psu.edu Model Fitting Non-linear regression Density (shape) estimation
More informationCOS513 LECTURE 8 STATISTICAL CONCEPTS
COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions
More informationAn Extended BIC for Model Selection
An Extended BIC for Model Selection at the JSM meeting 2007 - Salt Lake City Surajit Ray Boston University (Dept of Mathematics and Statistics) Joint work with James Berger, Duke University; Susie Bayarri,
More informationModel selection in penalized Gaussian graphical models
University of Groningen e.c.wit@rug.nl http://www.math.rug.nl/ ernst January 2014 Penalized likelihood generates a PATH of solutions Consider an experiment: Γ genes measured across T time points. Assume
More informationModelling the Covariance
Modelling the Covariance Jamie Monogan Washington University in St Louis February 9, 2010 Jamie Monogan (WUStL) Modelling the Covariance February 9, 2010 1 / 13 Objectives By the end of this meeting, participants
More informationCovariance function estimation in Gaussian process regression
Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian
More informationover the parameters θ. In both cases, consequently, we select the minimizing
MONOTONE REGRESSION JAN DE LEEUW Abstract. This is an entry for The Encyclopedia of Statistics in Behavioral Science, to be published by Wiley in 200. In linear regression we fit a linear function y =
More informationStatistical Inference
Statistical Inference Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Week 12. Testing and Kullback-Leibler Divergence 1. Likelihood Ratios Let 1, 2, 2,...
More informationLabel Switching and Its Simple Solutions for Frequentist Mixture Models
Label Switching and Its Simple Solutions for Frequentist Mixture Models Weixin Yao Department of Statistics, Kansas State University, Manhattan, Kansas 66506, U.S.A. wxyao@ksu.edu Abstract The label switching
More informationInformation Theory Based Estimator of the Number of Sources in a Sparse Linear Mixing Model
Information heory Based Estimator of the Number of Sources in a Sparse Linear Mixing Model Radu Balan University of Maryland Department of Mathematics, Center for Scientific Computation And Mathematical
More informationGibbs Sampling in Linear Models #2
Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling
More informationCOMPARING TRANSFORMATIONS USING TESTS OF SEPARATE FAMILIES. Department of Biostatistics, University of North Carolina at Chapel Hill, NC.
COMPARING TRANSFORMATIONS USING TESTS OF SEPARATE FAMILIES by Lloyd J. Edwards and Ronald W. Helms Department of Biostatistics, University of North Carolina at Chapel Hill, NC. Institute of Statistics
More informationSize and Shape of Confidence Regions from Extended Empirical Likelihood Tests
Biometrika (2014),,, pp. 1 13 C 2014 Biometrika Trust Printed in Great Britain Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests BY M. ZHOU Department of Statistics, University
More informationSemiparametric Gaussian Copula Models: Progress and Problems
Semiparametric Gaussian Copula Models: Progress and Problems Jon A. Wellner University of Washington, Seattle 2015 IMS China, Kunming July 1-4, 2015 2015 IMS China Meeting, Kunming Based on joint work
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationIntroduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf
1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 2013-14 We know that X ~ B(n,p), but we do not know p. We get a random sample
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction
More informationOn the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models
On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models Thomas Kneib Department of Mathematics Carl von Ossietzky University Oldenburg Sonja Greven Department of
More informationConsistency of test based method for selection of variables in high dimensional two group discriminant analysis
https://doi.org/10.1007/s42081-019-00032-4 ORIGINAL PAPER Consistency of test based method for selection of variables in high dimensional two group discriminant analysis Yasunori Fujikoshi 1 Tetsuro Sakurai
More informationGeneralized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses
Ann Inst Stat Math (2009) 61:773 787 DOI 10.1007/s10463-008-0172-6 Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Taisuke Otsu Received: 1 June 2007 / Revised:
More informationAkaike criterion: Kullback-Leibler discrepancy
Model choice. Akaike s criterion Akaike criterion: Kullback-Leibler discrepancy Given a family of probability densities {f ( ; ψ), ψ Ψ}, Kullback-Leibler s index of f ( ; ψ) relative to f ( ; θ) is (ψ
More informationis a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.
Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable
More informationEmpirical likelihood-based methods for the difference of two trimmed means
Empirical likelihood-based methods for the difference of two trimmed means 24.09.2012. Latvijas Universitate Contents 1 Introduction 2 Trimmed mean 3 Empirical likelihood 4 Empirical likelihood for the
More informationBayesian Analysis of Risk for Data Mining Based on Empirical Likelihood
1 / 29 Bayesian Analysis of Risk for Data Mining Based on Empirical Likelihood Yuan Liao Wenxin Jiang Northwestern University Presented at: Department of Statistics and Biostatistics Rutgers University
More informationMaximum likelihood: counterexamples, examples, and open problems. Jon A. Wellner. University of Washington. Maximum likelihood: p.
Maximum likelihood: counterexamples, examples, and open problems Jon A. Wellner University of Washington Maximum likelihood: p. 1/7 Talk at University of Idaho, Department of Mathematics, September 15,
More informationA MODIFIED LIKELIHOOD RATIO TEST FOR HOMOGENEITY IN FINITE MIXTURE MODELS
A MODIFIED LIKELIHOOD RATIO TEST FOR HOMOGENEITY IN FINITE MIXTURE MODELS Hanfeng Chen, Jiahua Chen and John D. Kalbfleisch Bowling Green State University and University of Waterloo Abstract Testing for
More informationF & B Approaches to a simple model
A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys
More informationDiscrepancy-Based Model Selection Criteria Using Cross Validation
33 Discrepancy-Based Model Selection Criteria Using Cross Validation Joseph E. Cavanaugh, Simon L. Davies, and Andrew A. Neath Department of Biostatistics, The University of Iowa Pfizer Global Research
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationSimulation of truncated normal variables. Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris
Simulation of truncated normal variables Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris Abstract arxiv:0907.4010v1 [stat.co] 23 Jul 2009 We provide in this paper simulation algorithms
More informationCh. 5 Hypothesis Testing
Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,
More informationLarge Sample Properties of Estimators in the Classical Linear Regression Model
Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in
More informationPractice Problems Section Problems
Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,
More informationReview and continuation from last week Properties of MLEs
Review and continuation from last week Properties of MLEs As we have mentioned, MLEs have a nice intuitive property, and as we have seen, they have a certain equivariance property. We will see later that
More informationLecture 4 September 15
IFT 6269: Probabilistic Graphical Models Fall 2017 Lecture 4 September 15 Lecturer: Simon Lacoste-Julien Scribe: Philippe Brouillard & Tristan Deleu 4.1 Maximum Likelihood principle Given a parametric
More information11 Survival Analysis and Empirical Likelihood
11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with
More informationEstimating GARCH models: when to use what?
Econometrics Journal (2008), volume, pp. 27 38. doi: 0./j.368-423X.2008.00229.x Estimating GARCH models: when to use what? DA HUANG, HANSHENG WANG AND QIWEI YAO, Guanghua School of Management, Peking University,
More informationICSA Applied Statistics Symposium 1. Balanced adjusted empirical likelihood
ICSA Applied Statistics Symposium 1 Balanced adjusted empirical likelihood Art B. Owen Stanford University Sarah Emerson Oregon State University ICSA Applied Statistics Symposium 2 Empirical likelihood
More informationInformation in Data. Sufficiency, Ancillarity, Minimality, and Completeness
Information in Data Sufficiency, Ancillarity, Minimality, and Completeness Important properties of statistics that determine the usefulness of those statistics in statistical inference. These general properties
More informationMcGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper
McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section
More informationQuantifying Stochastic Model Errors via Robust Optimization
Quantifying Stochastic Model Errors via Robust Optimization IPAM Workshop on Uncertainty Quantification for Multiscale Stochastic Systems and Applications Jan 19, 2016 Henry Lam Industrial & Operations
More informationA Bayesian Multiple Comparison Procedure for Order-Restricted Mixed Models
A Bayesian Multiple Comparison Procedure for Order-Restricted Mixed Models Junfeng Shang 1, Joseph E. Cavanaugh 2, Farroll T. Wright 3, 1 Department of Mathematics and Statistics, 450 Math Science Building,
More informationModel Selection in the Presence of Incidental Parameters
Model Selection in the Presence of Incidental Parameters Yoonseok Lee University of Michigan February 2012 Abstract This paper considers model selection of nonlinear panel data models in the presence of
More informationInvestigation of goodness-of-fit test statistic distributions by random censored samples
d samples Investigation of goodness-of-fit test statistic distributions by random censored samples Novosibirsk State Technical University November 22, 2010 d samples Outline 1 Nonparametric goodness-of-fit
More informationEstimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk
Ann Inst Stat Math (0) 64:359 37 DOI 0.007/s0463-00-036-3 Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk Paul Vos Qiang Wu Received: 3 June 009 / Revised:
More informationAugmented Likelihood Estimators for Mixture Models
Augmented Likelihood Estimators for Mixture Models Markus Haas Jochen Krause Marc S. Paolella Swiss Banking Institute, University of Zurich What is mixture degeneracy? mixtures under study are finite convex
More informationA union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling
A union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling Min-ge Xie Department of Statistics, Rutgers University Workshop on Higher-Order Asymptotics
More informationIntroduction to graphical models: Lecture III
Introduction to graphical models: Lecture III Martin Wainwright UC Berkeley Departments of Statistics, and EECS Martin Wainwright (UC Berkeley) Some introductory lectures January 2013 1 / 25 Introduction
More information7. Estimation and hypothesis testing. Objective. Recommended reading
7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing
More informationSTAT 535 Lecture 5 November, 2018 Brief overview of Model Selection and Regularization c Marina Meilă
STAT 535 Lecture 5 November, 2018 Brief overview of Model Selection and Regularization c Marina Meilă mmp@stat.washington.edu Reading: Murphy: BIC, AIC 8.4.2 (pp 255), SRM 6.5 (pp 204) Hastie, Tibshirani
More informationEmpirical Likelihood Methods for Sample Survey Data: An Overview
AUSTRIAN JOURNAL OF STATISTICS Volume 35 (2006), Number 2&3, 191 196 Empirical Likelihood Methods for Sample Survey Data: An Overview J. N. K. Rao Carleton University, Ottawa, Canada Abstract: The use
More informationChapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools. Joan Llull. Microeconometrics IDEA PhD Program
Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools Joan Llull Microeconometrics IDEA PhD Program Maximum Likelihood Chapter 1. A Brief Review of Maximum Likelihood, GMM, and Numerical
More informationEstimation in Generalized Linear Models with Heterogeneous Random Effects. Woncheol Jang Johan Lim. May 19, 2004
Estimation in Generalized Linear Models with Heterogeneous Random Effects Woncheol Jang Johan Lim May 19, 2004 Abstract The penalized quasi-likelihood (PQL) approach is the most common estimation procedure
More informationMixture Modeling. Identifying the Correct Number of Classes in a Growth Mixture Model. Davood Tofighi Craig Enders Arizona State University
Identifying the Correct Number of Classes in a Growth Mixture Model Davood Tofighi Craig Enders Arizona State University Mixture Modeling Heterogeneity exists such that the data are comprised of two or
More informationCausal Model Selection Hypothesis Tests in Systems Genetics
1 Causal Model Selection Hypothesis Tests in Systems Genetics Elias Chaibub Neto and Brian S Yandell SISG 2012 July 13, 2012 2 Correlation and Causation The old view of cause and effect... could only fail;
More informationModel Selection in the Presence of Incidental Parameters
Model Selection in the Presence of Incidental Parameters Yoonseok Lee University of Michigan July 2012 Abstract This paper considers model selection of nonlinear panel data models in the presence of incidental
More informationLTI Systems, Additive Noise, and Order Estimation
LTI Systems, Additive oise, and Order Estimation Soosan Beheshti, Munther A. Dahleh Laboratory for Information and Decision Systems Department of Electrical Engineering and Computer Science Massachusetts
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More information