Minimax estimators of the coverage probability of the impermissible error for a location family
|
|
- Elfrieda Craig
- 5 years ago
- Views:
Transcription
1 Minimax estimators of the coverage probability of the impermissible error for a location family by Miguel A. Arcones Binghamton University arcones@math.binghamton.edu Talk based on: Arcones, M. A. (2008). Minimax estimators of the coverage probability of the impermissible error for a location family. Statistics & Decisions. 28, Available at 1
2 1 Confidence regions with a constrained Lebesgue measure for a location family. Let {X j } j=1 be a sequence of Rd valued i.i.d.r.v. s with a p.d.f. belonging to the family {f( θ) : θ R d }. Fix L > 0. We would like to find a translation equivariant confidence region C HL,L (X 1,..., X n ) such that: 1. For each x := (x 1,..., x n ), R m I(y C HL,L ( x)) dy L. (1) 2. For any confidence region C(X 1,..., X n ) satisfying (1), inf P θ{θ C HL,L ( X)} inf P θ{θ C( X)} θ R d θ R d 2
3 If C( X) is a translation equivariant confidence region, then C( X) = X n + C(Z 1,..., Z n 1, 0) = X n + C ( Z). We would like to find C ( Z) maximizing inf θ R d P θ {θ X n + C ( Z)} = P 0 { X n C ( Z)} (2) = I( x n C ( z)) n 1 j=1 f(z j + x n ) f(x n ) dx n d z = I(y C ( z)) n 1 j=1 f(z j y) f( y) dy d z, subject that for each z, I(y x n +C ( z)) dy = I(y C ( z)) dy L. (3) R m R m Conditioning on z, we get that the optimal region is C HL( z) = {y R d : λ HL ( z) where λ HL ( z) is such that It is easy to see that n 1 j=1 λ HL ( z) n 1 j=1 f(z j y) f( y)) f(z j y) f( y))}. dy = L. C HL,L (x 1,..., x n ) = x n + C HL ( z) = {θ R d : λ HL ( z) n j=1 f(x j θ)}. 3
4 Theorem 1. Suppose that: (i) For each 0 < M <, inf{f(x) : x M} > 0. (ii) lim x f(x) = 0. (iii) sup x R d f(x) <. Let C(X 1,..., X n ) be a confidence region for θ such that for each x 1,..., x n, R d I(y C(x 1,..., x n )) dy L. Then, inf θ R d P θ {θ C HL,L ( X)} inf θ R d P θ {θ C( X)}. 4
5 2 Impermissible error Let T n := T n ( X) be an estimator of θ. The error of the estimator T n is T n θ. Suppose that we select a number b, b > 0, such that any error of estimation bigger than b is impermissible. The maximum coverage probability of the impermissible error of the estimator T n is sup P θ { T n θ > b}. (4) θ Θ We would like to obtain an estimator T HL,b minimizing (4) for a location family. Finding T HL,b is equivalent to find a confidence interval of length 2b with coverage probability as large as possible. If d = 1, and C HL,2b ( X) is an interval, we take T n,hl,b ( X) as the middle point of C HL,2b ( X). By Theorem 1, for each estimator T n, sup P θ { T n θ > b} sup P θ { T n,hl,b θ > b}. θ R θ R 5
6 3 Estimation via the m.l.e. An mle ˆθ n of θ, over a parametric family {f(x, θ) : θ Θ} is a r.v. such that n n 1 log f(x j, ˆθ n ) = sup n 1 θ Θ j=1 n log f(x j, θ). (5) Suppose that there exists a r.v. ˆθn = ˆθ n (X 1,..., X n ) satisfying (5) when f(x, θ) = f(x θ). We may assume that ˆθ n is equivariant. This implies that ˆθ n θ is a pivotal quantity. Suppose that θ ˆθ n has a distribution absolutely continuous with respect to the Lebesgue measure, when θ obtains. Let h n ( ) be the pdf of θ ˆθ n when θ obtains. Let L > 0. Given C B(R d ), a confidence region for θ based on θ ˆθ n is determined by {θ R d : θ ˆθ n C} = {θ R d : θ ˆθ n + C}. j=1 6
7 Between all sets C such that I(θ ˆθ n + C) dθ = a set C which maximizes is P θ {θ ˆθ n C} = C mle,l = {x R d : λ mle,l h n (x)} where λ mle,l is such that {x R d :λ mle,l h n (x)} C C 1 dθ L, (6) h n (θ) dθ (7) 1 dx = L. Hence, using the mle, a confidence region satisfying (6) and maximizing (7) is ˆθ n + C mle,l = {θ R d : λ mle,l h n (θ ˆθ n )}. Notice h n (θ x), x R, is the p.d.f. of ˆθ n when θ obtains. 7
8 The obtained confidence regions maximize the coverage probability over all the regions with expected Lebesgue measure less or equal than L: Theorem 2. Let C(ˆθ n ) be a confidence region for θ such that Then, [ ] sup E θ I(y C(ˆθ n )) dy θ R d R d L. (8) inf P θ{θ C(ˆθ n )} inf P θ{θ ˆθ n + C mle,l }. (9) θ R d θ R d The previous theorem implies that if C(ˆθ n ) is a confidence region for θ such that for each x R d, then, (9) holds. R d I(y C(x)) dy L, (10) 8
9 Suppose that d = 1 and we would like to find the estimator T (ˆθ n ) (based on the m.l.e.) minimizing: sup P θ { T (ˆθ n ) θ > b}. (11) θ R If C mle,2b is an interval, then the middle point T n,mle,b of ˆθ n + C mle,2b minimizes (11) between all the estimators based on the m.l.e. 9
10 4 Asymptotic results. Let {X j } j=1 {f( θ) : θ R}. be a sequence of i.i.d.r.v. s from the family Theorem 3. Suppose that: (i) For each x R, f(x) > 0. (ii) log f( ) is a strictly concave function (iii) For each t R, there exists λ t > 0 such that Then, where E 0 [exp(λ t log f(x t) )] <. lim n n 1 log P θ {θ C HL,L (X 1,..., X n )} = S(L), Besides, S(L) = inf λ R log R (f(x L))λ (f(x)) 1 λ dx. (a) S is increasing on [0, ) and S is decreasing in (, 0]. (b) S is a continuous function. (c) lim t S(t) =. 10
11 Theorem 4. Suppose that: (i) For each x R, f(x) > 0. (ii) log f is a strictly concave function. (iii) R f (x) dx = 0. (iv) E[(f(X)) 2 (f (X)) 2 ] <. Then, the m.l.e. ˆθ n is well defined and θ ˆθ n, when θ obtains, has pdf ( n ) h n (t) := E 0 [I j=1 (f(x j + t)) 1 f (X j + t) > 0 ] n j=1 (f(x j)) 1 f (X j ), t R. Besides, h n is nonincreasing in [0, ) and nondecreasing in (, 0]. 11
12 Theorem 5. Suppose that: (i) For each x R, f(x) > 0. (ii) log f is a concave function. (iii) E 0 [(f(x)) 1 f (X)] = 0. (iv) E 0 [(f(x)) 2 (f (X)) 2 ] <. (v) ˆθ n θ satisfies the LDP with speed n and continuous rate function R(t) := inf λ R log E 0[exp(λ(f(X t)) 1 f (X t))], t R. Then, lim n n 1 log P θ {θ ˆθ n C mle,l (ˆθ n )} = S mle (L) where S mle (L) := inf{u 0 : x:r(x) u 1 dx L}. Besides, (i) For each t R, R(t) <. (ii) R(0) = 0 (iii) R is increasing in [0, ) and decreasing in (, 0]. (iv) R is continuous in R. (v) There exists a t 0 (0, L) such that R(t) = R(t 0 L) = S mle (L). 12
13 Example 1. If f(x) = (2π) 1/2 σ 1 exp( 2 1 σ 2 x 2 ), x R, (12) where σ > 0, then, it is easy to see that: (i) T HL,b = T mle,b = ˆθ n = X := n 1 n j=1 X j. (ii) C HL,2b (X 1,..., X n ) = C mle,2b (ˆθ n ) = [ X b, X + b]. (iii) S mle (2b) = S(2b) = R(b) = 2 1 σ 2 b 2. P (iv) For each θ R, T θ HL,b θ. 13
14 Example 2. If f(x) = (Γ(α)) 1 γ α α exp (αγx αe γx ) (13) where α > 0 and γ 0, then, it is possible to see that: (i) ˆθ ( n = γ 1 log n 1 ) n j=1 eγx j. (ii) C HL,2b (X 1,..., X n ) = C mle,2b (ˆθ n ) = [ˆθ n + γ 1 log ( 2 1 γ 1 b 1 (e 2γb 1) ) 2b, ˆθ n + γ 1 log ( 2 1 γ 1 b 1 (e 2γb 1) ) ]. (iii) T HL,b = T mle,b = ˆθ n +γ 1 log ( 2 1 γ 1 b 1 (e γb e γb ) ). (iv) For each b > 0, S(2b) = S mle (2b) = R(b) = α log(2 1 γ 1 b 1 (e 2 γ b 1)) α(e 2 γ b 1) 1 (e 2 γ b 1 2 γ b). (v) For each b > 0 and each θ R, P T θ HL,b θ + γ 1 log(2 1 γ 1 b 1 (e γb e γb )) θ. By Theorem 2 in Ferguson [1] the location families in examples 1 and 2 are the only one dimensional location families, which are exponential families. 14
15 Example 3. If f(x) = c exp( a 1 exp(τ 1 x) a 2 exp( τ 2 x)), (14) where a 1, a 2, τ 1, τ 2 > 0 and c makes f a p.d.f. Then, (i) ( n j=1 eτ 1X j, n j=1 e τ 2X j ) is a minimal sufficient stat. (ii) The family {f( θ) : θ R} is a curved exponential family. (iii) The m.l.e. ˆθ n of θ is (τ 1 + τ 2 ) 1 log (iv) where a 1 2 τ 1 2 a 1τ 1 n j=1 e τ 2X j 1 n j=1 e τ 1X j C HL,2b (X 1,..., X n ) = C mle,2b (ˆθ n ) = ˆθ n + [t 0 2b, t 0 ], (v) t 0 = (τ 1 + τ 2 ) 1 log ( τ 1 1 τ 2(e 2bτ 1 1)(1 e 2bτ 2 ) 1). T HL,b = T mle,b = ˆθ n +(τ 1 + τ 2 ) 1 log ( τ 1 1 τ 2(e bτ 1 e bτ 1)(e bτ 2 e bτ 2) 1). 15.
16 Example 4. If f(x) = 2 1 exp( x ), x R, (15) then, ˆθ n is not uniquely defined. is concave, but not strictly concave. The function log f( ) Let X (1),..., X (n) be the order statistics. X 1,..., X n are all different wit probability one. Assume that X 1,..., X n are all different. If n is odd, ˆθ n = X (2 1 (n+1)). If n is even, then ˆθ n = X (2 1 n) and ˆθ n = X (2 1 n+1) are both m.l.e. s. It is easy to see that whatever choice for the m.l.e. is made theorems 1 5 apply giving that (i) C HL,2b (X 1,..., X n ) C mle,2b (ˆθ n ). (ii) For each b > 0, S(2b) = b log(1 + b). (iii) For each b > 0, S mle (2b) = b 2 1 log ( 2e b 1 ). (iii) For each b > 0, S mle (2b) < S(2b). P (iv) For each θ R, T θ HL,b θ. 16
17 Theorem 6. Suppose that: (i) For each x R, f(x) > 0. (ii) log f is a concave function. (iii) f is even. Then, (i) For each b > 0, R(b) S(2b). (ii) Assume that f has a third derivative and that f is not a p.d.f. from the families of p.d.f. s in examples 1 and 3, then for some b 0, R(b) < S(2b). References [1] Ferguson, T. S. (1962). Location and scale parameters in exponential families of distributions. Ann. Mathemat. Statist
2014/2015 Smester II ST5224 Final Exam Solution
014/015 Smester II ST54 Final Exam Solution 1 Suppose that (X 1,, X n ) is a random sample from a distribution with probability density function f(x; θ) = e (x θ) I [θ, ) (x) (i) Show that the family of
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationDefinition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by speci
Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by specifying one or more values called parameters. The number
More informationChapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets
Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Confidence sets X: a sample from a population P P. θ = θ(p): a functional from P to Θ R k for a fixed integer k. C(X): a confidence
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth
More informationLarge deviations of empirical processes
Large deviations of empirical processes Miguel A. Arcones Abstract. We give necessary and sufficient conditions for the large deviations of empirical processes and of Banach space valued random vectors.
More informationSection 8: Asymptotic Properties of the MLE
2 Section 8: Asymptotic Properties of the MLE In this part of the course, we will consider the asymptotic properties of the maximum likelihood estimator. In particular, we will study issues of consistency,
More informationLecture 17: Likelihood ratio and asymptotic tests
Lecture 17: Likelihood ratio and asymptotic tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f
More informationA Very Brief Summary of Bayesian Inference, and Examples
A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X
More informationTesting Algebraic Hypotheses
Testing Algebraic Hypotheses Mathias Drton Department of Statistics University of Chicago 1 / 18 Example: Factor analysis Multivariate normal model based on conditional independence given hidden variable:
More informationStatistical Inference
Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will
More informationLecture 26: Likelihood ratio tests
Lecture 26: Likelihood ratio tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f θ0 (X) > c 0 for
More informationAn exponential family of distributions is a parametric statistical model having densities with respect to some positive measure λ of the form.
Stat 8112 Lecture Notes Asymptotics of Exponential Families Charles J. Geyer January 23, 2013 1 Exponential Families An exponential family of distributions is a parametric statistical model having densities
More informationStatistical Inference
Statistical Inference Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA. Asymptotic Inference in Exponential Families Let X j be a sequence of independent,
More informationReview and continuation from last week Properties of MLEs
Review and continuation from last week Properties of MLEs As we have mentioned, MLEs have a nice intuitive property, and as we have seen, they have a certain equivariance property. We will see later that
More informationSTAT215: Solutions for Homework 2
STAT25: Solutions for Homework 2 Due: Wednesday, Feb 4. (0 pt) Suppose we take one observation, X, from the discrete distribution, x 2 0 2 Pr(X x θ) ( θ)/4 θ/2 /2 (3 θ)/2 θ/4, 0 θ Find an unbiased estimator
More informationEconometrics I, Estimation
Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the
More informationLecture 32: Asymptotic confidence sets and likelihoods
Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence
More information) ) = γ. and P ( X. B(a, b) = Γ(a)Γ(b) Γ(a + b) ; (x + y, ) I J}. Then, (rx) a 1 (ry) b 1 e (x+y)r r 2 dxdy Γ(a)Γ(b) D
3 Independent Random Variables II: Examples 3.1 Some functions of independent r.v. s. Let X 1, X 2,... be independent r.v. s with the known distributions. Then, one can compute the distribution of a r.v.
More informationECE598: Information-theoretic methods in high-dimensional statistics Spring 2016
ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture : Mutual Information Method Lecturer: Yihong Wu Scribe: Jaeho Lee, Mar, 06 Ed. Mar 9 Quick review: Assouad s lemma
More informationChapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1)
Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1) Detection problems can usually be casted as binary or M-ary hypothesis testing problems. Applications: This chapter: Simple hypothesis
More informationMaximum Likelihood Estimation
Maximum Likelihood Estimation Assume X P θ, θ Θ, with joint pdf (or pmf) f(x θ). Suppose we observe X = x. The Likelihood function is L(θ x) = f(x θ) as a function of θ (with the data x held fixed). The
More informationStatistics Ph.D. Qualifying Exam
Department of Statistics Carnegie Mellon University May 7 2008 Statistics Ph.D. Qualifying Exam You are not expected to solve all five problems. Complete solutions to few problems will be preferred to
More informationlim n C1/n n := ρ. [f(y) f(x)], y x =1 [f(x) f(y)] [g(x) g(y)]. (x,y) E A E(f, f),
1 Part I Exercise 1.1. Let C n denote the number of self-avoiding random walks starting at the origin in Z of length n. 1. Show that (Hint: Use C n+m C n C m.) lim n C1/n n = inf n C1/n n := ρ.. Show that
More informationLecture 25: Review. Statistics 104. April 23, Colin Rundel
Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April
More informationOptimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X.
Optimization Background: Problem: given a function f(x) defined on X, find x such that f(x ) f(x) for all x X. The value x is called a maximizer of f and is written argmax X f. In general, argmax X f may
More informationAsymptotics of minimax stochastic programs
Asymptotics of minimax stochastic programs Alexander Shapiro Abstract. We discuss in this paper asymptotics of the sample average approximation (SAA) of the optimal value of a minimax stochastic programming
More informationInvariant HPD credible sets and MAP estimators
Bayesian Analysis (007), Number 4, pp. 681 69 Invariant HPD credible sets and MAP estimators Pierre Druilhet and Jean-Michel Marin Abstract. MAP estimators and HPD credible sets are often criticized in
More informationHomework 27. Homework 28. Homework 29. Homework 30. Prof. Girardi, Math 703, Fall 2012 Homework: Define f : C C and u, v : R 2 R by
Homework 27 Define f : C C and u, v : R 2 R by f(z) := xy where x := Re z, y := Im z u(x, y) = Re f(x + iy) v(x, y) = Im f(x + iy). Show that 1. u and v satisfies the Cauchy Riemann equations at (x, y)
More informationSTAT 135 Lab 2 Confidence Intervals, MLE and the Delta Method
STAT 135 Lab 2 Confidence Intervals, MLE and the Delta Method Rebecca Barter February 2, 2015 Confidence Intervals Confidence intervals What is a confidence interval? A confidence interval is calculated
More informationBrief Review on Estimation Theory
Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on
More informationStatistical Inference
Statistical Inference Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Week 12. Testing and Kullback-Leibler Divergence 1. Likelihood Ratios Let 1, 2, 2,...
More information5.1 Uniformly Most Accurate Families of Confidence
Chapter 5 Confidence Estimation Let (X, B X, P ), P {P θ θ Θ} be sample space and {Θ, B Θ, µ} be measure space with some σ finite measure µ. 5.1 Uniformly Most Accurate Families of Confidence Sets Def.
More informationChapter 4. Theory of Tests. 4.1 Introduction
Chapter 4 Theory of Tests 4.1 Introduction Parametric model: (X, B X, P θ ), P θ P = {P θ θ Θ} where Θ = H 0 +H 1 X = K +A : K: critical region = rejection region / A: acceptance region A decision rule
More informationStatistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That
Statistics Lecture 4 August 9, 2000 Frank Porter Caltech The plan for these lectures: 1. The Fundamentals; Point Estimation 2. Maximum Likelihood, Least Squares and All That 3. What is a Confidence Interval?
More information1 Likelihood. 1.1 Likelihood function. Likelihood & Maximum Likelihood Estimators
Likelihood & Maximum Likelihood Estimators February 26, 2018 Debdeep Pati 1 Likelihood Likelihood is surely one of the most important concepts in statistical theory. We have seen the role it plays in sufficiency,
More informationStat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota
Stat 5102 Lecture Slides Deck 3 Charles J. Geyer School of Statistics University of Minnesota 1 Likelihood Inference We have learned one very general method of estimation: method of moments. the Now we
More informationOptimum Joint Detection and Estimation
20 IEEE International Symposium on Information Theory Proceedings Optimum Joint Detection and Estimation George V. Moustakides Department of Electrical and Computer Engineering University of Patras, 26500
More informationSuggested solutions to written exam Jan 17, 2012
LINKÖPINGS UNIVERSITET Institutionen för datavetenskap Statistik, ANd 73A36 THEORY OF STATISTICS, 6 CDTS Master s program in Statistics and Data Mining Fall semester Written exam Suggested solutions to
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationTheory of Maximum Likelihood Estimation. Konstantin Kashin
Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical
More informationA BAYESIAN MATHEMATICAL STATISTICS PRIMER. José M. Bernardo Universitat de València, Spain
A BAYESIAN MATHEMATICAL STATISTICS PRIMER José M. Bernardo Universitat de València, Spain jose.m.bernardo@uv.es Bayesian Statistics is typically taught, if at all, after a prior exposure to frequentist
More informationStep-Stress Models and Associated Inference
Department of Mathematics & Statistics Indian Institute of Technology Kanpur August 19, 2014 Outline Accelerated Life Test 1 Accelerated Life Test 2 3 4 5 6 7 Outline Accelerated Life Test 1 Accelerated
More informationLikelihood based Statistical Inference. Dottorato in Economia e Finanza Dipartimento di Scienze Economiche Univ. di Verona
Likelihood based Statistical Inference Dottorato in Economia e Finanza Dipartimento di Scienze Economiche Univ. di Verona L. Pace, A. Salvan, N. Sartori Udine, April 2008 Likelihood: observed quantities,
More information9 Bayesian inference. 9.1 Subjective probability
9 Bayesian inference 1702-1761 9.1 Subjective probability This is probability regarded as degree of belief. A subjective probability of an event A is assessed as p if you are prepared to stake pm to win
More informationMath 211 Business Calculus TEST 3. Question 1. Section 2.2. Second Derivative Test.
Math 211 Business Calculus TEST 3 Question 1. Section 2.2. Second Derivative Test. p. 1/?? Math 211 Business Calculus TEST 3 Question 1. Section 2.2. Second Derivative Test. Question 2. Section 2.3. Graph
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing
More informationMTH 3102 Complex Variables Solutions: Practice Exam 2 Mar. 26, 2017
Name Last name, First name): MTH 31 omplex Variables Solutions: Practice Exam Mar. 6, 17 Exam Instructions: You have 1 hour & 1 minutes to complete the exam. There are a total of 7 problems. You must show
More informationELEMENTS OF PROBABILITY THEORY
ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable
More informationHW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.
HW1 solutions Exercise 1 (Some sets of probability distributions.) Let x be a real-valued random variable with Prob(x = a i ) = p i, i = 1,..., n, where a 1 < a 2 < < a n. Of course p R n lies in the standard
More informationSTAT 461/561- Assignments, Year 2015
STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationChapter 8.8.1: A factorization theorem
LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.
More informationStatistical Theory MT 2008 Problems 1: Solution sketches
Statistical Theory MT 008 Problems : Solutio sketches. Which of the followig desities are withi a expoetial family? Explai your reasoig. a) Let 0 < θ < ad put fx, θ) = θ)θ x ; x = 0,,,... b) c) where α
More informationISyE 6644 Fall 2014 Test 3 Solutions
1 NAME ISyE 6644 Fall 14 Test 3 Solutions revised 8/4/18 You have 1 minutes for this test. You are allowed three cheat sheets. Circle all final answers. Good luck! 1. [4 points] Suppose that the joint
More informationStatistical Estimation: Data & Non-data Information
Statistical Estimation: Data & Non-data Information Roger J-B Wets University of California, Davis & M.Casey @ Raytheon G.Pflug @ U. Vienna, X. Dong @ EpiRisk, G-M You @ EpiRisk. a little background Decision
More informationWeek 2 Spring Lecture 3. The Canonical normal means estimation problem (cont.).! (X) = X+ 1 X X, + (X) = X+ 1
Week 2 Spring 2009 Lecture 3. The Canonical normal means estimation problem (cont.). Shrink toward a common mean. Theorem. Let X N ; 2 I n. Let 0 < C 2 (n 3) (hence n 4). De ne!! (X) = X+ 1 Then C 2 X
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationPart III. A Decision-Theoretic Approach and Bayesian testing
Part III A Decision-Theoretic Approach and Bayesian testing 1 Chapter 10 Bayesian Inference as a Decision Problem The decision-theoretic framework starts with the following situation. We would like to
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More informationGibbs Voronoi tessellations: Modeling, Simulation, Estimation.
Gibbs Voronoi tessellations: Modeling, Simulation, Estimation. Frédéric Lavancier, Laboratoire Jean Leray, Nantes (France) Work with D. Dereudre (LAMAV, Valenciennes, France). SPA, Berlin, July 27-31,
More informationUNIVERSITY OF MANITOBA
Question Points Score INSTRUCTIONS TO STUDENTS: This is a 6 hour examination. No extra time will be given. No texts, notes, or other aids are permitted. There are no calculators, cellphones or electronic
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationSolutions to Exam 1, Math Solution. Because f(x) is one-to-one, we know the inverse function exists. Recall that (f 1 ) (a) =
Solutions to Exam, Math 56 The function f(x) e x + x 3 + x is one-to-one (there is no need to check this) What is (f ) ( + e )? Solution Because f(x) is one-to-one, we know the inverse function exists
More informationSolutions to the Exercises of Section 2.11.
Solutions to the Exercises of Section 2.11. 2.11.1. Proof. Let ɛ be an arbitrary positive number. Since r(τ n,δ n ) C, we can find an integer n such that r(τ n,δ n ) C ɛ. Then, as in the proof of Theorem
More informationOn Consistent Hypotheses Testing
Mikhail Ermakov St.Petersburg E-mail: erm2512@gmail.com History is rather distinctive. Stein (1954); Hodges (1954); Hoefding and Wolfowitz (1956), Le Cam and Schwartz (1960). special original setups: Shepp,
More informationStatistical Theory MT 2009 Problems 1: Solution sketches
Statistical Theory MT 009 Problems : Solutio sketches. Which of the followig desities are withi a expoetial family? Explai your reasoig. (a) Let 0 < θ < ad put f(x, θ) = ( θ)θ x ; x = 0,,,... (b) (c) where
More informationChoosing among models
Eco 515 Fall 2014 Chris Sims Choosing among models September 18, 2014 c 2014 by Christopher A. Sims. This document is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported
More informationStatistical Theory MT 2007 Problems 4: Solution sketches
Statistical Theory MT 007 Problems 4: Solution sketches 1. Consider a 1-parameter exponential family model with density f(x θ) = f(x)g(θ)exp{cφ(θ)h(x)}, x X. Suppose that the prior distribution has the
More informationMIT Spring 2016
MIT 18.655 Dr. Kempthorne Spring 2016 1 MIT 18.655 Outline 1 2 MIT 18.655 Decision Problem: Basic Components P = {P θ : θ Θ} : parametric model. Θ = {θ}: Parameter space. A{a} : Action space. L(θ, a) :
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationHypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes
Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis
More informationChapter 4: Asymptotic Properties of the MLE
Chapter 4: Asymptotic Properties of the MLE Daniel O. Scharfstein 09/19/13 1 / 1 Maximum Likelihood Maximum likelihood is the most powerful tool for estimation. In this part of the course, we will consider
More informationLikelihood inference in the presence of nuisance parameters
Likelihood inference in the presence of nuisance parameters Nancy Reid, University of Toronto www.utstat.utoronto.ca/reid/research 1. Notation, Fisher information, orthogonal parameters 2. Likelihood inference
More informationBrownian survival and Lifshitz tail in perturbed lattice disorder
Brownian survival and Lifshitz tail in perturbed lattice disorder Ryoki Fukushima Kyoto niversity Random Processes and Systems February 16, 2009 6 B T 1. Model ) ({B t t 0, P x : standard Brownian motion
More informationA Kullback-Leibler Divergence for Bayesian Model Comparison with Applications to Diabetes Studies. Chen-Pin Wang, UTHSCSA Malay Ghosh, U.
A Kullback-Leibler Divergence for Bayesian Model Comparison with Applications to Diabetes Studies Chen-Pin Wang, UTHSCSA Malay Ghosh, U. Florida Lehmann Symposium, May 9, 2011 1 Background KLD: the expected
More informationProbability Theory and Statistics. Peter Jochumzen
Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................
More informationHARMONIC ANALYSIS. Date:
HARMONIC ANALYSIS Contents. Introduction 2. Hardy-Littlewood maximal function 3. Approximation by convolution 4. Muckenhaupt weights 4.. Calderón-Zygmund decomposition 5. Fourier transform 6. BMO (bounded
More informationMultivariate Analysis and Likelihood Inference
Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density
More informationINDIAN INSTITUTE OF TECHNOLOGY BOMBAY MA205 Complex Analysis Autumn 2012
INDIAN INSTITUTE OF TECHNOLOGY BOMBAY MA205 Complex Analysis Autumn 2012 September 5, 2012 Mapping Properties Lecture 13 We shall once again return to the study of general behaviour of holomorphic functions
More informationPrelim Examination 2010 / 2011 (Assessing Units 1 & 2) MATHEMATICS. Advanced Higher Grade. Time allowed - 2 hours
Prelim Examination 00 / 0 (Assessing Units & ) MATHEMATICS Advanced Higher Grade Time allowed - hours Read Carefully. Calculators may be used in this paper.. Candidates should answer all questions. Full
More informationPointwise convergence rates and central limit theorems for kernel density estimators in linear processes
Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract Convergence
More informationInformation in Data. Sufficiency, Ancillarity, Minimality, and Completeness
Information in Data Sufficiency, Ancillarity, Minimality, and Completeness Important properties of statistics that determine the usefulness of those statistics in statistical inference. These general properties
More informationBerge s Maximum Theorem
Berge s Maximum Theorem References: Acemoglu, Appendix A.6 Stokey-Lucas-Prescott, Section 3.3 Ok, Sections E.1-E.3 Claude Berge, Topological Spaces (1963), Chapter 6 Berge s Maximum Theorem So far, we
More informationWe say that the function f obtains a maximum value provided that there. We say that the function f obtains a minimum value provided that there
Math 311 W08 Day 10 Section 3.2 Extreme Value Theorem (It s EXTREME!) 1. Definition: For a function f: D R we define the image of the function to be the set f(d) = {y y = f(x) for some x in D} We say that
More informationGeneration from simple discrete distributions
S-38.3148 Simulation of data networks / Generation of random variables 1(18) Generation from simple discrete distributions Note! This is just a more clear and readable version of the same slide that was
More informationApplied Statistics Lecture Notes
Applied Statistics Lecture Notes Kosuke Imai Department of Politics Princeton University February 2, 2008 Making statistical inferences means to learn about what you do not observe, which is called parameters,
More informationStatistical Theory MT 2006 Problems 4: Solution sketches
Statistical Theory MT 006 Problems 4: Solution sketches 1. Suppose that X has a Poisson distribution with unknown mean θ. Determine the conjugate prior, and associate posterior distribution, for θ. Determine
More informationis a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.
Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable
More informationLecture 35: December The fundamental statistical distances
36-705: Intermediate Statistics Fall 207 Lecturer: Siva Balakrishnan Lecture 35: December 4 Today we will discuss distances and metrics between distributions that are useful in statistics. I will be lose
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More information3 a = 3 b c 2 = a 2 + b 2 = 2 2 = 4 c 2 = 3b 2 + b 2 = 4b 2 = 4 b 2 = 1 b = 1 a = 3b = 3. x 2 3 y2 1 = 1.
MATH 222 LEC SECOND MIDTERM EXAM THU NOV 8 PROBLEM ( 5 points ) Find the standard-form equation for the hyperbola which has its foci at F ± (±2, ) and whose asymptotes are y ± 3 x The calculations b a
More informationHypothesis testing: theory and methods
Statistical Methods Warsaw School of Economics November 3, 2017 Statistical hypothesis is the name of any conjecture about unknown parameters of a population distribution. The hypothesis should be verifiable
More information4 Invariant Statistical Decision Problems
4 Invariant Statistical Decision Problems 4.1 Invariant decision problems Let G be a group of measurable transformations from the sample space X into itself. The group operation is composition. Note that
More informationBEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.
BEST TESTS Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. 1. Most powerful test Let {f θ } θ Θ be a family of pdfs. We will consider
More informationLecture 4 September 15
IFT 6269: Probabilistic Graphical Models Fall 2017 Lecture 4 September 15 Lecturer: Simon Lacoste-Julien Scribe: Philippe Brouillard & Tristan Deleu 4.1 Maximum Likelihood principle Given a parametric
More informationStatistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm
Statistics GIDP Ph.D. Qualifying Exam Theory Jan, 06, 9:00am-:00pm Instructions: Provide answers on the supplied pads of paper; write on only one side of each sheet. Complete exactly 5 of the 6 problems.
More information