SOLUTION FOR HOMEWORK 7, STAT p(x σ) = (1/[2πσ 2 ] 1/2 )e (x µ)2 /2σ 2.
|
|
- Harriet Logan
- 5 years ago
- Views:
Transcription
1 SOLUTION FOR HOMEWORK 7, STAT We have (for a general case) Denote p (x) p(x σ)/ σ. Then p(x σ) (1/[2πσ 2 ] 1/2 )e (x µ)2 /2σ 2. p (x σ) p(x σ) 1 (x µ)2 +. σ σ 3 Then E{ p (x σ) p(x σ) } σ 2 2σ 4 E(X µ) 2 + σ 6 E(X µ) 4 σ 2 2σ 2 + 3σ 4 /σ 6 2σ 2 : I(σ). Note that for an optimal ˆσ this implies that n 1/2 (ˆσ σ) L N(0, σ 2 /2). Now let θ σ 2 be the estimand. Then, using the Delta method, n 1/2 (ˆσ 2 σ 2 ) L N(0, (2σ) 2 (σ 2 /2)) N(0, 2σ 4 ). But we know (or you can calculate it) that I(θ) 1/[2σ 4 ], that is, we got the efficient estimator. 2. Let p(x θ) p(x θ), x (, ). Then [change of variable u x θ] I(θ) [p (x θ)] 2 p 1 (x θ)dx [p (u)] 2 p 1 (u)du C. In other words, the Fisher information for location family is constant. Thus, for an unbiased estimate θ n based on n observations we have E θ {( θ n θ) 2 } [nc] Consider p(x θ) θ 1 f(x/θ), θ > 0. Then p (x θ) (1/θ 2 )f(x/θ) θ 3 xf (x/θ). We conclude that I(θ) θ 2 θ 2 [f(x/θ) + xθ 1 f (x/θ)] dx θ 1 f(x/θ) 1
2 [make change of variable u x/θ so du dx/θ] θ 2 [f(u) + uf (u)] 2 du f(u) θ 2 [uf (u)/f(u) + 1] 2 f(u)du. 4. Here the idea is that if Y θz > 0 then ln(y ) ln(θ) + ln(z) becomes the location model for which, as we know from Problem 1, the Fisher information is constant. Now let us prove the wished assertion. The density is e ξ f(xe ξ ). Then I(ξ) [change of variable u xe ξ, so du e ξ dx] [ e xi f(xe ξ ) xe 2ξ f (xe ξ )] 2 dx e ξ f(xe ξ ) [f(u) + uf (u)] 2 du. f(u) As you see, the last expression does not depend on ξ. 5. Let us recall the idea of the Information inequality, 0 [ p θ (x) θ]dx p θ (x)dx. Then θ E θ{δ(x)} p θ (x)δ(x)dx p θ (x)(δ(x) g(θ)dx. Then we use Cauchy-Schwarz for p θ x) p θ (x) [ p θ (x)(δ(x) g(θ))]dx, and get the Information inequality. Now remember that we have equality in Cauchy-Schwarz iff p θ (x) p θ (x) C p θ (x)(δ(x) g(θ)) which yields δ(x) g(θ) + C p θ(x) p θ (x). 2
3 Finally E(δ(X) g(θ)) 2 (g(θ))2 I(θ), and this implies that C 2 [g (θ)] 2 /I 2 (θ). We conclude that δ(x) g(θ) + g (θ) I(θ) p θ(x)/ θ 6. Here X Norm(θ, 1), and then Further, g(θ) + g (θ) I(θ) ln(p θ(x))/ θ. p X X (a,b) θ (x) (2π)1/2 e (x θ)2 /2 b a (2π)1/2 e (u θ)2 /2 du. P θ (X (b ǫ, b) X (a, b)) F θ(b) F θ (b ǫ) F θ (b) F θ (a) F(b ǫ θ). F(a θ) Now, if θ then F(x θ) becomes small. It is known that F(x) e x2 2π x (1 + o x (1)), x Thus as θ. Also, from and (1) we get as θ. F(a θ) O(e2(b a)θ ) (1) P θ (X (b ǫ, b) X (a, b)) 1 F(b ǫ θ) O(e 2ǫθ ) 0 F(b ǫ θ) 7. Here P θ (X i x) θ x e θ /[x!(1 e θ )], x 1, 2,... We are seeking the maximum of L(x n, θ) ln(p θ (X x i )) x i ln(θ) nθ ln(x i!) n ln(1 e θ). i1 i1 i1 3
4 Set the derivative to zero, 0 L (X n, θ) X i /θ n ne θ i1 1 e θ ni1 x i n[1 + e θ θ 1 e n x θ i /θ n/(1 e θ ) 0. i1 We get x θ/(1 e θ ). Now consider a function We write f(θ) : θ 1 e θ. f (θ) (1 e θ ) θe θ (1 e θ ) 2 1 (1 + θ)e θ (1 e θ ) 2. As we see, f (θ) is positive for both small and large θ. Let us check the possibility of f (θ) 0. In this case e θ 1 + θ, and this occurs only if θ 0, and this case is not of interest to us. We conclude that f(θ) is increasing in θ and the solution x f(θ) is unique. 8. Here L(θ) : L(X n, θ) ln(p(x n θ)). Then [using Taylor s formula] L(θ 0 + 1/ n) L(θ 0 ) + (1/2)I(θ 0 ) n 1/2 L (θ 0 ) + [2n] 1 L (θ 0 ) + [6n 3/2 ] 1 L ( θ) + (1/2)I(θ 0 ). By the CLT By the LLN n 1/2 L (θ 0 ) n 1/2 n i1 p (X i θ) p(x i θ) L N(0, I(θ)). (2n) 1 L (θ 0 ) P (1/2)E θ0 { p (X θ 0 )p(x θ 0 ) (p (X θ 0 )) 2 } (1/2)I(θ p 2 0 ). (X θ 0 ) Also note that n 1 L ( θ) < C. These relations imply that [L(θ 0 + 1/n 1/2 ) L(θ 0 ) + (1/2)I(θ 0 )]/I 1/2 (θ 0 ) L N(0, 1). 9. Assume that f(x) is differentiable. Then ln(f(x)) ψ(x) and we get for the considered function that ψ (x) < 0. Further, f (x)/f(x) ψ (x). We need to ask about f(x) 0 as x. Then f (x) has the sign of ψ (x). Further, ψ (x) 4
5 cannot strictly positive or negative, so f(x) increases and then decreases. This yields one mode if we assume that ψ (x) < Let us consider just the case of one observation. Then L (x, θ) 0 implies f (x)/f(x) 0 because strictly monotone in θ implies unique root. If f (x)/f(x) is decreasing then L(x, θ) is convex (or the send derivative is positive). For n observations, L(X n, θ) n i1 ln(p(x l θ), and L (X n, θ) n i1 f (X i θ)/f(x i θ). Then you just repeat the above-provided arguments. 5
f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain
0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher
More information4 Invariant Statistical Decision Problems
4 Invariant Statistical Decision Problems 4.1 Invariant decision problems Let G be a group of measurable transformations from the sample space X into itself. The group operation is composition. Note that
More informationMath 362, Problem set 1
Math 6, roblem set Due //. (4..8) Determine the mean variance of the mean X of a rom sample of size 9 from a distribution having pdf f(x) = 4x, < x
More information5.2 Fisher information and the Cramer-Rao bound
Stat 200: Introduction to Statistical Inference Autumn 208/9 Lecture 5: Maximum likelihood theory Lecturer: Art B. Owen October 9 Disclaimer: These notes have not been subjected to the usual scrutiny reserved
More informationChapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1
Chapter 4 HOMEWORK ASSIGNMENTS These homeworks may be modified as the semester progresses. It is your responsibility to keep up to date with the correctly assigned homeworks. There may be some errors in
More informationMath 3338, Homework 4. f(u)du. (Note that f is discontinuous, whereas F is continuous everywhere.)
Math 3338, Homework 4. Define a function f by f(x) = { if x < if x < if x. Calculate an explicit formula for F(x) x f(u)du. (Note that f is discontinuous, whereas F is continuous everywhere.) 2. Define
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationChapter 3. Point Estimation. 3.1 Introduction
Chapter 3 Point Estimation Let (Ω, A, P θ ), P θ P = {P θ θ Θ}be probability space, X 1, X 2,..., X n : (Ω, A) (IR k, B k ) random variables (X, B X ) sample space γ : Θ IR k measurable function, i.e.
More informationECE 275B Homework # 1 Solutions Version Winter 2015
ECE 275B Homework # 1 Solutions Version Winter 2015 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2
More informationZ-estimators (generalized method of moments)
Z-estimators (generalized method of moments) Consider the estimation of an unknown parameter θ in a set, based on data x = (x,...,x n ) R n. Each function h(x, ) on defines a Z-estimator θ n = θ n (x,...,x
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationChapter 4: Asymptotic Properties of the MLE (Part 2)
Chapter 4: Asymptotic Properties of the MLE (Part 2) Daniel O. Scharfstein 09/24/13 1 / 1 Example Let {(R i, X i ) : i = 1,..., n} be an i.i.d. sample of n random vectors (R, X ). Here R is a response
More informationSOLUTION FOR HOMEWORK 4, STAT 4352
SOLUTION FOR HOMEWORK 4, STAT 4352 Welcome to your fourth homework. Here we begin the study of confidence intervals, Errors, etc. Recall that X n := (X 1,...,X n ) denotes the vector of n observations.
More informationMarch 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS
March 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS Abstract. We will introduce a class of distributions that will contain many of the discrete and continuous we are familiar with. This class will help
More informationFor iid Y i the stronger conclusion holds; for our heuristics ignore differences between these notions.
Large Sample Theory Study approximate behaviour of ˆθ by studying the function U. Notice U is sum of independent random variables. Theorem: If Y 1, Y 2,... are iid with mean µ then Yi n µ Called law of
More informationECE 275B Homework # 1 Solutions Winter 2018
ECE 275B Homework # 1 Solutions Winter 2018 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2 < < x n Thus,
More informationP n. This is called the law of large numbers but it comes in two forms: Strong and Weak.
Large Sample Theory Large Sample Theory is a name given to the search for approximations to the behaviour of statistical procedures which are derived by computing limits as the sample size, n, tends to
More informationAnnouncements. Topics: Homework:
Announcements Topics: - sections 7.3 (the definite integral +area), 7.4 (FTC), 7.5 (additional techniques of integration) * Read these sections and study solved examples in your textbook! Homework: - review
More informationLecture 4 September 15
IFT 6269: Probabilistic Graphical Models Fall 2017 Lecture 4 September 15 Lecturer: Simon Lacoste-Julien Scribe: Philippe Brouillard & Tristan Deleu 4.1 Maximum Likelihood principle Given a parametric
More informationReview and continuation from last week Properties of MLEs
Review and continuation from last week Properties of MLEs As we have mentioned, MLEs have a nice intuitive property, and as we have seen, they have a certain equivariance property. We will see later that
More informationHomework 10 (due December 2, 2009)
Homework (due December, 9) Problem. Let X and Y be independent binomial random variables with parameters (n, p) and (n, p) respectively. Prove that X + Y is a binomial random variable with parameters (n
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationHW4 : Bivariate Distributions (1) Solutions
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 7 Néhémy Lim HW4 : Bivariate Distributions () Solutions Problem. The joint probability mass function of X and Y is given by the following table : X Y
More informationAnswers to the 8th problem set. f(x θ = θ 0 ) L(θ 0 )
Answers to the 8th problem set The likelihood ratio with which we worked in this problem set is: Λ(x) = f(x θ = θ 1 ) L(θ 1 ) =. f(x θ = θ 0 ) L(θ 0 ) With a lower-case x, this defines a function. With
More informationConcave and Convex Functions 1
John Nachbar Washington University March 6, 2017 Concave and Convex Functions 1 1 Basic Definitions Definition 1 Let C R N be convex and let f : C R 1 (a) f is concave iff for any a, b C and any θ [0,
More informationStat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota
Stat 5102 Lecture Slides Deck 3 Charles J. Geyer School of Statistics University of Minnesota 1 Likelihood Inference We have learned one very general method of estimation: method of moments. the Now we
More informationLecture 1: Introduction
Principles of Statistics Part II - Michaelmas 208 Lecturer: Quentin Berthet Lecture : Introduction This course is concerned with presenting some of the mathematical principles of statistical theory. One
More informationChapter 4: Unconstrained nonlinear optimization
Chapter 4: Unconstrained nonlinear optimization Edoardo Amaldi DEIB Politecnico di Milano edoardo.amaldi@polimi.it Website: http://home.deib.polimi.it/amaldi/opt-15-16.shtml Academic year 2015-16 Edoardo
More informationConvergence Tests. Theorem. (the divergence test)., then the series u k diverges. k k. (b) If lim u = 0, then the series u may either converge
Convergence Tests We are now interested in developing tests to tell whether or not a series converges, without having to guess at its sum. The first such test is entirely a negative result. Theorem. (the
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More information2014/2015 Smester II ST5224 Final Exam Solution
014/015 Smester II ST54 Final Exam Solution 1 Suppose that (X 1,, X n ) is a random sample from a distribution with probability density function f(x; θ) = e (x θ) I [θ, ) (x) (i) Show that the family of
More informationAnnouncements. Topics: Homework:
Announcements Topics: - sections 7.4 (FTC), 7.5 (additional techniques of integration), 7.6 (applications of integration) * Read these sections and study solved examples in your textbook! Homework: - review
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationConcave and Convex Functions 1
John Nachbar Washington University March 27, 2018 Concave and Convex Functions 1 1 Basic Definitions. Definition 1. Let C R N be non-empty and convex and let f : C R. 1. (a) f is concave iff for any a,
More informationNotes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed
18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,
More informationFisher Information & Efficiency
Fisher Information & Efficiency Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA 1 Introduction Let f(x θ) be the pdf of X for θ Θ; at times we will also consider a
More information557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE
557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE An estimator, T (X), of θ can be evaluated via its statistical properties. Typically, two aspects are considered: Expectation Variance either in terms
More information40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology
Singapore University of Design and Technology Lecture 9: Hypothesis testing, uniformly most powerful tests. The Neyman-Pearson framework Let P be the family of distributions of concern. The Neyman-Pearson
More informationA Few Notes on Fisher Information (WIP)
A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More informationINTRODUCTION TO BAYESIAN METHODS II
INTRODUCTION TO BAYESIAN METHODS II Abstract. We will revisit point estimation and hypothesis testing from the Bayesian perspective.. Bayes estimators Let X = (X,..., X n ) be a random sample from the
More informationλ(x + 1)f g (x) > θ 0
Stat 8111 Final Exam December 16 Eleven students took the exam, the scores were 92, 78, 4 in the 5 s, 1 in the 4 s, 1 in the 3 s and 3 in the 2 s. 1. i) Let X 1, X 2,..., X n be iid each Bernoulli(θ) where
More informationStatistics Ph.D. Qualifying Exam: Part I October 18, 2003
Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer
More informationTheory of Statistical Tests
Ch 9. Theory of Statistical Tests 9.1 Certain Best Tests How to construct good testing. For simple hypothesis H 0 : θ = θ, H 1 : θ = θ, Page 1 of 100 where Θ = {θ, θ } 1. Define the best test for H 0 H
More informationMathematics 426 Robert Gross Homework 9 Answers
Mathematics 4 Robert Gross Homework 9 Answers. Suppose that X is a normal random variable with mean µ and standard deviation σ. Suppose that PX > 9 PX
More informationIntroduction to Estimation Methods for Time Series models Lecture 2
Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:
More informationLecture 17: Likelihood ratio and asymptotic tests
Lecture 17: Likelihood ratio and asymptotic tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f
More informationIntegration by Substitution
November 22, 2013 Introduction 7x 2 cos(3x 3 )dx =? 2xe x2 +5 dx =? Chain rule The chain rule: d dx (f (g(x))) = f (g(x)) g (x). Use the chain rule to find f (x) and then write the corresponding anti-differentiation
More informationRearrangement on Conditionally Convergent Integrals in Analogy to Series
Electronic Proceedings of Undergraduate Mathematics Day, Vol. 3 (8), No. 6 Rearrangement on Conditionally Convergent Integrals in Analogy to Series Edward J Timko University of Dayton Dayton OH 45469-36
More informationStatistics 300B Winter 2018 Final Exam Due 24 Hours after receiving it
Statistics 300B Winter 08 Final Exam Due 4 Hours after receiving it Directions: This test is open book and open internet, but must be done without consulting other students. Any consultation of other students
More informationStat 411 Lecture Notes 03 Likelihood and Maximum Likelihood Estimation
Stat 411 Lecture Notes 03 Likelihood and Maximum Likelihood Estimation Ryan Martin www.math.uic.edu/~rgmartin Version: August 19, 2013 1 Introduction Previously we have discussed various properties of
More informationSTA 260: Statistics and Probability II
Al Nosedal. University of Toronto. Winter 2017 1 Properties of Point Estimators and Methods of Estimation 2 3 If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition
More informationQualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf
Part : Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section
More information7 Influence Functions
7 Influence Functions The influence function is used to approximate the standard error of a plug-in estimator. The formal definition is as follows. 7.1 Definition. The Gâteaux derivative of T at F in the
More informationExam 3 review for Math 1190
Exam 3 review for Math 9 Be sure to be familiar with the following : Extreme Value Theorem Optimization The antiderivative u-substitution as a method for finding antiderivatives Reimann sums (e.g. L 6
More informationCIVL 7012/8012. Continuous Distributions
CIVL 7012/8012 Continuous Distributions Probability Density Function P(a X b) = b ò a f (x)dx Probability Density Function Definition: and, f (x) ³ 0 ò - f (x) =1 Cumulative Distribution Function F(x)
More informationECONOMICS 207 SPRING 2006 LABORATORY EXERCISE 5 KEY. 8 = 10(5x 2) = 9(3x + 8), x 50x 20 = 27x x = 92 x = 4. 8x 2 22x + 15 = 0 (2x 3)(4x 5) = 0
ECONOMICS 07 SPRING 006 LABORATORY EXERCISE 5 KEY Problem. Solve the following equations for x. a 5x 3x + 8 = 9 0 5x 3x + 8 9 8 = 0(5x ) = 9(3x + 8), x 0 3 50x 0 = 7x + 7 3x = 9 x = 4 b 8x x + 5 = 0 8x
More informationSOLUTION FOR HOMEWORK 6, STAT 6331
SOLUTION FOR HOMEWORK 6, STAT 633. Exerc.7.. It is given that X,...,X n is a sample from N(θ, σ ), and the Bayesian approach is used with Θ N(µ, τ ). The parameters σ, µ and τ are given. (a) Find the joinf
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More information1 Introduction. 2 Measure theoretic definitions
1 Introduction These notes aim to recall some basic definitions needed for dealing with random variables. Sections to 5 follow mostly the presentation given in chapter two of [1]. Measure theoretic definitions
More informationMath 172 Problem Set 8 Solutions
Math 72 Problem Set 8 Solutions Problem. (i We have (Fχ [ a,a] (ξ = χ [ a,a] e ixξ dx = a a e ixξ dx = iξ (e iax e iax = 2 sin aξ. ξ (ii We have (Fχ [, e ax (ξ = e ax e ixξ dx = e x(a+iξ dx = a + iξ where
More informationMISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30
MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)
More informationMath 0230 Calculus 2 Lectures
Math 00 Calculus Lectures Chapter 8 Series Numeration of sections corresponds to the text James Stewart, Essential Calculus, Early Transcendentals, Second edition. Section 8. Sequences A sequence is a
More informationChapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic
Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic Unbiased estimation Unbiased or asymptotically unbiased estimation plays an important role in
More informationStatistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That
Statistics Lecture 2 August 7, 2000 Frank Porter Caltech The plan for these lectures: The Fundamentals; Point Estimation Maximum Likelihood, Least Squares and All That What is a Confidence Interval? Interval
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationInference in non-linear time series
Intro LS MLE Other Erik Lindström Centre for Mathematical Sciences Lund University LU/LTH & DTU Intro LS MLE Other General Properties Popular estimatiors Overview Introduction General Properties Estimators
More informationSTAT 450: Final Examination Version 1. Richard Lockhart 16 December 2002
Name: Last Name 1, First Name 1 Stdnt # StudentNumber1 STAT 450: Final Examination Version 1 Richard Lockhart 16 December 2002 Instructions: This is an open book exam. You may use notes, books and a calculator.
More informationLecture 28: Asymptotic confidence sets
Lecture 28: Asymptotic confidence sets 1 α asymptotic confidence sets Similar to testing hypotheses, in many situations it is difficult to find a confidence set with a given confidence coefficient or level
More informationSTA 732: Inference. Notes 10. Parameter Estimation from a Decision Theoretic Angle. Other resources
STA 732: Inference Notes 10. Parameter Estimation from a Decision Theoretic Angle Other resources 1 Statistical rules, loss and risk We saw that a major focus of classical statistics is comparing various
More informationLecture 32: Asymptotic confidence sets and likelihoods
Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence
More informationStatistics 581 Revision of Section 4.4: Consistency of Maximum Likelihood Estimates Wellner; 11/30/2001
Statistics 581 Revision of Section 4.4: Consistency of Maximum Likelihood Estimates Wellner; 11/30/2001 Some Uniform Strong Laws of Large Numbers Suppose that: A. X, X 1,...,X n are i.i.d. P on the measurable
More informationMaximum Likelihood Estimation
Chapter 8 Maximum Likelihood Estimation 8. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function
More informationGraduate Econometrics I: Maximum Likelihood I
Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood
More informationn=0 ( 1)n /(n + 1) converges, but not
Math 07H Topics for the third exam (and beyond) (Technically, everything covered on the first two exams plus...) Absolute convergence and alternating series A series a n converges absolutely if a n converges.
More informationLecture 2: Basic Concepts of Statistical Decision Theory
EE378A Statistical Signal Processing Lecture 2-03/31/2016 Lecture 2: Basic Concepts of Statistical Decision Theory Lecturer: Jiantao Jiao, Tsachy Weissman Scribe: John Miller and Aran Nayebi In this lecture
More informationMaximum Likelihood Asymptotic Theory. Eduardo Rossi University of Pavia
Maximum Likelihood Asymtotic Theory Eduardo Rossi University of Pavia Slutsky s Theorem, Cramer s Theorem Slutsky s Theorem Let {X N } be a random sequence converging in robability to a constant a, and
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationSOLUTION FOR HOMEWORK 12, STAT 4351
SOLUTION FOR HOMEWORK 2, STAT 435 Welcome to your 2th homework. It looks like this is the last one! As usual, try to find mistakes and get extra points! Now let us look at your problems.. Problem 7.22.
More informationStat410 Probability and Statistics II (F16)
Stat4 Probability and Statistics II (F6 Exponential, Poisson and Gamma Suppose on average every /λ hours, a Stochastic train arrives at the Random station. Further we assume the waiting time between two
More informationwhere r n = dn+1 x(t)
Random Variables Overview Probability Random variables Transforms of pdfs Moments and cumulants Useful distributions Random vectors Linear transformations of random vectors The multivariate normal distribution
More informationMAC 2311 Calculus I Spring 2004
MAC 2 Calculus I Spring 2004 Homework # Some Solutions.#. Since f (x) = d dx (ln x) =, the linearization at a = is x L(x) = f() + f ()(x ) = ln + (x ) = x. The answer is L(x) = x..#4. Since e 0 =, and
More informationDifferentiation. Table of contents Definition Arithmetics Composite and inverse functions... 5
Differentiation Table of contents. Derivatives................................................. 2.. Definition................................................ 2.2. Arithmetics...............................................
More informationParametric Inference
Parametric Inference Moulinath Banerjee University of Michigan April 14, 2004 1 General Discussion The object of statistical inference is to glean information about an underlying population based on a
More informationDescent methods. min x. f(x)
Gradient Descent Descent methods min x f(x) 5 / 34 Descent methods min x f(x) x k x k+1... x f(x ) = 0 5 / 34 Gradient methods Unconstrained optimization min f(x) x R n. 6 / 34 Gradient methods Unconstrained
More informationExpectation Maximization (EM) Algorithm. Each has it s own probability of seeing H on any one flip. Let. p 1 = P ( H on Coin 1 )
Expectation Maximization (EM Algorithm Motivating Example: Have two coins: Coin 1 and Coin 2 Each has it s own probability of seeing H on any one flip. Let p 1 = P ( H on Coin 1 p 2 = P ( H on Coin 2 Select
More informationMath 152 Take Home Test 1
Math 5 Take Home Test Due Monday 5 th October (5 points) The following test will be done at home in order to ensure that it is a fair and representative reflection of your own ability in mathematics I
More informationJim Lambers MAT 169 Fall Semester Practice Final Exam
Jim Lambers MAT 169 Fall Semester 2010-11 Practice Final Exam 1. A ship is moving northwest at a speed of 50 mi/h. A passenger is walking due southeast on the deck at 4 mi/h. Find the speed of the passenger
More informationTopology Homework Assignment 1 Solutions
Topology Homework Assignment 1 Solutions 1. Prove that R n with the usual topology satisfies the axioms for a topological space. Let U denote the usual topology on R n. 1(a) R n U because if x R n, then
More informationMaximum Likelihood Estimation
Chapter 7 Maximum Likelihood Estimation 7. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More information1 Introduction to Optimization
Unconstrained Convex Optimization 2 1 Introduction to Optimization Given a general optimization problem of te form min x f(x) (1.1) were f : R n R. Sometimes te problem as constraints (we are only interested
More informationMatrix concentration inequalities
ELE 538B: Mathematics of High-Dimensional Data Matrix concentration inequalities Yuxin Chen Princeton University, Fall 2018 Recap: matrix Bernstein inequality Consider a sequence of independent random
More informationRegression Estimation Least Squares and Maximum Likelihood
Regression Estimation Least Squares and Maximum Likelihood Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 3, Slide 1 Least Squares Max(min)imization Function to minimize
More informationLecture 6 Ecient estimators. Rao-Cramer bound.
Lecture 6 Eciet estimators. Rao-Cramer boud. 1 MSE ad Suciecy Let X (X 1,..., X) be a radom sample from distributio f θ. Let θ ˆ δ(x) be a estimator of θ. Let T (X) be a suciet statistic for θ. As we have
More informationLecture 8 October Bayes Estimators and Average Risk Optimality
STATS 300A: Theory of Statistics Fall 205 Lecture 8 October 5 Lecturer: Lester Mackey Scribe: Hongseok Namkoong, Phan Minh Nguyen Warning: These notes may contain factual and/or typographic errors. 8.
More informationContinuity. Chapter 4
Chapter 4 Continuity Throughout this chapter D is a nonempty subset of the real numbers. We recall the definition of a function. Definition 4.1. A function from D into R, denoted f : D R, is a subset of
More informationInterval Estimation. Chapter 9
Chapter 9 Interval Estimation 9.1 Introduction Definition 9.1.1 An interval estimate of a real-values parameter θ is any pair of functions, L(x 1,..., x n ) and U(x 1,..., x n ), of a sample that satisfy
More informationSolutions: Homework 8
Solutions: Homework 8 1 Chapter 7.1 Problem 3 This is one of those problems where we substitute in solutions and just check to see if they work or not. (a) Substituting y = e rx into the differential equation
More informationLECTURE 7 NOTES. x n. d x if. E [g(x n )] E [g(x)]
LECTURE 7 NOTES 1. Convergence of random variables. Before delving into the large samle roerties of the MLE, we review some concets from large samle theory. 1. Convergence in robability: x n x if, for
More information