January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS

Similar documents
Parameter, Statistic and Random Samples

Estimation for Complete Data

7.1 Convergence of sequences of random variables

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

Lecture 2: Monte Carlo Simulation

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS

The standard deviation of the mean

Distribution of Random Samples & Limit theorems

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

7.1 Convergence of sequences of random variables

SOME THEORY AND PRACTICE OF STATISTICS by Howard G. Tucker

Convergence of random variables. (telegram style notes) P.J.C. Spreij

This section is optional.

Random Variables, Sampling and Estimation

Chapter 6 Principles of Data Reduction

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes.

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

Lecture 3: MLE and Regression

Lecture 19: Convergence

Lecture 01: the Central Limit Theorem. 1 Central Limit Theorem for i.i.d. random variables

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.

STAT Homework 2 - Solutions

Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman:

Math 525: Lecture 5. January 18, 2018

Mathematical Statistics - MS

4. Partial Sums and the Central Limit Theorem

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 22

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 19

SDS 321: Introduction to Probability and Statistics

EE 4TM4: Digital Communications II Probability Theory

Please do NOT write in this box. Multiple Choice. Total

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

Lecture 20: Multivariate convergence and the Central Limit Theorem

Direction: This test is worth 150 points. You are required to complete this test within 55 minutes.

STATISTICAL INFERENCE

Machine Learning Brett Bernstein

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1

Stat410 Probability and Statistics II (F16)

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Review Questions, Chapters 8, 9. f(y) = 0, elsewhere. F (y) = f Y(1) = n ( e y/θ) n 1 1 θ e y/θ = n θ e yn

Notes 5 : More on the a.s. convergence of sums

Lecture 2: Concentration Bounds

Advanced Stochastic Processes.

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

An Introduction to Asymptotic Theory

Stat 319 Theory of Statistics (2) Exercises

Lecture 6 Simple alternatives and the Neyman-Pearson lemma

Expectation and Variance of a random variable

HOMEWORK I: PREREQUISITES FROM MATH 727

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence

Mathematics 170B Selected HW Solutions.

ST5215: Advanced Statistical Theory

Exponential Families and Bayesian Inference

Statistical Theory MT 2009 Problems 1: Solution sketches

Solutions: Homework 3

6.3 Testing Series With Positive Terms

Random Matrices with Blocks of Intermediate Scale Strongly Correlated Band Matrices

Discrete Mathematics and Probability Theory Summer 2014 James Cook Note 15

Learning Theory: Lecture Notes

Parameter, Statistic and Random Samples

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett

Clases 7-8: Métodos de reducción de varianza en Monte Carlo *

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

STAT Homework 1 - Solutions

6. Sufficient, Complete, and Ancillary Statistics

Lecture Chapter 6: Convergence of Random Sequences

Lecture 7: Density Estimation: k-nearest Neighbor and Basis Approach

Homework for 2/3. 1. Determine the values of the following quantities: a. t 0.1,15 b. t 0.05,15 c. t 0.1,25 d. t 0.05,40 e. t 0.

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise

Lecture 12: September 27

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013. Large Deviations for i.i.d. Random Variables

An Introduction to Randomized Algorithms

Statistical Theory MT 2008 Problems 1: Solution sketches

Lecture 7: Properties of Random Samples

MATH 472 / SPRING 2013 ASSIGNMENT 2: DUE FEBRUARY 4 FINALIZED

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

The Random Walk For Dummies

Topic 9: Sampling Distributions of Estimators

Randomized Algorithms I, Spring 2018, Department of Computer Science, University of Helsinki Homework 1: Solutions (Discussed January 25, 2018)

Math 152. Rumbos Fall Solutions to Review Problems for Exam #2. Number of Heads Frequency

LECTURE 8: ASYMPTOTICS I

Confidence Level We want to estimate the true mean of a random variable X economically and with confidence.

n outcome is (+1,+1, 1,..., 1). Let the r.v. X denote our position (relative to our starting point 0) after n moves. Thus X = X 1 + X 2 + +X n,

32 estimating the cumulative distribution function

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall Midterm Solutions

Statistical Inference Based on Extremum Estimators

Unbiased Estimation. February 7-12, 2008

Monte Carlo Integration

STA Object Data Analysis - A List of Projects. January 18, 2018

Final Review for MATH 3510

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain

1 = δ2 (0, ), Y Y n nδ. , T n = Y Y n n. ( U n,k + X ) ( f U n,k + Y ) n 2n f U n,k + θ Y ) 2 E X1 2 X1

MATH/STAT 352: Lecture 15

Chapter 5. Inequalities. 5.1 The Markov and Chebyshev inequalities

April 18, 2017 CONFIDENCE INTERVALS AND HYPOTHESIS TESTING, UNDERGRADUATE MATH 526 STYLE

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator

( θ. sup θ Θ f X (x θ) = L. sup Pr (Λ (X) < c) = α. x : Λ (x) = sup θ H 0. sup θ Θ f X (x θ) = ) < c. NH : θ 1 = θ 2 against AH : θ 1 θ 2

Problem Set 4 Due Oct, 12

Transcription:

Jauary 25, 207 INTRODUCTION TO MATHEMATICAL STATISTICS Abstract. A basic itroductio to statistics assumig kowledge of probability theory.. Probability I a typical udergraduate problem i probability, we are told or implicity give the distributio of a radom variable X ad ask to compute certai quatities. I more advaced courses, we are asked to prove certai assertios about X. Exercise. Let p (0, ). Let X Ber(p), so that P(X = ) = p ad P(X = 0) = p. For which value of p is the variace maximized? Exercise 2. Suppose T is the umber of flips of a fair coi that are required for a head to appear. Fid ET. Exercise 3. Suppose that X = (X, X 2 ) are two cards dealt from a well-shuffled stadard 52 card deck. What is the probability that I get a pair? Exercise 4. Let X, X 2,... be a sequece of i.i.d. radom variables all with mea µ ad variace σ 2 <. Fid lim Xi 2. Although, the task may seem simple, problems that are easy to state may ot be so simple to solve. Let X, X 2,..., be a sequece of i.i.d. discrete radom variables that are uiformly distributed o {e,..., e }, where the e i are the stadard uit basis vectors i Z d so that e = (, 0,..., 0) ad e d = (0, 0,..., ). Let S = X + + X. The sequece (S ) is called a simple symmtric radom walk o Z d. Set T = if { : S = 0 Z d}. We say that that (S ) is recurret if P(T < ) =. Theorem 5 (Polya). The simple symmetric radom walk o Z d is recurret for dimesios d =, 2 ad ot recurret for all dimesios d 3. i=

2 INTRODUCTION TO MATHEMATICAL STATISTICS I statistics, we are iterested i the dual problem of fidig the distributio of X; we wat to kow whether a coi is actually fair or ot. Typically this ivovles observig may istaces of X; that is flippig the coi may times. Based o our observatios, we make a iferece about the distributio. 2. Radom samples ad poit estimators Suppose that f θ is a probability desity fuctio for each θ Θ, for a discrete of coituous radom variable. Here, Θ is usually a subset of R or R d. Let θ Θ. We say that X = (X,..., X ) is a radom sample from f θ if X,..., X are i.i.d. radom variables all with the same pdf f θ. I a typical statistics problem we do ot kow the value of θ, but we do kow that θ Θ ad we observe the values of a radom sample ad hope that it will tell us somethig about θ. If T = g(x) for some determiisitic fuctio, the we say that T is a statistic. If T does ot deped o θ, ad we use it to estimate θ, the we say that it is a poit estimator. If we observe the value T = t, the we say that t is a poit estimate. I udergraduate statistics course we were already itroduced to two importat poit estimators. Let X,..., X be a radom sample. The sample mea is give by X = X i. i= The sample variace is give by S 2 = (X i X) 2. i= A poit estimate T of θ is ubiased if ET = θ for all θ Θ. A sequece of poit estimates T = T are cosistet if T coverges i probability to θ, as, for all θ Θ. Whe covergece i probability is replaced with the stroger almost sure covergece, sometimes the termiology strogly cosistet is used. Exercise 6. Show that the sample mea ad variace are ubiased ad cosistet estimators for the true mea ad variace. Exercise 7. Let X,..., X be a radom sample from the Beroulli family; that is, X i Ber(p), where p (0, ). Let σ 2 be the variace of X. Cosider the poit estimate for σ 2 give by X( X). Is it ubiased? Is it cosistet?

INTRODUCTION TO MATHEMATICAL STATISTICS 3 Exercise 8. Suppose X,..., X is a radom sample, where X i U(0, θ), for some θ (0, ), so that X i are uiformly distributed i (0, θ). Cosider M := max {X,..., X } is a poit estimator for θ. Is it ubiased? Is it cosistet? 3. Sample Variace I this sectio, we will work out Exercise 6. We kow that E( X) = µ ad the cosistecy of X follows from the law of large umbers. Some algebra will give the result for S 2. First, set σ 2 := Var(X ) ad µ := EX. A short-cut formula gives that Var(Y ) := E(Y EY ) 2 = EY 2 (EY ) 2, from which if follows that EX 2 = σ 2 + µ 2. Secod, ote that (X i X) 2 = Xi 2 2X i X + X2 = Xi 2 2 X i X j 2 X2 i + X 2. j=,j i Third, takig expectatios, usig the idepedece of the X i, ad summig over i, gives ( )ES 2 = (σ 2 + µ 2 ) 2( )µ 2 2(σ 2 + µ 2 ) + E( X 2 ). Fially, recall that Var( X) = σ2, so that E( X 2 ) = σ 2 / + µ 2. Some algebra gives ( )ES 2 = ( )σ 2, so we are doe. To show cosistecy, first, recall the followig short-cut formula. Exercise 9. Let x,... x R ad x = / i= x i. Prove the socalled short-cut formula. (x i x) 2 = x 2 i ( ) 2. x i i= The Exercise 9 gives that ( )S 2 = i= i= Xi 2 ( ) 2. X i Fially, ote that the law of large umbers gives that Xi 2 EX 2 = σ 2 + µ 2 i= i= i=

4 INTRODUCTION TO MATHEMATICAL STATISTICS ad ( ) 2 X i µ 2 i= almost surely, as. So we are doe. 4. Method of momets The momets of a radom variable X are give by µ k := EX k. Uder certai coditios, if all the momets of X are fiite, the they determie the distributio for X. If X,..., X is a radom sample from f θ, the momet estimators are give by M k := Xi k. Exercise 0. Show that the momet estimators are ubiased ad cosistet estimators for µ k. If θ = (θ,..., θ 9 ), the it possible that if we compute at least 9 of momets, the we might be able to have equatios for each the θ i i terms of the momets, which we ca estimate usig the momet estimators. We call this method of fidig estimators the method of momets ad estimator obtaied usig this method are sometimes called the method of momet estimators. i= Exercise. Let X,..., X be a radom sample, where X i Uif(0, θ), where θ is ukow. Use the method of momets to fid a estimator for θ. Solutio. Note that EX = θ/2, so that θ = 2µ. Thus the method of momet estimator for θ is give by 2 X. Exercise 2. Suppose that X Bi(, p). Solve for ad p i terms of µ ad µ 2. Solutio. We have that µ = EX = p ad µ 2 = EX 2 = Var(X ) + (p) 2 = p( p) + (p) 2. Replacig p with µ / i the expressio for µ 2 gives µ 2 = µ µ 2 / + µ 2, from which it follows that = µ 2 /(µ + µ 2 µ 2 ). Exercise 3. Let X,..., X k be a radom sample, where X i Bi(, p), where both ad p are ukow. Use the method of momets to fid estimators for ad p.

INTRODUCTION TO MATHEMATICAL STATISTICS 5 Solutio. Notice that the short-cut formula gives that k (X i k X) 2 = k Xi 2 ( k ) 2. X i k k i= Thus from the previous exercise ad the short-cut formula, we have that the estimator for is give by N := ( X) 2 / ( X k k S2) ad so the estimator of p is give by X/N. Sometimes, rather tha cosiderig the secod momet i= µ 2 = EX 2 it is easier (ad equivalet) to cosider i= µ 2 = E(X µ ) 2 = Var(X ) = σ 2, which has a method of momet estimator give by S2 = (X i X) 2. i= Agai, this follows from the short-cut formula. Exercise 4. Let X,..., X be a radom sample, where X Gamma(α, β), where both α, β > 0 are ukow. Fid the method of momet estimators for α ad β. Solutio. Recall that the pdf X is give by f(x ; α, β) = β α Γ(α) xα e x/β [x > 0], ad µ = E(X ) = αβ ad σ 2 = Var(X ) = αβ 2. α = µ/β ad β 2 = σ 2 /α = σ 2 β/µ so that β = σ 2 /µ. Thus the method of momet estimators for β ad α are give by B = S2 X ad 2 ( X) A = S. 2

6 INTRODUCTION TO MATHEMATICAL STATISTICS Exercise 5. Let X,..., X be a radom sample, where X i Uif(a, b), where both a ad b are ukow. Fid the method of momet estimators for a ad b.