Solutions. Name and surname: Instructions

Similar documents
First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise

Estimation for Complete Data

AMS570 Lecture Notes #2

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain

Mathematical Statistics - MS

Review Questions, Chapters 8, 9. f(y) = 0, elsewhere. F (y) = f Y(1) = n ( e y/θ) n 1 1 θ e y/θ = n θ e yn

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes.

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

Economics Spring 2015

Final Examination Statistics 200C. T. Ferguson June 10, 2010

Direction: This test is worth 150 points. You are required to complete this test within 55 minutes.

DS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10

Simulation. Two Rule For Inverting A Distribution Function

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d

Chapter 22. Comparing Two Proportions. Copyright 2010 Pearson Education, Inc.

University of California, Los Angeles Department of Statistics. Practice problems - simple regression 2 - solutions

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

STA Learning Objectives. Population Proportions. Module 10 Comparing Two Proportions. Upon completing this module, you should be able to:

Topic 9: Sampling Distributions of Estimators

STATISTICAL INFERENCE

Statistics 20: Final Exam Solutions Summer Session 2007

of the matrix is =-85, so it is not positive definite. Thus, the first

Monte Carlo Integration

Lecture 7: Properties of Random Samples

5. Fractional Hot deck Imputation

TMA4245 Statistics. Corrected 30 May and 4 June Norwegian University of Science and Technology Department of Mathematical Sciences.

Chapter 22. Comparing Two Proportions. Copyright 2010, 2007, 2004 Pearson Education, Inc.

Random Variables, Sampling and Estimation

Bayesian Methods: Introduction to Multi-parameter Models

Parameter, Statistic and Random Samples

Lecture 2: Monte Carlo Simulation

KLMED8004 Medical statistics. Part I, autumn Estimation. We have previously learned: Population and sample. New questions

Introductory statistics

TAMS24: Notations and Formulas

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

Econ 371 Exam #1. Multiple Choice (5 points each): For each of the following, select the single most appropriate option to complete the statement.

Properties and Hypothesis Testing

Statistical and Mathematical Methods DS-GA 1002 December 8, Sample Final Problems Solutions

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

On stratified randomized response sampling

Read through these prior to coming to the test and follow them when you take your test.

January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS

Lecture 33: Bootstrap

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

AAEC/ECON 5126 FINAL EXAM: SOLUTIONS

10-701/ Machine Learning Mid-term Exam Solution

7-1. Chapter 4. Part I. Sampling Distributions and Confidence Intervals

A Relationship Between the One-Way MANOVA Test Statistic and the Hotelling Lawley Trace Test Statistic

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

NANYANG TECHNOLOGICAL UNIVERSITY SYLLABUS FOR ENTRANCE EXAMINATION FOR INTERNATIONAL STUDENTS AO-LEVEL MATHEMATICS

Kurskod: TAMS11 Provkod: TENB 21 March 2015, 14:00-18:00. English Version (no Swedish Version)

Common Large/Small Sample Tests 1/55

Machine Learning Brett Bernstein

Introduction to Econometrics (3 rd Updated Edition) Solutions to Odd- Numbered End- of- Chapter Exercises: Chapter 3

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Mathematical Notation Math Introduction to Applied Statistics

MATH 472 / SPRING 2013 ASSIGNMENT 2: DUE FEBRUARY 4 FINALIZED

Stat 139 Homework 7 Solutions, Fall 2015

Chapter 8: STATISTICAL INTERVALS FOR A SINGLE SAMPLE. Part 3: Summary of CI for µ Confidence Interval for a Population Proportion p

(all terms are scalars).the minimization is clearer in sum notation:

Monte Carlo method and application to random processes

STAT 515 fa 2016 Lec Sampling distribution of the mean, part 2 (central limit theorem)

Element sampling: Part 2

Mathacle. PSet Stats, Concepts In Statistics Level Number Name: Date: Confidence Interval Guesswork with Confidence

ORF 245 Fundamentals of Engineering Statistics. Final Exam

In this section we derive some finite-sample properties of the OLS estimator. b is an estimator of β. It is a function of the random sample data.

Stat 400: Georgios Fellouris Homework 5 Due: Friday 24 th, 2017

If, for instance, we were required to test whether the population mean μ could be equal to a certain value μ

Chapter 6 Sampling Distributions

MBACATÓLICA. Quantitative Methods. Faculdade de Ciências Económicas e Empresariais UNIVERSIDADE CATÓLICA PORTUGUESA 9. SAMPLING DISTRIBUTIONS

Expectation and Variance of a random variable

Lecture 6: Coupon Collector s problem

Chapter 6 Principles of Data Reduction

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

Confidence Level We want to estimate the true mean of a random variable X economically and with confidence.

Lecture 3. Properties of Summary Statistics: Sampling Distribution

Topic 9: Sampling Distributions of Estimators

Economics 250 Assignment 1 Suggested Answers. 1. We have the following data set on the lengths (in minutes) of a sample of long-distance phone calls

CSE 527, Additional notes on MLE & EM

Stat 319 Theory of Statistics (2) Exercises

Topic 9: Sampling Distributions of Estimators

STAT 350 Handout 19 Sampling Distribution, Central Limit Theorem (6.6)

Massachusetts Institute of Technology

Final Examination Solutions 17/6/2010

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA STATISTICAL THEORY AND METHODS PAPER I

6 Sample Size Calculations

Solutions: Homework 3

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

Questions and Answers on Maximum Likelihood

Stat 421-SP2012 Interval Estimation Section

Understanding Dissimilarity Among Samples

IIT JAM Mathematical Statistics (MS) 2006 SECTION A

4. Partial Sums and the Central Limit Theorem

Read carefully the instructions on the answer book and make sure that the particulars required are entered on each answer book.

STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. Comments:

STATS 200: Introduction to Statistical Inference. Lecture 1: Course introduction and polling

Transcription:

Uiversity of Ljubljaa, Faculty of Ecoomics Quatitative fiace ad actuarial sciece Probability ad statistics Writte examiatio September 4 th, 217 Name ad surame: Istructios Read the problems carefull before starig your work There are four problems You ca have a sheet with formulae ad a mathematical hadbook Write your solutios o the paper provided You have two hours ID: Problem a b c d 1 2 3 4 Total Solutios

1 (25) Ofte it is difficult to obtai hoest aswers from sample subjects to questios like Have you ever used heroi or Have you ever cheated o a exam To reduce bias the method or radomized respose is used The sample subject is give oe of the two statemets below at radom: (1) I have property A (2) I do ot have property A The subject respods YES or NO to the give questio The pollster does ot kow to which of the two statemets the subject is respodig We assume: Let The subjects are a simple radom sample of size from a larger populatio of size N The statemets are assiged to the chose subjects idepedetly The assigmet of statemets is idepedet of the samplig procedure The subjects respod hoestly to the statemets they are give p be the probability the a subject will be assiged the statemet (1) This probability is kow ad is part of the desig q be the proportio of subjects i the populatio with property A r be the probability that a radomly selected subject respods YES to the statemet assiged R be the proportio of subjects i the sample who respod YES a (1) Compute the probability r that a radomly selected subject i the populatio respods YES to the statemet assiged Show that R is a ubiased estimate of r Take ito accout that the assigmet of statemets is idepedet of the selectio procedure Solutio: The probability that a radomly selected subject i the has property A is q The assigmet of statemets is radom ad idepedet of the samplig procedure which meas that the subject is assiged statemet (1) with probability p idepedetly from whether it has property A or ot Sice we assume that the subjects give a hoest aswer, the coditioal probability that the subject respods YES give that she is assiged statemet (i) is q Similarly if the subject is assiged statemet (2) the coditioal probability the she will repod YES is 1 q By the law of total probabilities we have that r = pq+(1 p)(1 q) = 1 p q + 2pq To see that R is a ubiased estimator of r write R = 1 i=1 I iwhere I i is the idicator of the evet that the i-th subject i the sample respods YES Above we have show that E(I i ) = P (I i = 1) = r which proves the claim b (5) Suggest a ubiased estimator of q Whe is this possible? Express the variace of the estimator with var(r) Solutio: We have q = p + r 1 2p 1, 2

which suggests that Q = p + R 1 2p 1 is a plausible choice We have show that R is a ubiased estimator of r By liearity of expectatio Q is ubiased The above is oly possible if p 1/2 If p = 1/2 we ote that a give subject will repod YES with probability 1/2 irrespective of the value of q Whatever expressio we take its expectatio will ot deped o q which meas that it caot be a ubiased estimator For the variace we compute var(q) = var(r) (2p 1) 2 c (1) Let N Y be the umber of sample subjects who respod YES to the questio ad N A a umber of sample subjects who have property A Assume as kow that var(n Y ) = p(1 p) + (2p 1) 2 var (N A ) Compute var(r) Use this to give the stadard error of the ubiased estimator of q Hit: what quatity does N A / estimate? Solutio: By formulae for simple radom samplig we have var( A ) = N q(1 q), N 1 Takig ito accout that R = Y / we fially get var(r) = p(1 p) + N N 1 From this the stadard error is derived easily (2p 1) 2 q(1 q) 3

2 (25) Assume the observed values x 1, x 2,, x were geerated as radom variables X 1, X 2,, X with desity for x, µ > f(x) = 1 (1 µx)2 e 2x 2πx 3 a (5) Fid the maximum likelihood estimate of µ Rešitev: Zapišemo logaritemsko fukcijo verjetja kot l = 2 log 2π 3 (1 µx k ) 2 log x k 2 2x k Odvajaje po µ am da eačbo (1 µx k ) = Iskaa ceilka je torej ˆµ = x 1 + x 2 + + x = 1 x b (5) Ca you fix the maximum likelihood estimator to be ubiased? Assume as kow: The desity of X 1 + + X is for x > f (x) = ( µx)2 e 2x 2πx 3 Assume as kow that for a, b > we have x 5/2 e ax b x dx = π ( 1 + 2 ab ) Rešitev: Naj ima X gostoto f (x) Račuamo ( ) 1 E = X x f (x) dx = 2 eµ 2π 2b 3/2 x 5/2 e µ2 2 x 2 2x dx = 2 eµ 1 + µ 2π e µ 2π 3 = µ + 1 e 2 ab 4

Nepristraska ceilka bi bila torej µ = 1 X 1 c (1) Compute the variace of the maximum likelihood estimator of µ Assume as kow that for a, b > we have ( ) π 3 + 6 ab + 4ab x 7/2 e ax b x dx = e 2 ab 4b 5/2 Rešitev: Za slučajo spremeljivko X z gostoto f (x) račuamo ( ) 2 2 E = X 2 x f (x) dx 2 Variaca je torej eaka = 3 eµ 2π x 7/2 e µ2 2 x 2 2x dx = 3 eµ 2π(3 + 3µ + 2 µ 2 ) e µ 2π 5 = 3 2 + 3µ + µ2 var(ˆµ) = E(ˆµ 2 ) ( E(ˆµ) ) 2 = µ + 2 2 d (5) What approximatio the the stadard error of the maximum likelihood estimator do we get if we use the Fisher iformatio? Assume as kow that x 1/2 e ax b x dx = π a e 2 ab Rešitev: Z odvajajem dobimo, da za = 1 velja Sledi, da je l = x I(µ) = E(X) = eµ 1 e µ2 x 2 1 2x dx 2π x = eµ 2π 2πµe µ = 1 µ 5

Aproksimacija variace po Fisherju je torej µ, kar je vodili čle v izrazu za pravo variaco 6

3 (25) Gauss s gamma distributio is give by the desity 2λ f(x, y) = π y e y e λy(x µ) 2 2 for < x < ad y > ad (µ, λ) R (, ) Assume that the observatios are pairs (x 1, y 1 ), (x 2, y 2 ),, (x, y ) geerated as idepedet radom pairs (X 1, Y 1 ),, (X, Y ) with desity f(x, y) We would like to test H : µ = versus H 1 : µ a (15) Compute the maximum likelihood estimates of the parameters Compute the maximum likelihood estimate of λ whe µ = Rešitev: Logaritemska fukcija verjetja je l = ( ) 2λ 2 log + π (log y k y k ) λ 2 y k (x k µ) 2 Parciale odvode izeačimo z i dobimo eačbi 2λ 1 2 y k (x k µ) 2 = i Iz druge eačbe sledi, da je λ y k (x k µ) = ˆµ = x ky k y k Vstavimo lahko v prvo eačbo i sledi ˆλ = y k(x k ˆµ) 2 V primeru, ko privzamemo µ =, je ustreza ceilka določea prav s prej omejeo prvo eačbo Vajo torej vstavimo µ = i dobimo ceilko λ = x2 k y k b (1) Fid the likelihood ratio statistics for the above testig problem What is its approximate distributio uder H? 7

Rešitev: Testa statistika je eaka [ ] λ = 2 l(ˆλ, ˆµ x, y) l( λ, x, y) = (log ˆλ log λ) ˆλ y k (x k ˆµ) 2 + λ x 2 ky k Toda iz eačb, iz katerih smo dobili ceilke, sledi ˆλ y k (x k ˆµ) 2 = λ x 2 ky k =, torej je kar λ = log ˆλ λ Po Wilksovem izreku je aproksimativa porazdelitev te statistike, če H eaka χ 2 (1) drži, 8

4 (25) Assume the regressio model Y = Xβ + ɛ, where E(ɛ) = ad var(ɛ) = σ 2 V Assume V is a ivertible symmetric matrix Assume that X is of the form 1 x 1 1 x 2 X = 1 x with x k =, ad the kow matrix V of the form with ρ 1/( 1) ad ρ 1 V = (1 ρ)i + ρ11 T a (5) Show that ˆβ = ( X T V 1 X ) 1 X T V 1 Y is the best ubiased liear estimator of β Rešitev: Regresijsko eačbo pomožimo a obeh straeh z V 1/2 i ozačimo Ỹ = V 1/2 Y i podobo za X ter ɛ S tem regresijski model prevedemo a stadardo obliko i po izreku Gauss-Markova lahko zapišemo zgorjo ceilko, ki je ajboljša b (1) Show that the best liear ubiased estimator of β is just the usual ordiary least squares estimator that we get assumig that V = I Hit: the iverse matrix V 1 is of the form ai + b11 T Rešitev: Z možejem dobimo V ( ai + b11 T ) = a(1 ρ)i + aρ11 T + b(1 ρ)11 T + bρ11 T 11 T Poeostavimo v Sledi, da je Račuamo i podobo a(1 ρ)i + (aρ + b(1 ρ) + bρ)11 T V 1 = 1 1 ρ I ρ (1 ρ)(1 + ( 1)ρ) 11T X T V 1 X = ( ) a + b 2 a x2 k ( (a + b) X T V 1 Y = Y ) k a x ky k 9

Sledi ( ) Ȳ ˆβ = x ky k / x2 k To ceilko bi dobili v primeru, če bi bilo V = I c (1) Compute ( ) E ˆɛ 2 k Hit: for a arbitrary vector a we have a T a = Tr(aa T ), ad for matrices A ad B we have Tr(AB) = Tr(BA) Rešitev: Račuamo E ( ˆɛ 2 k ) = E (ˆɛ T ˆɛ ) = E ( )) T Tr (ˆɛˆɛ = Tr (var(ˆɛ)) = Tr ( σ 2 (I H)V(I H) ) = Tr ( σ 2 V(I H)(I H) ) = Tr ( σ 2 V(I H) ) = σ 2( Tr (V) Tr(VH) ), kjer je H = X ( X T X ) 1 X T Preprosto je videti, da je Tr (V) = Nadalje je Tr (VH) = Tr ( (1 ρ)h + ρ11 T H ) = 2(1 ρ) + ρ 1 T H1 = 2(1 ρ) + ρ 1 T X ( X T X ) 1 X T 1 = 2(1 ρ) + ρ ( ) ( ) 1 ( ) 1 x2 k = 2(1 ρ) + ρ Sledi ( ) E = ( 2)(1 ρ)σ 2 ˆɛ 2 k 1