Statistics 135 Fall 2007 Midterm Exam

Similar documents
Stat 135 Fall 2013 FINAL EXAM December 18, 2013

Margin of Error for Proportions

STAT 135 Lab 3 Asymptotic MLE and the Method of Moments

This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

Statistics 135 Fall 2008 Final Exam

Practice Exam 1. (A) (B) (C) (D) (E) You are given the following data on loss sizes:

Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices.

Math 494: Mathematical Statistics

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

Binomial and Poisson Probability Distributions

STAB57: Quiz-1 Tutorial 1 (Show your work clearly) 1. random variable X has a continuous distribution for which the p.d.f.

Point Estimation and Confidence Interval

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics Ph.D. Qualifying Exam: Part II November 3, 2001

Mathematical statistics

STAT 516 Midterm Exam 2 Friday, March 7, 2008

Statistics 135: Fall 2004 Final Exam

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

Statistics 3858 : Maximum Likelihood Estimators

Parameter Estimation

AMCS243/CS243/EE243 Probability and Statistics. Fall Final Exam: Sunday Dec. 8, 3:00pm- 5:50pm VERSION A

Stat 5102 Final Exam May 14, 2015

Some Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2

Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2

You may use your calculator and a single page of notes.

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

STAT 461/561- Assignments, Year 2015

Estimation and Confidence Intervals

STAT 135 Lab 2 Confidence Intervals, MLE and the Delta Method

The Random Variable for Probabilities Chris Piech CS109, Stanford University

Point Estimation. Vibhav Gogate The University of Texas at Dallas

(4) One-parameter models - Beta/binomial. ST440/550: Applied Bayesian Statistics

18.05 Practice Final Exam

Likelihood and Bayesian Inference for Proportions

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1

Review. December 4 th, Review

A proportion is the fraction of individuals having a particular attribute. Can range from 0 to 1!

MAS223 Statistical Inference and Modelling Exercises

BTRY 4090: Spring 2009 Theory of Statistics

Statistics. Statistics

ST 371 (IX): Theories of Sampling Distributions

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

How to Use the Internet for Election Surveys

MA 320 Introductory Probability, Section 1 Spring 2017 Final Exam Practice May. 1, Exam Scores. Question Score Total. Name:

ACMS Statistics for Life Sciences. Chapter 9: Introducing Probability

Probability & Statistics - FALL 2008 FINAL EXAM

Eco517 Fall 2004 C. Sims MIDTERM EXAM

Exam C Solutions Spring 2005

Likelihood and Bayesian Inference for Proportions

CSC321 Lecture 18: Learning Probabilistic Models

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages

Discrete Distributions

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS

Master s Written Examination

CSE 312: Foundations of Computing II Quiz Section #10: Review Questions for Final Exam (solutions)

Loglikelihood and Confidence Intervals

CS 361 Sample Midterm 2

Spring 2012 Math 541B Exam 1

APPM/MATH 4/5520 Solutions to Exam I Review Problems. f X 1,X 2. 2e x 1 x 2. = x 2

GALLUP NEWS SERVICE JUNE WAVE 1

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing

Lecture 25. Ingo Ruczinski. November 24, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University

Review of Discrete Probability (contd.)

STAT 285: Fall Semester Final Examination Solutions

Estimation of Quantiles

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Introduction to Machine Learning. Lecture 2

1 Introduction. P (n = 1 red ball drawn) =

EXAM # 2. Total 100. Please show all work! Problem Points Grade. STAT 301, Spring 2013 Name

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices.

PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Content by Week Week of October 14 27

Probabilistic Graphical Models

Page Max. Possible Points Total 100

6.867 Machine Learning

# of 6s # of times Test the null hypthesis that the dice are fair at α =.01 significance

Statistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm

HT Introduction. P(X i = x i ) = e λ λ x i

Estimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1

Department of Statistical Science FIRST YEAR EXAM - SPRING 2017

Probability and Estimation. Alan Moses

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Lecture 17: The Exponential and Some Related Distributions

First Year Examination Department of Statistics, University of Florida

What does a population that is normally distributed look like? = 80 and = 10

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

This does not cover everything on the final. Look at the posted practice problems for other topics.

Central Limit Theorem ( 5.3)

The comparative studies on reliability for Rayleigh models

1. Fisher Information

Imputation of rounded data

Ph.D. Preliminary Examination Statistics June 2, 2014

The Bayesian Paradigm

Institute of Actuaries of India

Probability Theory for Machine Learning. Chris Cremer September 2015

Transcription:

Name: Student ID Number: Statistics 135 Fall 007 Midterm Exam Ignore the finite population correction in all relevant problems. The exam is closed book, but some possibly useful facts about probability distributions are listed on the last page. Show your work in answering all questions. There are six questions. 1. True or False and Why? (1 point each) (a) The mean of a simple random sample is a random variable. T. The sample values are randomly determined and hence their average is random (b) The population variance is a random variable. F. This is a population parameter (c) The estimated standard error of the mean from a simple random sample is a random variable. T. It is a function of the sample values, which are random (d) A 90% confidence interval for a mean, based on a simple random sample, contains 90% of the population values with probability 0.90. F. It contains the population mean with 90% probability (e) A 90% confidence interval for mean, based on a simple random sample of size 50, would be about twice the length of an interval based on a sample of size 100. F. The width is inversely proportional to the square root of the sample size. 1

(f) A histogram of the values in a simple random sample of size 1000 should look approximately like a normal distribution. F. It will look like the population histogram (g) As the number of bootstrap replications gets larger and larger, the bootstrap estimate of the standard error of a parameter estimate will get closer and closer to the actual standard error of the estimate. F. The bootstrap standard error is based on generating random samples with the estimated value of θ, not the actual value of θ (h) In a simple random sample of size 100 to estimate a proportion, the variance of ˆp is always less than or equal to 0.05. T. There was a typo and the number should have been 0.005. This follows from the variance of ˆp being maximal when p = 0.5 (i) Suppose X 1, X,..., X n are the values of the observations of a simple random sample from a population with mean µ. X 1 is an unbiased estimate of µ. T. E(X i ) = µ for all i (j) In the previous problem, 3X 1 X is an unbiased estimate of µ T. E(3X 1 X ) = 3E(X 1 ) E(X ) = µ from above (k) In the previous two problems, the estimate X 1 has a smaller variance than the estimate 3X 1 X. T. The variance of the latter is 13σ and the variance of the former is σ. () According to a recent Gallup poll, President George W. Bush s job approval rating from the American public is an anemic 3%. Results are based on telephone interviews with 1,010 national adults, aged 18 and older, conducted Oct. -7, 007. For results based on the total sample of national adults, one can say with 95% confidence that the maximum margin of sampling error is ± 3

percentage points. In addition to sampling error, question wording and practical difficulties in conducting surveys can introduce error or bias into the findings of public opinion polls. Can you explain how this figure of ± 3% is arrived at?.3.68 sˆp = = 0.017 1009 1.96sˆp = 0.09 3 3. () Two populations are independently surveyed, both with simple random samples of size 100. In one sample the proportion of unemployed is ˆp 1 = 0.15 and in the other, ˆp = 0.1. Find an approximate 90% confidence interval for the difference of unemployment rates, p 1 p. s ˆp1 ˆp =.15.85 + 99 = 0.085.1.88 99 A 90% confidence interval is thus ( ˆp 1 ˆp ) ± 1.68s ˆp1 ˆp or (-0.05, 0.11). The Pareto distribution is sometimes used to model heavy tailed distributions. Consider a Pareto distribution with density function f(x θ) = (θ 1)x θ, θ >, 1 x <. Suppose that X 1, X,..., X n are i.i.d. with density f(x θ) (a) () Find the method of moments estimate of θ. E(X) = (θ 1 1)x θ+1 dx = θ 1 = µ θ θ = 1 µ 1 µ 1 X ˆθ = 1 X (b) () Find the maximum likelihood estimate of θ. l(θ) = n log(θ 1) θ log X i l (θ) = n log X θ 1 1 n ˆθ = P log Xi + 1 3

(c) () Find the asymptotic variance of the maximum likelihood estimate. log f(x θ) = 1 θ (θ 1) V ar(ˆθ) (θ 1) n (d) () Suppose that in a sample of size n = 100, the maximum likelihood estimate is ˆθ = 3.. Give an approximate 90% confidence interval for θ. From the previous part, sˆθ =. = 0. An approximate confidence interval is thus 3. ± 1.68. or 10 (.83,3.57) (e) () If the data are as above, explain clearly and succinctly how the bootstrap could be used to estimate the standard error of ˆθ. Generate a large number of samples of size 100 from a Pareto distribution with parameter θ = 3.. Estimate θ from each of these samples and then find the standard deviation of these estimates. 5. () Suppose a sequence of independent trials, each with probability of success θ, are performed until there are 3 total successes. Let X denote the total number of trials. Then the distribution of X is negative binomial: ( ) k 1 P (X = k) = θ 3 (1 θ) k 3 Let θ denote a player s probability of throwing a ringer in the game of horseshoes. (If you don t know this game, don t worry just regard a ringer as a success.) The player tries until he gets three ringers, performing 18 throws in all. If in a Bayesian analysis, the prior distribution of θ is uniform on [0,1], what is the posterior distribution? What is the mean of the posterior distribution? Since the prior distribution of θ is uniform, the posterior distribution is ( ) x 1 f Θ X (θ x) f X Θ (x θ) = θ 3 (1 θ) x 3 This can be recognized as a beta distribution with a =, b = 16 and mean a/(a + b) = 1/5

6. Suppose that in a population of twins, males (M) and females (F ) are equally likely to occur and that the proportion of identical twins is θ). Identical twins have the same gender. (a) () Show that, ignoring birth order, P (MM) = P (F F ) = (1+θ)/ and P (MF ) = (1 θ)/. P (MM) = P (MM Identical)P (Identical) + P (MM F raternal)p (F raternal) = 1 θ + 1 (1 θ) = 1 + θ = P (F F ) P (MF ) = 1 P (MM) P (F F ) = (1 θ)/ (b) () Suppose that in a sample of n twins, n 1 are F F, n are MM and n 3 are MF. Find the maximum likelihood estimate of θ. From the multinomial distribution, l(θ) = Constant + (n 1 + n ) log(1 + θ) + n 3 log(1 θ) l (θ) = n 1 + n 1 + θ n 3 1 θ ˆθ = n 1 + n n 3 n 1 + n + n 3 If the value above is negative, then ˆθ = 0 Some Reminders For the standard normal distribution, P (Z.57).5%, P (Z.33) 1%, P (Z 1.96).5% and P (Z 1.68) 5%. The multinomial distribution is where n = x i. p(x 1, x,..., x m ) = n! x 1!x! x m! px 1 1 p x p xm m The beta density is f(x) = Γ(a+b) Γ(a)Γ(b) xa 1 (1 x) b 1. E(X) = a/(a + b) 5