Problems ( ) 1 exp. 2. n! e λ and

Similar documents
Topic 15: Simple Hypotheses

1 Presessional Probability

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

Mathematical Statistics

Chapters 10. Hypothesis Testing

Introductory Econometrics. Review of statistics (Part II: Inference)

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

Master s Written Examination

This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

Class 26: review for final exam 18.05, Spring 2014

F79SM STATISTICAL METHODS

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

CSE 312 Final Review: Section AA

Institute of Actuaries of India

f (1 0.5)/n Z =

Quick Tour of Basic Probability Theory and Linear Algebra

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Chapters 10. Hypothesis Testing

18.175: Lecture 13 Infinite divisibility and Lévy processes

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

Random Variables and Their Distributions

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Hypothesis testing: theory and methods

Chapters 9. Properties of Point Estimators

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

ECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172.

MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation)

This paper is not to be removed from the Examination Halls

Exercises with solutions (Set D)

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

Topic 10: Hypothesis Testing

Math Bootcamp 2012 Miscellaneous

Notes on the Multivariate Normal and Related Topics

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n

Chapter 5. Chapter 5 sections

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Qualifying Exam in Probability and Statistics.

MA 519 Probability: Review

CONTINUOUS RANDOM VARIABLES

1. Let A be a 2 2 nonzero real matrix. Which of the following is true?

1 Probability Model. 1.1 Types of models to be discussed in the course

Problem 1 (20) Log-normal. f(x) Cauchy

Non-parametric Inference and Resampling

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages

Topic 10: Hypothesis Testing

14.30 Introduction to Statistical Methods in Economics Spring 2009

Conditional distributions (discrete case)

Master s Written Examination - Solution

EC2001 Econometrics 1 Dr. Jose Olmo Room D309

Chapter 4. Chapter 4 sections

T has many other desirable properties, and we will return to this example

2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Spring 2012 Math 541A Exam 1. X i, S 2 = 1 n. n 1. X i I(X i < c), T n =

Math 180B Problem Set 3

Introduction to Bayesian Statistics

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Summary: the confidence interval for the mean (σ 2 known) with gaussian assumption

Homework for 1/13 Due 1/22

UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences FINAL EXAMINATION, APRIL 2013

Two hours. Statistical Tables to be provided THE UNIVERSITY OF MANCHESTER. 14 January :45 11:45

18.05 Practice Final Exam

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

Lecture 13: p-values and union intersection tests

Statistical Preliminaries. Stony Brook University CSE545, Fall 2016

If we want to analyze experimental or simulated data we might encounter the following tasks:

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

Special distributions

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

SDS 321: Introduction to Probability and Statistics

Performance Evaluation and Comparison

Ph.D. Qualifying Exam Monday Tuesday, January 4 5, 2016

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

Introduction to Bayesian Methods. Introduction to Bayesian Methods p.1/??

18.175: Lecture 17 Poisson random variables

Multiple Testing in Group Sequential Clinical Trials

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

More on Distribution Function

STT 843 Key to Homework 1 Spring 2018

ST5215: Advanced Statistical Theory

First Year Examination Department of Statistics, University of Florida

Simple Linear Regression

ECON Fundamentals of Probability

Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic

6 The normal distribution, the central limit theorem and random samples

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics

Probability Theory and Statistics. Peter Jochumzen

Chapter 9: Hypothesis Testing Sections

Probability and Distributions

Chapter 4 Sampling Distributions and Limits

Summary of Chapters 7-9

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018

Transcription:

Problems The expressions for the probability mass function of the Poisson(λ) distribution, and the density function of the Normal distribution with mean µ and variance σ 2, may be useful: ( ) 1 exp. 2πσ 2 λ n n! e λ and (x µ)2 2σ 2 1. (10 points) For each of the following questions, exactly one answer is correct. Each correct answer gives 1 point, and each incorrect answer results in a 1/2 point reduction. The minimal possible total score for the full problem is 0. Some of the results may be useful in later problems! a) Let X 1, X 2,... be an i.i.d. sequence of random variables with 0 < E(X 2 1) <. Which of the following statements is a consequence of the strong law of large numbers? 1. lim sup n X n = + almost surely. 2. lim n (X 2 1 + + X 2 n)/n = 0 almost surely. 3. sup n 1 X n < almost surely. b) Let ϕ 1 and ϕ 2 be the characteristic functions of two random variables X 1 and X 2. What does the identity theorem for characteristic functions tell us? 1. If ϕ 1 (u) = ϕ 2 (u) for all u R then X 1 and X 2 have the same distribution. 2. If ϕ 1 (u) = ϕ 2 (u) for all u R then X 1 and X 2 are equal almost surely. 3. If ϕ 1 (u) = ϕ 2 (u) for all u R then X 1 and X 2 are independent. c) What does it mean for a statistical test to have significance level 0.05? 1. The Null hypothesis will be accepted with probability at least 0.05. 2. If the Null hypothesis is true, then the probability to reject is at most 0.05. 3. If the Null hypothesis is false, then the probability to reject is at least 0.95. d) The characteristic function ϕ of a Poisson(λ) random variable is given by 1. ϕ(u) = exp(u(e iλ 1)). 2. ϕ(u) = exp(λ(e iu 1)). 3. ϕ(u) = exp(λ(iu 1)). 1

e) Suppose a statistical test resulted in a p-value of 0.07. Which of the following statements is correct? 1. The probability that the Null hypothesis is false is 0.07. 2. The Null hypothesis could not be rejected at significance level 0.06. 3. The Null hypothesis could not be rejected at significance level 0.08. f) Suppose X 1,..., X n are i.i.d. N(µ, σ 2 ). What is the distribution of 1 n n i=1 1. N(0, 1). 2. N(0, 1/n). 3. Student t with n degrees of freedom. X i µ σ? g) Suppose X 1,..., X n are i.i.d. Bernoulli(p) for some 0 < p < 1. What is the distribution of X 1 + + X n? 1. Binomial(n, p). 2. Poisson(np). 3. Geometric(n, p). h) Which of the following formulas is not correct in general? 1. P(A B) = P(A) + P(B) P(A B). 2. P(A B) = P(A) + P(A c B). 3. P(A B) = P(A) + P(A B c ). i) Which of the following statements is correct for any random variables X, Y with unit variance, Var(X) = Var(Y ) = 1? 1. If Cov(X, Y ) = 0 then X and Y are independent. 2. If X = a + by for some constants a R and b > 0, then Cov(X, Y ) = 1. 3. If X = Y 2, then Cov(X, Y ) = 1. j) What is the value of n 1 k=0 2k? 1. 2 n 1. 2. n 2 1. 3. 2 n n. 2

2. (10 points) A team of particle physicists has performed a series of experiments using the Large Hadron Collider (LHC) to study Higgs boson production. The total number of produced Higgs bosons is N Poisson(λ). Unfortunately, due to background effects, other particles are also produced during the experiments. The total number of such particles is M Poisson(γ), with M and N independent. The parameters λ and γ are nonnegative and unknown. The detector unit can only measure the total number of produced particles, X = M + N. a) Show that X Poisson(λ + γ). The parameter of interest is λ, which controls the number of Higgs bosons produced. To proceed, further information about the background parameter γ is needed. Therefore, the physicists perform a separate experiment from which a random variable Y N(γ, σ 2 ) is observed independently of X, where σ 2 > 0 is known. b) The joint distribution of (X, Y ) given the parameters θ = (λ, γ) is of the form Determine f(n, y; θ). P θ (X = n, Y z) = Consider the likelihood function L(θ) = f(x, Y ; θ). z f(n, y; θ)dy. c) Find the maximum likelihood estimate ( λ, γ) of (λ, γ), knowing X and Y, in the case where 0 < Y < X. 3

3. (15 points) Consider a series of coin tosses modeled by a sequence X 1, X 2,... of i.i.d. random variables with P(X n = 1) = P(X n = 1) = 1 2. Here X n = 1 signifies heads and X n = 1 tails. Before each toss, you have the opportunity to bet an amount V n. If the coin comes up heads (X n = 1), you receive V n. Otherwise you lose V n. Therefore the total gain up to and including the nth toss is G n = n V k X k. k=1 Note that G n may be negative, which signifies a loss. Suppose you use the following strategy: Set V 1 = 1. For k 1, if X k = 1 you stop, meaning that you set V n = 0 for n k + 1. If X k = 1 and you have not yet stopped, double your bet, that is, set V k+1 = 2 V k. Furthermore, let T denote the time you stop, T = min{k 1 : X k = 1}, with T = if X k = 1 for all k. a) Compute G n for 1 n < T. b) Compute G T for T <. c) Compute P(T < ). At this point you may feel that this is too good to be true and indeed there is a catch: d) Compute the expected maximal intermediate shortfall. That is, compute E ( min G ) n. 1 n T To make matters worse, in reality you would face a limit on credit. Specifically, assume you are only allowed to use strategies {V n : n 1} such that the gain satisfies G n b almost surely for all n, where b (1, ) is some fixed constant. Modify the above strategy so that you stop as soon as there is a risk of violating the bound. That is, specify V k as before, except that if you have not stopped at k, and G k 2 V k < b, then set V n = 0 for all n k + 1. Let T denote the time you stop under this modified strategy. e) Compute your expected gain E(G T ) under these new rules. In particular, how does this depend on the bound b? 4

4. (15 points) Celebrities Saylor Twift and Bustin Jieber are active on Twitter. Saylor has a total of n followers. There is some overlap between Twist and Jieber followers: a fraction γ (0, 1) of Saylor s followers also follow Bustin (that is, [γn] people follow both Saylor and Bustin, where [x] denotes the integer part of a real number x). Let A denote the event that Bustin tweets Saylor Twift got a #beautifulvoice. If a follower of Saylor reads this tweet, then this follower will re-tweet Saylor s tweets with probability p S = 0.3. Otherwise, the probability is only q S = 0.1. Followers make tweeting decisions independently of each other, and do not re-tweet each other s tweets. One day, Saylor tweets today glimpsed a new world who knew that #statsrulez?. Let N denote the total number of re-tweets of Saylor s tweet. a) Assuming that A did not happen, find E(N) and Var(N). b) Assuming that A did happen, find E(N) and Var(N). Suppose you work for Twitter and a colleague tells you the value of N. You are currently offline and cannot check specific tweets. You are interested in if A happened or not. c) Under each of the distributions in a) and b), show that N E(N) Var(N) converges weakly to the standard Normal distribution as n, with p S, q S, and γ held fixed. Hint: It may be helpful to write N as a sum of Bernoulli random variables. Furthermore, you are allowed to use the following result: Lemma: Let U 1, U 2,... and V 1, V 2,... be two sequences of random variables that both converge weakly to the standard Normal distribution. Suppose U n and V n are independent for each n. Let ϱ n (0, 1) and assume lim n ϱ n = ϱ (0, 1). Then ϱ n U n + 1 ϱ 2 nv n converges weakly to the standard Normal distribution. d) Consider the hypotheses { Null: Alternative: The event A happened; The event A did not happen. In this problem, you should approximate the corresponding distributions of N by Normals with the same mean and variance. Show that one can find a most powerful test at a given level α of the form reject the Null N + 0.05 n > c for some constant c. (You do not have to determine c; it depends on α and the distribution of N under the Null hypothesis.) 5

5. (15 points) A common measure of the quality of an estimator is the mean squared error (MSE). In this problem, you will see that the MSE may behave in unexpected ways. This was first noticed by Stein in 1956 and developed further by James and Stein in 1961. Fix n 2. Let X = (X 1,, X n ) be a random vector, where X 1,..., X n are independent with X i N(θ i, 1) for some unknown parameters θ = (θ 1,..., θ n ). a) Consider the estimator θ i = X i for i = 1,..., n, and set θ = ( θ 1,..., θ n ). Compute the mean squared error E( θ θ 2 ) Consider now the alternative estimator θ JS = ( θ 1 JS JS,..., θ n ), where θ JS i = X i n 2 X 2 X i, i = 1,..., n. b) Let h : R n R be continuously differentiable and such that h(x) = 0 whenever x is sufficiently large (that is, for h(x) = 0 whenever x > C for some constant C). Show that ( ) ( ) h E (X i θ i )h(x) = E (X) for all i = 1,..., n. (1) x i Hint: Use that φ (t) = tφ(t), where φ denotes the N(0, 1) density function. Equation (1) actually holds for more general functions h. In particular, for any i {1,..., n}, one can take h(x) = x i / x 2 (setting h(0) = 0 and h x i (0) = 0), as long as n 3. You may use this fact without proof. c) For n 3, show that θ JS has smaller mean squared error than θ. That is, show that E( θ θ JS 2 ) < E( θ θ 2 ). This result is counterintuitive, because the X i are independent. It is tempting to conclude that the quality of an estimate can be improved by simultaneously estimating independent variables. Food for thought... 6