Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Similar documents
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Monday Tuesday, January 4 5, 2016

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

P n. This is called the law of large numbers but it comes in two forms: Strong and Weak.

Qualifying Exam in Probability and Statistics.

Applied Statistics Comprehensive Exam

For iid Y i the stronger conclusion holds; for our heuristics ignore differences between these notions.

Problem Selected Scores

1 One-way analysis of variance

Statistics Ph.D. Qualifying Exam

Statistics Ph.D. Qualifying Exam: Part II November 9, 2002

Master s Written Examination

Statistics & Data Sciences: First Year Prelim Exam May 2018

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Spring 2012 Math 541B Exam 1

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018

Qualifying Exam in Probability and Statistics.

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.

Last lecture 1/35. General optimization problems Newton Raphson Fisher scoring Quasi Newton

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Mathematics Qualifying Examination January 2015 STAT Mathematical Statistics

Statistics Ph.D. Qualifying Exam: Part II November 3, 2001

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

Master s Written Examination

Central Limit Theorem ( 5.3)

COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION

A Very Brief Summary of Statistical Inference, and Examples

Link lecture - Lagrange Multipliers

EM Algorithm II. September 11, 2018

Statistics 3858 : Maximum Likelihood Estimators

Review and continuation from last week Properties of MLEs

Statistics 135 Fall 2008 Final Exam

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Lecture 6: Linear models and Gauss-Markov theorem

Optimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X.

Qualifying Exam in Probability and Statistics.

First Year Examination Department of Statistics, University of Florida

Statistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm

Chapter 1. Linear Regression with One Predictor Variable

A Very Brief Summary of Statistical Inference, and Examples

Information in Data. Sufficiency, Ancillarity, Minimality, and Completeness

Linear Models in Machine Learning

Chapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests

STAT 512 sp 2018 Summary Sheet

Lecture 4: Types of errors. Bayesian regression models. Logistic regression

STA 2201/442 Assignment 2

1. Let A be a 2 2 nonzero real matrix. Which of the following is true?

UNIVERSITY OF MASSACHUSETTS. Department of Mathematics and Statistics. Basic Exam - Applied Statistics. Tuesday, January 17, 2017

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Probability and Statistics qualifying exam, May 2015

AMCS243/CS243/EE243 Probability and Statistics. Fall Final Exam: Sunday Dec. 8, 3:00pm- 5:50pm VERSION A

Master s Written Examination - Solution

Some explanations about the IWLS algorithm to fit generalized linear models

Chapter 7. Hypothesis Testing

Ph.D. Qualifying Exam: Algebra I

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1

Linear model A linear model assumes Y X N(µ(X),σ 2 I), And IE(Y X) = µ(x) = X β, 2/52

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

Statistical Estimation

Statistics 135: Fall 2004 Final Exam

Model Checking and Improvement

Linear Methods for Prediction

Estimation and Model Selection in Mixed Effects Models Part I. Adeline Samson 1

LIST OF FORMULAS FOR STK1100 AND STK1110

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012

Maximum Likelihood Estimation

BTRY 4090: Spring 2009 Theory of Statistics

Maximum Likelihood Estimation. only training data is available to design a classifier

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

Regression Models - Introduction

Masters Comprehensive Examination Department of Statistics, University of Florida

Linear Regression (9/11/13)

Statistics and Econometrics I

Gibbs Sampling in Linear Models #2

Spring 2012 Math 541A Exam 1. X i, S 2 = 1 n. n 1. X i I(X i < c), T n =

PhD Qualifying Examination Department of Statistics, University of Florida

1 Introduction to Generalized Least Squares

Generalized Linear Models. Kurt Hornik

Linear Models A linear model is defined by the expression

Math 494: Mathematical Statistics

Linear Methods for Prediction

Stat 5102 Final Exam May 14, 2015

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

BIOS 2083 Linear Models c Abdus S. Wahed

Probability and Statistics Notes

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

[y i α βx i ] 2 (2) Q = i=1

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

Summary of Chapters 7-9

Logistic Regression and Generalized Linear Models

Simple Linear Regression for the MPG Data

Regression Estimation Least Squares and Maximum Likelihood

Finite Singular Multivariate Gaussian Mixture

REGRESSION WITH SPATIALLY MISALIGNED DATA. Lisa Madsen Oregon State University David Ruppert Cornell University

Chap 2. Linear Classifiers (FTH, ) Yongdai Kim Seoul National University

Chapter 4: Asymptotic Properties of the MLE (Part 2)

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

Lecture 14 More on structural estimation

Transcription:

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently drawn from distributions with means µ 1 and µ 2 and the same variance σ 2. Observation ID 1 2 3 4 5 6 7 8 9 10 x x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 x 10 y y 1 y 2 y 3 y 4 y 5 y 6 y 7 y 8 y 9 y 10 Describe how to test hypotheses on the means of the two distributions based on the two samples. State any additional assumptions on the distributions you need for the test. Give your procedure and formulas. Describe how to test the two distributions using a nonparametric method such as the Wilcoxon Rank-Sum test. Give your procedure and formulas. Describe how to test the mean difference (µ 1 µ 2 ) using a randomization distribution method. Give your procedure and compare this method with the other two methods in and. Problem 2. (5166) linear model: In a two-way factorial experiment with replicates, consider the following Y ijk = µ + α i + β j + ɛ ijk, i = 1,..., a; j = 1,..., b; k = 1,..., r, where µ is the overall mean, α i is the effect for the ith level of factor A, and β j is the effect for the jth level of factor B. The random errors {ɛ ijk } are assumed to be independent and normally distributed with mean zero and variance Var(ɛ ijk ) = σ 2 j. What are the distributions of Ȳi and Ȳ? Let S A = br a i=1 (Ȳi Ȳ ) 2, S B = ar b j=1 (Ȳ j Ȳ ) 2, and S R = a i=1 Find the mean values of S A and S R. b j=1 k=1 r (Y ijk Ȳi Ȳ j + Ȳ ) 2. Show that the decomposition of variation of this model is: where S D = S A + S B + S R a b r S D = (Y ijk Ȳ ) 2. i=1 j=1 k=1 1

Problem 3. (5167) Consider the linear regression model: Y = Xβ + ξ, where Y = (y 1,..., y n ), ξ = (ξ 1,..., ξ n ), β = (β 1,..., β p ) and X is an n p full-rank matrix. The process {ξ i } is generated by the moving-average model: ξ i = ɛ i θ 1 ɛ i 1, where {ɛ i, i = 0, 1,..., n} are iid N(0, σ 2 ) variables. Calculate the variance-covariance matrix of ξ. Describe how to use the least squares method, the weighted least squares method, and the maximum likelihood method to estimate the coefficients β in the model. Give details of the three procedures. Let ˆβ be the least squares estimate of β. Define Ŷ = X ˆβ, and ˆξ = Y Ŷ. Is ˆβ an unbiased estimate of β? What is the variance matrix of ˆβ? Are Ŷ and ˆξ independent? Show your reasons. Problem 4. (5167) Consider a multiple linear regression model with the form: y i = β 0 + β 1 x i1 + + β p x ip + ɛ i, i = 1,, n, where {ɛ i, i = 1,..., n} are independent random variables with mean zero and variance σ 2 i. Define β = (β 0, β 1,..., β p ). Assume that all the observations {y i ; i = 1,, n} are positive values. Describe how to choose a transformation of the response y to make the distribution of y i approximately normal with constant variance. Give a procedure and reasons. Describe how to select variables {x 1,, x p } in this linear model. Give at least two procedures for model selection and compare the procedures. Give a procedure to evaluate the influence of each case {y i, x i1, x i2,, x ip } and give your justifications for the procedure. 2

Problem 5. (5106) Let (X 1, Y 1 ), (X 2, Y 2 ),, (X n, Y n ) be a sequence of i.i.d. paired samples. Conditioned on X i, the value Y i follows a Poisson distribution with parameter λ Xi = e βx i. That is, Y i X i, β Poisson(λ Xi ) Our goal is to find the maximum likelihood estimate (MLE) of β with the observations {(X i, Y i )} n i=1. Derive an expression for the log-likelihood function l(β) = n log(f(y i X i, β)), i=1 such that the MLE is given by ˆβ = argmax β l(β). Find the expressions for l(β) and l(β), the first and second derivatives of l with respect to β. Verify that l(β) < 0. Write out the Newton-Raphson algorithm to find the root of l(β). Problem 6. (5106) Find the maximum likelihood estimate of θ where θ is a parameter in the multinomial distribution: (x 1, x 2, x 3, x 4 ) M(n; 0.1(1 θ), 0.4(2 3θ), 0.1(1 + 3θ), θ) Choose a variable for the missing data and use the EM algorithm for iteratively estimating θ. Let θ (m) be the current value of the unknown. Derive the mathematical formula to update for θ (m+1). Let L(θ) denote the likelihood with parameter θ. Prove that L(θ (m+1) ) L(θ (m) ). 3

Put your solution to each problem on a separate sheet of paper. Problem 7. (5326) Let X 1, X 2, X 3,... be iid Uniform(0, 1). Find both P (X 2 1 + X 2 2 < 1) and P (X 2 2X 1 > 0). Find the density of Y = X 2 X 1. Find the moment generating function of Z = 2(X 1 + X 2 + + X 20 ) 5. (d) What is the approximate distribution of Z = 2(X 1 + X 2 + + X 20 ) 5? (Give the name of the distribution and the values of any parameters.) Problem 8. (5326) A fair dodecahedron (a 12-sided solid) has its sides labeled 1, 2,..., 12. The dodecahedron is rolled five times. Let (X 1, X 2, X 3, X 4, X 5 ) denote the observed sequence of values. Find P (X 1 < X 2 < X 3 < X 4 < X 5 ). Find P (X 1 X 2 X 3 X 4 X 5 ). Find the probability that the sequence (X 1, X 2, X 3, X 4, X 5 ) contains three consecutive consecutive integers. (In other words, (X i, X i+1, X i+2 ) = (y, y + 1, y + 2) for some values i and y. Some possible sequences (X 1,..., X 5 ) which satisfy this are (8, 6, 7, 8, 3), (2, 8, 9, 10, 11), and (4, 5, 6, 7, 8).) Problem 9. (5327) The following parts use the family of densities f(x θ) = θ2 log x x θ+1 for 1 x, θ > 0. Suppose X 1,..., X n are iid f(x θ). Find the MLE for θ. Suppose that we have two independent random samples: X 1,..., X n are iid f(x θ), and Y 1,..., Y m are iid f(x β). Find the likelihood ratio test (LRT) of H 0 : θ = β versus H 1 : θ β. Show that the test in part can be based on the statistic U = n i=1 log X i n i=1 log X i + m i=1 log Y. i 4

Problem 10. (5327) Let X 1, X 2,..., X n be a random sample from the density f(x θ) = θ(θ + 1) x(1 x) θ 1 for 0 < x < 1, θ > 0. Answer the following. Justify your answers. Find a complete sufficient statistic for θ. Is X i sufficient for θ? Find I(θ), the Fisher information in a single observation from f(x θ). Problem 11. (6346) Let (Ω, F, µ) be a probability space and (O, G) be a measurable space. If f : Ω [0, ) is a non-negative simple function with fdµ < and B is a set with µ(b) = 0, show fdµ = 0. B Extend part to include non-negative measurable functions f. Suppose a non-negative random variable Y : Ω [0, ) has EY <, and X : Ω O is a measurable function. Show that there exists a function g such that Y (ω)dµ(ω) = g(ω)dµ X (ω) for any A G. X A A Problem 12. (6346) For the parts below, you may assume that X n and X are random variables with all moments finite. State Markov s inequality. Define convergence in probability. Define convergence in L p. L p (d) Show X n X P Xn X for finite p. (e) Prove Lyapunov s inequality: For 0 < q < p, [E ( X q )] 1/q [E ( X p )] 1/p (f) Show that finite second moment implies finite variance. 5