Eco517 Fall 2004 C. Sims MIDTERM EXAM

Similar documents
MAXIMUM LIKELIHOOD, SET ESTIMATION, MODEL CRITICISM

Eco517 Fall 2014 C. Sims MIDTERM EXAM

Eco517 Fall 2014 C. Sims FINAL EXAM

(I AL BL 2 )z t = (I CL)ζ t, where

MODEL COMPARISON CHRISTOPHER A. SIMS PRINCETON UNIVERSITY

Model comparison. Christopher A. Sims Princeton University October 18, 2016

APPM/MATH 4/5520 Solutions to Problem Set Two. = 2 y = y 2. e 1 2 x2 1 = 1. (g 1

MID-TERM EXAM ANSWERS. p t + δ t = Rp t 1 + η t (1.1)

Testing Restrictions and Comparing Models

STA 2201/442 Assignment 2

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Hypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33

Bayesian Regression Linear and Logistic Regression

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1

STAT 830 Bayesian Estimation

Problem 1 (20) Log-normal. f(x) Cauchy

A Bayesian perspective on GMM and IV

DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling

First Year Examination Department of Statistics, University of Florida

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Review. December 4 th, Review

Hypothesis Testing. Part I. James J. Heckman University of Chicago. Econ 312 This draft, April 20, 2006

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics

Eco517 Fall 2013 C. Sims MCMC. October 8, 2013

A Bayesian Treatment of Linear Gaussian Regression

Review of Statistics

Line Broadening. φ(ν) = Γ/4π 2 (ν ν 0 ) 2 + (Γ/4π) 2, (3) where now Γ = γ +2ν col includes contributions from both natural broadening and collisions.

Fitting a Straight Line to Data

STAT215: Solutions for Homework 2

18.05 Practice Final Exam

2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.

Problem #1 #2 #3 #4 #5 #6 Total Points /6 /8 /14 /10 /8 /10 /56

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Bayesian Semiparametric GARCH Models

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω

Bayesian Semiparametric GARCH Models

Exam C Solutions Spring 2005

7. Estimation and hypothesis testing. Objective. Recommended reading

MCMC algorithms for fitting Bayesian models

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Bayesian Inference: Concept and Practice

Fundamental Probability and Statistics

An Introduction to Bayesian Linear Regression

Choosing among models

STA 294: Stochastic Processes & Bayesian Nonparametrics

Stat 451 Lecture Notes Numerical Integration

Statistical Data Analysis Stat 3: p-values, parameter estimation

7. Estimation and hypothesis testing. Objective. Recommended reading

Stat751 / CSI771 Midterm October 15, 2015 Solutions, Comments. f(x) = 0 otherwise

The strictly 1/2-stable example

STA 4273H: Sta-s-cal Machine Learning

Statistical Theory MT 2006 Problems 4: Solution sketches

Bayesian Econometrics

Principles of Bayesian Inference

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

Statistics 3858 : Maximum Likelihood Estimators

ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS

STA 4273H: Statistical Machine Learning

Eco Spring 2002 Chris Sims OLG EXERCISES

MAS223 Statistical Inference and Modelling Exercises

Stat 5102 Final Exam May 14, 2015

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

CHAPTER 1: Functions

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions

01 Probability Theory and Statistics Review

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

A Very Brief Summary of Statistical Inference, and Examples

Problem Selected Scores

Statistical Distribution Assumptions of General Linear Models

UQ, Semester 1, 2017, Companion to STAT2201/CIVL2530 Exam Formulae and Tables

(x 3)(x + 5) = (x 3)(x 1) = x + 5. sin 2 x e ax bx 1 = 1 2. lim

ECO 513 Fall 2008 C.Sims KALMAN FILTER. s t = As t 1 + ε t Measurement equation : y t = Hs t + ν t. u t = r t. u 0 0 t 1 + y t = [ H I ] u t.

Infinite Limits. Infinite Limits. Infinite Limits. Previously, we discussed the limits of rational functions with the indeterminate form 0/0.

Math 115 Spring 11 Written Homework 10 Solutions

Bayesian Linear Regression

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

STA 2101/442 Assignment 2 1

UNIFORMLY MOST POWERFUL CYCLIC PERMUTATION INVARIANT DETECTION FOR DISCRETE-TIME SIGNALS

Markov Chain Monte Carlo (MCMC)

The distribution inherited by Y is called the Cauchy distribution. Using that. d dy ln(1 + y2 ) = 1 arctan(y)

STA 4322 Exam I Name: Introduction to Statistics Theory

Unless provided with information to the contrary, assume for each question below that the Classical Linear Model assumptions hold.

review session gov 2000 gov 2000 () review session 1 / 38

Practice Problems Section Problems

Swarthmore Honors Exam 2012: Statistics

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Disjointness and Additivity

p(z)

Parametric Techniques

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley

Interpreting Regression Results

Bayesian Estimation An Informal Introduction

MCMC notes by Mark Holder

Transcription:

Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering imposing a tax to finance a public program. The benefits from the tax are B = 4(τ θτ 2 ), showing diminishing returns because as the tax increases the activity being taxed decreases. The deadweight loss from the tax is C = θτ 2. We are not sure of θ, which represents the sensitivity of the activity with respect to the tax. θ = 1 or θ = 2 are the only two possibilities. Net benefits of the activity are of course B C. Our options are restricted to setting τ to.1,.3, or.4. (a) Calculate the pairs of losses, under θ = 1 and θ = 2, for each of the three τ values. [It may help to line these up in a θ by τ table.] Net gains are 4τ 5θτ 2. Here is the table: τ θ 1 2.1.35.3.3.75.3.4.8 0 (b) Are any of the three τ choices inadmissible? Explain your answer. τ =.1 is inadmissible, since it produces lower gains than τ =.3 when θ = 1, and equal gains when θ = 2. (c) Which of the τ values represent Bayesian decision rules? For each of these, specify at least one probability distribution over θ that would support choice of that τ value. The other two choices are both admissible and Bayesian. τ =.4 is preferred if P[θ = 1] > 6/7. Otherwise τ =.3 is preferred. A possibly subtle point is that τ =.1 is also Bayesian. It delivers expected gains equal to those of τ =.3 in the case where P[τ = 2] = 1. This illustrates the point that Bayes decision rules developed from priors that put probability zero on a set can be inadmissible through being dominated on that zero-probability set. (2) Suppose we have an i.i.d. sample of T observations X t, with each X t distributed as gamma(p), i.e. with pdf x p 1 e x /Γ(p). (a) Show that the sample mean of the X t s, X = X t /T, is an unbiased estimator for p. c 2004 by Christopher A. Sims. This document may be reproduced for educational and research purposes, so long as the copies contain this notice and are retained for personal use or distributed free.

MIDTERM EXAM 2 It was acceptable simply to point out that the mean of a standard Gamma(p, α) is p/α, so that here each X t has mean p. Therefore E[ X p] = T p/t = p, and X is unbiased for p. One could also do the integral: E[X t p] = 0 x xp 1 e x dx Γ(p) = Γ(p + 1) Γ(p) = p, where we are using the fact that Γ(p) = (p 1)! for integer p. (b) Display the kernel of the Bayesian posterior pdf for p when the prior pdf for p has the form π(p) = λ p e λ /Γ(p + 1) over the integer values of p = 0, 1,...,, with λ > 0. (The kernel of the pdf is just any function that is proportional to the pdf, without necessarily integrating to one.) Note that here we are talking about a posterior concentrated on the nonnegative integers. The kernel is λ p Γ(p) T Γ(p + 1) ( T t=1 x t ) p 1 e T 1 x t λ. Since we are interested only in the behavior of this as a function of p, we can simplify it by removing all factors that don t depend on p, to arrive at λ p ( T t=1 x t Γ(p) T+1 p ) p 1 There is a glitch in the statement of this problem, though. The Gamma distribution and Γ(p) are undefined for p = 0, so at best this has to be interpreted as a kernel over p 1. Γ(p) is defined for all p > 0. What I meant to state as the prior was π(p) = λ p 1 e λ /Γ(p), for integer p > 0. This would have led to the posterior kernel ( ) p 1 λ t=1 T x t Γ(p) T+1 (c) Explain why the sample mean cannot be a Bayesian posterior mean for p, regardless of the prior distribution on p. A reasonable answer was simply that we showed in class that an unbiased estimator cannot be a Bayesian posterior mean with any proper prior, if the estimator has finite variance. However, to make this answer complete you should have checked the finite variance property, and the question did not explicitly rule out improper priors. A direct and clean argument is simply to note that the sufficient statistic for p here is x t (or equivalently ( x t ) 1/T, the geometric mean). Since the sample mean does not depend on the data exclusively.

MIDTERM EXAM 3 through the sufficient statistic, it is not determined by the likelihood function shape alone. It therefore can t be a Bayesian estimator, because Bayesian estimators, whether from proper or improper priors, and regardless of the loss function, depend on the data only through the likelihood function and (therefore) through the sufficient statistics. (3) Suppose we will have a single observation on a random variable X that is the sum of two independent U(0, a) (uniform on the interval (0, a)) random variables, where a > 0 is unknown. (a) Find the pdf of X a. There are several ways to get the answer, which is that the pdf of X is 1/a (x a)/a 2 for x (0, 2a), 0 elsewhere, i.e. a triangle on (0, 2a) with peak height 1/a at a. One way to get this result is to recall that the sum of two independent random variables is the convolution of their pdfs, so here we have, for x (0, a) x 1 0 a 2 dy = x a 2. It should be clear by the symmetry, that this linear behavior for x < a just gets reversed as x proceeds above a. One could also think directly about the 2- dimensional geometry. If U, V are the two independent uniforms that sum to X, they have a joint density spread evenly over the unit square in R 2. Transforming from u, v coordinates to x, v coordinates has a Jacobian of 1, but we have to recognize that for a fixed value of x, v can range only between max(0, x a) and min(x, a). When we then integrate w.r.t. y, we get the same result. (b) Show that X is an unbiased estimator for a. This is obvious, since the pdf is symmetric about a and bounded, so its expectation exists. (c) Construct a two-tail test of size.04 for the null hypothesis H 0 : a = a 0, for an arbitrary a 0 > 0. We want to find an x such that x 0 (x/a2 0 ) dx =.02. This can easily be seen to be x =.2a 0. Therefore the symmetric interval that, conditional on a 0, has probability.96, is (.2a 0, 1.8a 0 ). The test should therefore reject whenever X is outside this interval. (d) Construct a 96% confidence interval for a based on the family of tests you contructed in (3c) For the confidence interval, we just collect all values of a that are not rejected for a given observation on X. These are all a s such that X >.2a or X < 1.8a, or in other words all a s in the interval ( 5 9 X, 5X).

MIDTERM EXAM 4 (e) Calculate the likelihood function and sketch a plot of it against a. Will the posterior mean under a flat prior be X? Is it likely that a decision maker with loss function (a â) 2 will choose â = X? Why or why not. The likelihood is x/a 2 for a > x, 2/a x/a 2 for a (x/2, x). See the figure for a sketch. Since the pdf for large a is x/a 2, the flat-prior posterior pdf, though integrable, has infinite first moment, so certainly X is not the posterior mean. A decision-maker should be using a proper prior, so the fact that the flat-prior posterior implies infinite expected loss under the loss function given does not mean the decision-maker would see infinite expected loss. But if the decision-maker s prior were not sharply concentrated relative to the likelihood, the heavy skewness to the right of the likelihood would pull the posterior mean (which is the optimal choice under a quadratic loss function) well above X. As an aside, (you were not asked for this) the flat-prior 96% equal-tail interval for a is (.59X, 36X), much more skewed to the right than the confidence region. pdf 0.0 0.2 0.4 0.6 0.8 1.0 0 2 4 6 8 10 (f) Suppose we want to test H 0 : a < 1. Consider the following procedure: Reject H 0 if the confidence interval you constructed in (3d) is disjoint from (0, 1) (i.e., does not overlap it at all). What is the size of this test? Is it unbiased? When a = 1, this test rejects if and only if X is in the right tail of the rejection region for the part (3c) test of a as null hypothesis, so the rejection probability a

MIDTERM EXAM 5 is.02. The test rejects whenever X exceeds 5. The probability of such an openended-to-the-right interval, conditional on a, is strictly increasing in a. Thus for a s less than one the rejection probability is smaller, so the size of the test is.02, and for a s greater than one the rejection probability is larger, so the test is unbiased. (4) We have data for a standard normal linear model, {Y X} N(Xβ, σ 2 I). There are 10 observations and 1 1 1 X X = 1 2 2, (X X) 1 = 1 2 3 2 1 0 1 2 1 0 1 1 6 X Y = 11, Y Y = 80. 14 1, ˆβ OLS = 2, 3 (a) Find a 95% two-sided confidence interval for β 2 + β 3. The.025 tail for a t 7 (0, 1) distribution is at 2.36. First we need the usual estimate of residual variance, which here can be found as (Y Y ˆβ OLS X Y)/(T k) = 10/7. We are interested in the linear combination of β s with weights w = (0, 1, 1). We can therefore treat ˆβ OLS,2 β 2 + ˆβ OLS,3 β 3, conditional on β, as t 7 with variance parameter (10/7) w(x X) 1 w = 10/7. This makes the requested interval 5 ± (2.36 1.2) = (2.2, 7.8). (b) Explain how you would calculate a 95% HPD set for β 1 /(β 1 + β 2 ) under a dσ/σ prior, without actually trying to carry out the calculation. You should explain as if giving intructions to an undergraduate RA well trained in algebra and calculus and computer programming, but who knows nothing about probability or statistics. There are two possible approaches: Jacobian or Monte Carlo simulation. Jacobian: If we set γ = β 1 /(β 1 + β 2 ), the Jacobian is (β 1, β 2 (γ, β 2 ) = β 2 (1 γ) 2. Substitute β 1 = γβ 2 /(1 γ) for β 1 in the multivariate t 7 density for β, and multiply the result by the Jacobian. Numerically integrate this expression with respect to β 2 at a finely spaced set of values of γ, chosen to cover the region where the integrand is large. Take special care with γ = 1 because the Jacobian suggests the pdf will show an integrable discontinuity there; it may be necessary to use a finer grid in that neighborhood. Normalize the resulting vector of marginal pdf heights so they

MIDTERM EXAM 6 sum to one. For each point on this grid of γ values, calculate the marginal pdf height h(γ) and the cdf value F(γ)(found as the sum of pdf values over γ s less than or equal to the current one). Once this is done, examine all pairs γ 1, γ 2 such that F(γ 1 ) + F(γ 2 ) 1 =..05 Among these, pick the pair that most nearly satisfies h(γ 1 ) = h(γ 2 ). Note that, though 1/(1 γ) 2 is not integrable over domains that include γ = 1, the joint and marginal pdf s we are dealing with here must be integrable, because they are pdf s for functions of β, which has an integrable likelihood for T > k. As γ 1 the Jacobian goes to infinity, but at the same time β 1 = γβ 2 /(1 γ 1 ) goes rapidly to infinity, for fixed β 2, and this expression appears squared in the denominator of the t pdf. Monte Carlo: You can imagine that there is a program to make i.i.d. pseudorandom draws from a multivariate t distribution. If this were available, you would just instruct the RA to make a few thousand such draws, say 4000, with the degrees of freedom 7 and the covariance matrix (10/7)(X X) 1. From these draws, form the vector of values for γ = β 1 /(β 1 + β 2 ). Sort these values in increasing order. An equal-tail posterior probability interval is then approximately given by the range from the 200 th to the 3800 th value of γ in these draws. This is not the HPD interval, however, unless it happens that a good estimate of the pdf at these points matches. One could locate an estimate of the HPD interval by looking at the 41 intervals defined by, say, the 180 th to the 220 th draws paired with the corresponding 3780 th to 3820 th draws. The shortest of these intervals is likely a good estimate of the HPD, unless the shortest interval turns out very close to the (220,3820) or (180,3780) pair. In that case, the search should be extended further to the right or left (respectively). Since a multivariate t random generator is not in the R base package, a still better answer might have explained how to generate a multivariate t by drawing first from the Gamma(T) posterior on ν = 1/σ 2, then from the N ( ˆβ OLS, (νx X) 1) for β.