of the matrix is =-85, so it is not positive definite. Thus, the first

Similar documents
Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators

Properties and Hypothesis Testing

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes.

Random Variables, Sampling and Estimation

Lecture 7: Properties of Random Samples

Mathematical Statistics - MS

Simulation. Two Rule For Inverting A Distribution Function

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

Lecture 33: Bootstrap

Questions and Answers on Maximum Likelihood

TAMS24: Notations and Formulas

4. Partial Sums and the Central Limit Theorem

17. Joint distributions of extreme order statistics Lehmann 5.1; Ferguson 15

Direction: This test is worth 150 points. You are required to complete this test within 55 minutes.

Bayesian Methods: Introduction to Multi-parameter Models

Asymptotics. Hypothesis Testing UMP. Asymptotic Tests and p-values

Review Questions, Chapters 8, 9. f(y) = 0, elsewhere. F (y) = f Y(1) = n ( e y/θ) n 1 1 θ e y/θ = n θ e yn

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise

Stat 319 Theory of Statistics (2) Exercises

STATISTICAL INFERENCE

Stat 200 -Testing Summary Page 1

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting

Efficient GMM LECTURE 12 GMM II

Chapter 2 The Monte Carlo Method

Last Lecture. Wald Test

Homework 3 Solutions

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting

IIT JAM Mathematical Statistics (MS) 2006 SECTION A

( θ. sup θ Θ f X (x θ) = L. sup Pr (Λ (X) < c) = α. x : Λ (x) = sup θ H 0. sup θ Θ f X (x θ) = ) < c. NH : θ 1 = θ 2 against AH : θ 1 θ 2

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

Statistical Theory MT 2008 Problems 1: Solution sketches

Math 152. Rumbos Fall Solutions to Review Problems for Exam #2. Number of Heads Frequency

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d

Problem Set 4 Due Oct, 12

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain

Statistical Theory MT 2009 Problems 1: Solution sketches

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

EE 4TM4: Digital Communications II Probability Theory

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

Exponential Families and Bayesian Inference

1.010 Uncertainty in Engineering Fall 2008

Logit regression Logit regression

STAC51: Categorical data Analysis

Regression with an Evaporating Logarithmic Trend

Linear Regression Models

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator

Rule of probability. Let A and B be two events (sets of elementary events). 11. If P (AB) = P (A)P (B), then A and B are independent.

Expectation and Variance of a random variable

Statistical Inference Based on Extremum Estimators

Maximum Likelihood Estimation

Circle the single best answer for each multiple choice question. Your choice should be made clearly.

Probability and Statistics

November 2002 Course 4 solutions

S Y Y = ΣY 2 n. Using the above expressions, the correlation coefficient is. r = SXX S Y Y

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering

Probability and statistics: basic terms

LECTURE 8: ASYMPTOTICS I

Stat 139 Homework 7 Solutions, Fall 2015

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

Joint Probability Distributions and Random Samples. Jointly Distributed Random Variables. Chapter { }

TMA4245 Statistics. Corrected 30 May and 4 June Norwegian University of Science and Technology Department of Mathematical Sciences.

Lecture 18: Sampling distributions

Access to the published version may require journal subscription. Published with permission from: Elsevier.

Kurskod: TAMS11 Provkod: TENB 21 March 2015, 14:00-18:00. English Version (no Swedish Version)

5. Likelihood Ratio Tests

Final Examination Solutions 17/6/2010

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen)

Univariate Normal distribution. whereaandbareconstants. Theprobabilitydensityfunction(PDFfromnowon)ofZ andx is. ) 2π.

Economics 326 Methods of Empirical Research in Economics. Lecture 18: The asymptotic variance of OLS and heteroskedasticity

1 Inferential Methods for Correlation and Regression Analysis

Final Review. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

2. The volume of the solid of revolution generated by revolving the area bounded by the

1 Models for Matched Pairs

Important Formulas. Expectation: E (X) = Σ [X P(X)] = n p q σ = n p q. P(X) = n! X1! X 2! X 3! X k! p X. Chapter 6 The Normal Distribution.

7-1. Chapter 4. Part I. Sampling Distributions and Confidence Intervals

Asymptotic Results for the Linear Regression Model

Section 14. Simple linear regression.

Lecture 2: Monte Carlo Simulation

Solutions to Odd Numbered End of Chapter Exercises: Chapter 4

Linear Regression Models, OLS, Assumptions and Properties

CS284A: Representations and Algorithms in Molecular Biology

Homework for 2/3. 1. Determine the values of the following quantities: a. t 0.1,15 b. t 0.05,15 c. t 0.1,25 d. t 0.05,40 e. t 0.

Common Large/Small Sample Tests 1/55

Regression, Inference, and Model Building

MATH 472 / SPRING 2013 ASSIGNMENT 2: DUE FEBRUARY 4 FINALIZED

The Sampling Distribution of the Maximum. Likelihood Estimators for the Parameters of. Beta-Binomial Distribution

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

BIOSTATISTICS. Lecture 5 Interval Estimations for Mean and Proportion. dr. Petr Nazarov

Final Examination Statistics 200C. T. Ferguson June 10, 2010

This section is optional.

10-701/ Machine Learning Mid-term Exam Solution

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

Chi-Squared Tests Math 6070, Spring 2006

Interval Estimation (Confidence Interval = C.I.): An interval estimate of some population parameter is an interval of the form (, ),

Investigating the Significance of a Correlation Coefficient using Jackknife Estimates

Overview. p 2. Chapter 9. Pooled Estimate of. q = 1 p. Notation for Two Proportions. Inferences about Two Proportions. Assumptions

Transcription:

BOSTON COLLEGE Departmet of Ecoomics EC771: Ecoometrics Sprig 4 Prof. Baum, Ms. Uysal Solutio Key for Problem Set 1 1. Are the followig quadratic forms positive for all values of x? (a) y = x 1 8x 1 x + (11x ). (b) y = 5x 1 + x + 7x 3 + 4x 1 x + 6x 1 x 3 + 8x x 3. The first may be writte [ ] [ ] [ ] 1 14 x1 x 1 x. The determiat of the matrix is 11-196=-85, so it is ot positive defiite. Thus, the first 14 11 x quadratic form eed ot be positive. The secod uses the matrix 5 3 1 4 3 4 7 There are several ways to check the defiiteess of a matrix. Oe way is to check the sigs of the pricipal miors, which must be positive. The first two are 5 ad 5(1)-()=1, but the third, determiat is -34. Therefore, the matrix is ot positive defiite. Its three characteristic roots are 11.1,.9, ad -1. It follows, therefore, that there are values of x 1, x, ad x 3 for which the quadratic form is egative.. Compute the characteristic roots of 4 3 A = 4 8 6. 3 6 5. The roots are determied by A-λI =. For the matrix above, this is A-λI = ( λ)(8 λ)(5 λ) + 7 + 7 9(8 λ) 36( λ) 16(5 λ) = λ 3 + 15λ 5λ = λ(λ 15λ + 5) =. Oe solutio is obviously zero.(this might have bee apparet. The secod colum of the matrix is twice the first, so it has rak o more tha two, ad therefore o more tha two ozero roots.) The other two roots are (15± 5)/=.341 ad 4.659. 3. If x has a ormal distributio with mea 1 ad stadard deviatio 3, what are the followig? (a) P rob [ x > ]. (b) P rob [x > 1 x < 1.5]. Usig the ormal table, (a)p rob [ x > ] = 1 P rob [ x ] 1

= 1 P rob [ x ] = 1 P rob [( 1)/3 z ( 1)/3] = 1 [F (1/3) F ( 1)] = 1.636 +.1587 =.581. (b)p rob [x > 1 x < 1.5] = P rob [ 1 < x < 1.5] /P rob [x < 1.5] P rob [ 1 < x < 1.5] = P rob [( 1 1)/3 < z < (1.5 1)3] = P rob [z < 1/6] P rob [z < /3] =.566.55 =.3137. The coditioal probability is.3137/.566=.554. 4. If x has a ormal distributio with mea µ ad stadard deviatio σ, what is the probability distributio of y = e x? If y = e x, the x = ly ad the Jacobia is dx/dy = 1/y. substitutio, 1 f(y) = σy π e 1 [(ly µ)/σ] Makig the This is the desity of the logormal distributio. 5. The followig sample is draw from a ormal distributio with mea µ ad stadard deviatio σ: x = 1.3,.1,.4, 1.3,.5,., 1.8,.5, 1.9, 3.. Usig the data, test the followig hypotheses: (a) µ., (b) µ.7, (c) σ =.5. (d)usig a likelihood ratio test, test the hypothesis µ = 1.8, σ =.8. Mea, variace ad stadard deviatio of the sample are as follows, x = x i = 1.5, s = (x i x) =.9418, 1 s =.97 (a) We would reject the hypothesis if 1.5 is too small relative to the hypothesized value of. Sice the data are sampled from a ormal distributio, we may use a t-test to test the hypothesis. The t-ratio is t[9] = (1.5 )/[.97/ 1] = 1.47.

The 95% critical value from the t-distributio for a oe tailed test is -1.833. Therefore, we would ot reject the hypothesis at a sigificace level of 95%. (b) We would reject the hypothesis if 1.5 is excessively large relative to the hypothesized mea of.7. The t-ratio is t[9] = (1.5.7)/[.97/ 1] =.673. Usig the same critical value as i the previous problem, we would reject this hypothesis. (c) The statistic ( 1)s /σ is distributed as χ with 9 degrees of freedom. This is 9(.94)/.5=16.9. The 95% critical values from the chi-squared table for a two tailed test are.7 ad 19.. Thus we would ot reject the hypothesis. (d) The log-likelihood for a sample from a ormal distributio is l L = (/) l(π) (/) l σ 1 σ (x i µ) The sample values are ˆµ = x = 1.5, ˆσ = (xi x) =.8476. The maximized log-likelihood for the sample is -13.363. A useful shortcut for computig the log-likelihood at the hypothesized value is (x i µ) = (x i x) + ( x µ). For the hypothesized value of µ = 1.8, this is (x i 1.8) = 9.6. The log-likelihood is 5 l(π) 5 l.8 (1/1.6)9.6 = 13.861. The likelihood ratio statistic is -(l L r l L u )=.996. The critical value for a chi-squared with degrees of freedom is 5.99, so we would ot reject the hypothesis. 6. A commo method of simulatig radom draws from the stadard ormal distributio is to compute the sum of 1 draws from the uiform [,1] distributio ad subtract 6. Ca you justify this procedure? The uiform distributio has mea ad variace 1/1. Therefore, the statistic 1( x 1/) = 1 x i 6 is equivalet to z = ( x µ)/σ. As, this coverges to a stadard ormal variable. Experiece suggests that a sample of 1 is large eough to approximate this result. However, more recetly developed radom umber geerators usually use differet procedures based o the trucatio error which occurs i represetig real umbers i a digital computer. 7. The radom variable x has a cotiuous distributio f(x) ad cumulative distributio fuctio F (x). What is the probability distributio of the sample maximum?[hit: I a radom sample of observatios, x 1, x,..., x, if z is the maximum, the every observatio i the sample is less tha or equal to z. Use the cdf.] If z is the maximum, the every sample observatio is less tha or equal to z. The probability of this is Prob[x 1 z, x z,..., x z]=f (z)f (z)... F (z) = [F (z)]. The desity is the derivative, [F (z)] 1 f(z). 3

8. Testig for ormality. Oe method that has bee suggested for testig whether the distributio uderlyig a sample is ormal is to refer the statistics L = [skewess /6 + (kurtosis 3) /4] to the chi-squared distributio with two degrees of freedom. Usig the data i Exercise 5, carry out the test. The skewess coefficiet is.1419 ad the kurtosis is 1.8447. (These are the third ad fourth momets divided by the third ad fourth power of the sample stadard deviatio.) Isertig these i the expressio i the questio produces L= 1{.1419 /6 + (1.8447 3) /4} =.59. The critical value from the chi-squared distributio with degrees of freedom (95%) is 5.99. Thus, the hypothesis of ormality caot be rejected. 9. Mixture distributio. Suppose that the joit distributio of the two radom variables x ad y is f(x, y) = θe (β+θ)y (βy) x, β, θ >, y, x =, 1,,... x! Fid the maximum likelihood estimators of β ad θ ad their asymptotic joit distributio. The log-likelihood is l L = l θ (β + θ) y i + l β x i + x ilogy i log(x i!) The first ad secod derivatives are L/ θ = /θ l L/ β = y i + y i l L/ θ = /θ l L/ θ = x i /β l L/ β θ =. x i /β Therefore, the maximum likelihood estimators are[ ˆθ = 1/ȳ ad ˆβ = x/ȳ ad the /θ asymptotic covariace matrix is the iverse of E ]. x i/β I order to complete the derivatio, we will require the expected value of x i = E[x i ]. I order to obtai E[x i ], it is ecessary to obtai the margial distributio of x i, which is f(x) = θe (β+θ)y (βy) x /x! dy = β x (θ/x!) e (β+θ)y y x dy. This is β x (θ/x!) times a gamma itegral. This is f(x) = β x (θ/x!)[γ(x+1)]/(β + θ) x+1. But, Γ(x + 1) = x!, so the expressio reduces to f(x) = [θ/(β + θ)][β/(β + θ)] x. 4

Thus, x has a geometric distributio with parameter π = θ/(β +θ). (This is the distributio of the umber of tries util the first success of idepedet trials each with success probability 1-π.) Fially, we require the expected value of x i, which is E[x] = [θ/(β + θ)] x[β/(β + θ)] x ] = β/θ. x= The, the required asymptotic covariace matrix is [ ] /θ 1 [ /θ (β/θ)/β = (βθ)/ ]. 1. Empirical exercise. Usig the caed dataset discrim, which you may access with the commad.use http://fmwww.bc.edu/ec-p/data/wooldridge/discrim (a) How may of the observatios are from Pesylvaia? (hit: describe may be useful) (b) What is the average startig wage (first wave)? What are the mi ad max of this variable? (c) Test the hypothesis that average icomes i NJ ad PA are equal. (hit: help ttest) (d) Test the hypothesis that the average price of a etree (first wave) was $1.39. (e) Write a Stata program, usig the ml sytax, which will estimate the parameters of a k-variable liear regressio model with a costat term, icludig the σ parameter, via maximum likelihood. Use your program to estimate the model: pfries i = β + β 1 icome i + β prpblck i + ɛ i over the whole sample, ad (f) oly over the New Jersey observatios (those for which state equals 1). (g) Use Stata s regress commad to estimate the same liear regressio, ad use bootstrap to geerate bootstrap stadard errors for the ˆβ parameters. Discuss how the bootstrap cofidece itervals for icome ad prpblk compare with the covetioal cofidece itervals computed by regress. Tur i a pritout of your program illustratig the results of each estimatio. 5