Extensions to the Basic Framework II

Size: px
Start display at page:

Download "Extensions to the Basic Framework II"

Transcription

1 Topic 7 Extensions to the Basic Framework II ARE/ECN 240 A Graduate Econometrics Professor: Òscar Jordà

2 Outline of this topic Nonlinear regression Limited Dependent Variable regression Applications of MLE 2

3 Nonlinear Regression General set-up: y i = m(x i ; )+² i Remarks: Nonlinearity refers specifically to nonlinearity in the parameters. For example, the following regression can be handled with standard regression methods: y 2 i = 0 + 1x i + 2x i + ² i whereas the following regression cannot: y i = x i + ² i 3

4 Remarks (cont.) It will be important to ascertain identification. Remember, in linear regression we rule out multicolinearity, which is an example of lack of identification. I will come back to this issue. I will use the method of moments to think about what the estimator looks like. In general, we will have moment conditions such as: E(R(Z) 0 (y m(x; )) = 0 where R(Z) denotes not just variables Z but also nonlinear transformations of Z. 4

5 Example: R(Z) = 1 Let g i (y i ;X i ; ) =(y i m(x i ; )) = ² i or g(y; X; ) =(y m(x; )) = ² GMM: F.O.C. max ^Q n ( ) = g(y;x; ) 0 ^Wg(y; X; ) Let M n then ^M 0 ^W ^g =0! ^ Assuming a CLT holds such that t 1 nx d p ²i! N(0; Ð); with Ð = E(²² 0 ) =? ¾ 2 I n n i=1 5

6 Example (cont.) Apply M.V.T. to F.O.C. 0= ^M 0 ^W ^g = 0 0 ^M ^Wg+ ^M ^W M( ¹ ^ ) Usual assumptions: ^ p! =) ^M! p M; ¹ 2 [ ^ ; ] =) M ¹! p M; ^W! p W Identification assumption: p n( ^ ) = Ã ^M 0 ^W ¹ M n rank(h) =k and invertible! 1 Ã! ^M 0 ^Wg p n d! N(0;V) V =(M 0 WM) 1 M 0 W ÐWM(M 0 WM) 1 If Ð = ¾ 2 I and W ^W = I then V = ¾ 2 (M 0 M) 1 6

7 AMickeyMouse Mickey-Mouse Example Suppose I want to estimate the model y i = 2x i + ² i ; ² i i:i:d:» D(0;¾ 2 ) Clearly, I could estimate the model y i = x i + ² i p n(^ ) d! N(0;¾ 2 (X 0 X) 1 ) from which and using the delta method: p n(^ 1=2 )! d N(0;¾ )2 (X 0 X) 1 )=N(0; 4 2¾ 2 (X 0 X) 1 ) Using NLS (GMM) and asymptotic formula, verify you get the same answer 7

8 Statistical Properties General Comments Previously we focused on the matrix X and required that t it be full-rank k(i.e., no one variable can be expressed as a linear combination of others). With non-linear functions things are a little trickier now we require the Jacobian M to be full rank. More technically, one could have asymptotic identification even is identification in small samples fails. With non-linear functions, one has to be careful of local optima this will affect asymptotic results which usually assume global optimum. Often not a big deal as long as properties around the global optimum are OK 8

9 Statistical Properties (cont.) Nonlinearities come in many different forms: it is difficult to give general results. I did not talk about consistency because this gets technical really quickly. But generally speaking, the same arguments about normality, the efficient weighting matrix and feasible weighting g schemes go largely through with a little work. One are to be careful about is instrumental variables. Notice the moment condition is E(R(Z) 0 (y m(x; )) = 0 so instruments for X itself are not appropriate. 9

10 Nonlinear IV and NLTSLS GMM is always the safest way to go. But if you do TSLS, make sure the first stage regression refers to the nonlinear transformation of the x, not the x themselves. Example: y = x 2 + u Two possible first stage regressions: x 2 = z + v or x = ¼z + u Basic idea: d(x 2 ) 6= (^x) 2 10

11 Estimating Nonlinear Models Suppose your objective function is twice, continuously differentiable so that a second order Taylor series approximation around the point is: Q ( ) ¼ Q( (0) )+h 0 (0) 1 k (0) ( (0) )+ 1 2 ( (0)) 0 H (0) ( (0) ) k k where h (0) is the gradient of Q evaluated at (0) and where H (0) is the hessian. The first-order conditions for Q* clearly are: h (0) + H (0) ( (0) )=0 =) (1) = (0) H 1 (0) h (0) 11

12 Newton s Method If Q = Q* then clearly we are done but Q* is only an approximation. Newton s method is an algorithm, which applied sequentially, allows us to obtain the optimum of Q (there are mathematical foundations that tell you when the algorithm converges to the optimum and when it does so in finite time). Hence, given a set of initial conditions (arbitrarily chosen), iterate on (j) = (j 1) H 1 (j 1) h (j 1) 12

13 Apply to OLS The OLS objective function is already quadratic and easy to solve. But for illustration purposes: Q( ) =(y X ) 0 (y X ); h( ) = 2X 0 (y X ); H( ) = 2(X 0 X) 0 Suppose I guess (0) = b. Then Newton s method says (1) = b +(2X 0 X) 1 2(X 0 y Xb)=(X 0 X) 1 X 0 y And clearly ^ = (1) regardless of initial conditions. 13

14 GMM Example Suppose I am interested in the usual GMM problem: max ^Q n ( ) = g(y;x;z; ) 0 ^Wg(y;X;Z; ) With possibly nonlinear g. Denote: ^G j ^ Then Newton s method specialized to GMM is: h i 1 h i ^ j+1 = ^ j ^G0 j ^W ^Gj ^G0 j ^W ^gj 14

15 Practicalities Computers have algorithms that calculate derivatives numerically. However, if you can derive the expressions for the gradient and the Hessian analytic, then the optimization algorithm will go much faster. Poorly chosen initial values can result in: (a) slow convergence; and/or (b) getting stuck in local optima. Non-linear optimization is part science, part art. If the function you are optimizing has several local optima, be sure to try different starting values to make sure you get the global optimum 15

16 Stopping Rules When do we stop the algorithm? 3 methods: 1. When: ^ i;j+1 ^ i;j <±for i =1;:::;k for ± a userdefined tolerance level (e.g ) 2. When: Q ( ^ ^ n ( j+1 ) Q n ( j) < ± 3. When: h (j+1) <± When the model is poorly identified (the area around the optimum is very flat), criterion 1 may fail even when 2 and 3 are met. Best to check them all. 16

17 Quasi-Newton Methods Replace: with (j) = (j 1) H 11 h (j 1) h (j 1) (j) = (j 1) 1 (j 1) D (j 1) h (j 1) (j 1) h (j 1) Interpretation: The gradient h gives the direction of most direct improvement in the objective function. The Hessian H is sometimes difficult to obtain so it is useful to use stand-ins, D, that are easier to compute (j) is a step-size that can improve the speed of the algorithm 17

18 Common Algorithms Most have to do with how the Hessian is approximated since the step-size can be adapted to any nx Gauss-Newton/BHHH: i D (j) = ^ ^ (j) Marquardt: if D (j) is not negative definite in BHHH then use: nx D (j) = X i ^ ^ (j) (j) I 18

19 Remarks These algorithms are popular in econometrics because they exploit Fisher s information equality in MLE. Also think about the optimal weighting matrix in GMM. For more complex models, there are nonderivative methods: more robust but much slower. Usually, it is a good idea to mix the methods: use robust methods (such as BHHH/Marquardt or even non-derivative) away from the optimum and then switch to Newton method. 19

20 Binary Variables Binary variable: a variable that takes on values 0 or 1. e.g., 0f for male, 1f for female; 0f for unemployed, 1 for employed. etc. Discrete variable: a variable that takes on a small set of values, e.g. 1 for no high-school, 2 for high-school, 3 for college, 4 for post-graduate Binary and discrete variables are usually handled as any other variable when entering as regressors. The most important issue is to keep in mind that they generate colinearity with the constant term and this needs to be addressed. 20

21 Binary regressor Consider a dummy variable d i 2 0; 1, say 0 for male, 1 for female. Suppose you want to investigate the effect of class size on test scores broken down by boys and girls. Suppose first that sex only affects the average score. Two ways to run the regression: T estscore i = F d i + M(1 d i )+ STR STR i + ² i T estscore i = 0 + F d i + STR ST R i + ² i 21

22 Interpretation In this regression T estscore i = F d i + M(1 d i )+ STR STR i + ² i F is the average score for girls and M the average score for boys, when STR = 0. 0 In this regression T estscore i = 0 + F d i + STR ST R i + ² i 0 is the average score for boys and F is how much better/worse girls are on average with respect to the boys, when STR = 0. 22

23 Interactive binary variables We could also consider that the effect of the treatment itself varies by sex. For example, you may conjecture that boys may need more control in class but girls are more self-disciplined disciplined. Hence you may consider specifying: T estscore i = F d i + M(1 d i )+ STR F STR i d i + STR M STR i (1 d i )+² i 23 But statistically speaking, nothing really changes with respect to what we have discussed in previous topics

24 Limited Dependent Variables Things change when the dependent variable is binary, e.g what determines the decision to rent or own; to be in or out of the labor force; to take public or private transportation; etc. Let y i 2f0; 1g then in the model y i = X i + ² i notice that: E(y i jx i i) ) = X i = P (y i =1jX j i i) ),, but there is nothing that guarantees that Xi X i i ^ 2 f0; 1g 24

25 OLS with limited it dependent d variable Usual linear OLS works (with the caveats just mentioned). Be sure to use heteroskedasticity robust standard errors this is because the variance of y jx is X i (1 X i ) whichvarieswithx i y i jx i OLS is a good exploratory tool but it is conventional to estimate models based on MLE designed to deal with the idiosyncrasies of having a binary dependent variable. 25

26 Example: Boston HMDA Data Source: Alicia H. Munnell, Geoffrey M.B. Tootell, Lynne E. Browne, and James ME McEneaney, Mortgage Lending in Boston: Interpreting HMDA Data, American Economic Review, 1996, pp Here we look at mortgage applications for single- family residences for white and black applicants only. Variables: dependent variable is deny : 1 for application denied, 0 otherwise; pi_rat i is the payment to income ratio; black is 1 if applicant is african-american, 0 otherwise. 26

27 OLS Output. sum deny pi_rat black Variable Obs Mean Std. Dev. Min Max deny pi_rat black reg deny pi_rat black, robust Linear regression Number of obs = 2380 F( 2, 2377) = Prob > F = R-squared = Root MSE = Robust deny Coef. Std. Err. t P> t [95% Conf. Interval] pi_rat black _cons

28 The picture 28

29 Restricting the Response It seems natural then that we would specify the relation between the dependent variable and the regressors as y i = F (X i ) + ² i with F ( 1) =0;F (1) =1; andf(x) but this looks like a CDF df (u) du > 0 29

30 A Latent Variable Model Suppose the model for the latent variable y* is: y i = X i + ² i ; ² i» D(0; 1) Remarks: For reasons that will become clear momentarily, we assume the residuals are standardized. we suppose that y* itself is unobservable but it is related to the binary variable y as follows: y i =1ify i > 0; y i =0ify i 0 Notice then that: P (y i =1)=P (y i > 0) = P (X i + ² i > 0) = P (² i > X X i ) = P (² i X i ) = F (X i ) 30

31 Likelihood of the Latent Variable Model So, if we made an assumption about the distribution of the ² i, specifically, F(.), we could use MLE to estimate the model. Notice that the likelihood function is L(y; )= n i=1p (y i =1) y i P (y i =0) 1 y i n 1 y i = i=1 F (X i ) y i (1 F (X i )) The log-likelihood is: n nx L(y; )= y i log F (X i )+(1 y i )log(1 F (X i )) i=1 y i 31

32 MLE first-order conditions Recall: L(y; )= Then, or nx y i log F (X i )+(1 y i )log(1 F (X i )) = nx yi F ix 0 i 1 y i F ix 0 F i =0; i=1 i 1 F i F i = F (X i ); Fi 0 = F 0 (X i ); F 0 nx yi F i i F i 0 F X i =0 i=1 i(1 F i ) 32

33 Probit and Logit Models 33 The two most common assumptions for F(.) are the normal and the logistic distributions, i.e.: Probit model: F (X i ) = (X i ) Logit model: F (X i ) = exp(x i ) 1+exp(X = 1 i ) 1+exp( X i ) Remarks: for the logit the log-odds ratio is μ Pi log = X i ; P i = P (y i = 1) 1 P i for the logit, the first derivative is such that: 0 F 0 (u) = F (u)f ( u) = F (u)(1 F (u))

34 Logit MLE Recall: L(y; ) = F.O.C. XnX n y i log F (X i ) + (1 y i ) log(1 F (X i )) = n nx i=1 y i F F ix 0 i 1 y i F i 1 F ix 0 i =0; i F i = F (X i ); Fi 0 = F 0 (X i ); F 0 Recall that t for the logit F 0 (u) =F (u)f ( u) =F (u)(1 F (u)) Combining i things 34

35 Logit MLE (cont.) = nx i=1 F (X i )(1 F (X i )) X F (X iy 0 i + i ) F (X i )(1 F (X i )) X 1 F (X i(1 0 y i )=0 i ) nx = X iy i XiF 0 (X i ) =0 i=1 What does this condition remind you of? 35

36 Logit MLE (cont.) Remember, originally we thought of a model such as: y i = F (X i ) + ² i would be good when y is binary. Minimizing i i i the sum of squared residuals results in the GMM moment condition n g n (y i ;X i ; ) = 1 X X n i(y 0 i F (X i )) i=1 And the first order conditions of the GMM problem when ^W = I are just the MLE FOC! 36

37 Logit MLE 2 (y; 0 nx F (X i )(1 F (X i ))X ix 0 i i=1 It turns out the Newton method is particularly well suited for this model since ^ ^ (j+1) = (j) + (X 0 P ^P ^Q 1 (j) Q (j) X) 1 (X 0 y X 0 P ^P (j) ) where ^P (j) is an n n matrix with diagonal elements given by F (X i (j) ^ (j) ) and similarly for but with diagonal elements [1 F (X i ^ (j) )] Q ^Q(j) (j) 37

38 Remarks Because logit/probit models are inherently nonlinear, the can no longer be interpreted as measuring marginal effects directly. The reason we assumed that the model for the latent variable has standardized residuals is that the coefficients are identified up to scale only. Marginal effects (and the interpretation of the ) is different than in linear i =1jX i ij = F 0 (X i ) ij ; where F 0 38

39 Marginal Effects Marginal Effects Depend on Evaluation Point 0.2 Probability Regressor 39

40 HMDA Example Continued. reg deny pi_rat, robust Linear regression Number of obs = 2380 F( 1, 2378) = Prob > F = R-squared = Root MSE = Robust deny Coef. Std. Err. t P> t [95% Conf. Interval] pi_rat _cons probit deny pi_rat Iteration 0: log likelihood = Iteration 1: log likelihood = Iteration 2: log likelihood = Iteration 3: log likelihood = Probit regression Number of obs = 2380 LR chi2(1) = Prob > chi2 = Log likelihood = Pseudo R2 = deny Coef. Std. Err. z P> z [95% Conf. Interval] pi_rat _cons Note: 0 failures and 1 success completely determined.

41 HDMA Example (cont.). probit deny pi_rat Iteration 0: log likelihood = Iteration 1: log likelihood = Iteration 2: log likelihood = Iteration 3: log likelihood = Probit regression Number of obs = 2380 LR chi2(1) = Prob > chi2 = Log likelihood = Pseudo R2 = deny Coef. Std. Err. z P> z [95% Conf. Interval] pi_rat _cons Note: 0 failures and 1 success completely determined.. logit deny pi_rat Iteration 0: log likelihood = Iteration 1: log likelihood = Iteration 2: log likelihood = Iteration 3: log likelihood = Iteration 4: log likelihood = Logistic regression Number of obs = 2380 LR chi2(1) = Prob > chi2 = Log likelihood = Pseudo R2 = deny Coef. Std. Err. z P> z [95% Conf. Interval] pi_rat _cons

42 Comparing Logit and Probit 42

43 How good is the fit? McFadden s (1974) Pseudo-R 2 Pseudo-R 2 =1 P n i=1 y i ln ^p i +(1 y i )ln(1 ^p i ) n[¹y ln ¹y +(1 ¹y)ln(1 ¹y)] Predicted Outcomes: this is tricky to evaluate. Here is the STATA output:. estat classificationl i Classified + if predicted d Pr(D) >=.5 True D defined as deny!= 0 Logistic model for deny True Classified D ~D Total Total Sensitivity Pr( + D) 3.86% Specificity Pr( - ~D) 99.71% Positive predictive value Pr( D +) 64.71% Negative predictive value Pr(~D -) 88.40% False + rate for true ~D Pr( + ~D) 0.29% False - rate for true D Pr( - D) 96.14% False + rate for classified + Pr(~D +) 35.29% False - rate for classified - Pr( D -) 11.60% 43 Looks good!? Correctly classified 88.24%

44 But Remember that there are very few applications denied, so it is easy to just always predict deny = 0 and seemingly do well Classified + if predicted Pr(D) >=.5 True D defined as deny!= 0 Sensitivity Pr( + D) 3.86% Specificity Pr( - ~D) 99.71% Positive predictive value Pr( D +) 64.71% Negative predictive value Pr(~D -) 88.40% False + rate for true ~D Pr( + ~D) 0.29% False - rate for true D Pr( - D) 96.14% False + rate for classified + Pr(~D +) 35.29% False - rate for classified - Pr( D -) 11.60% Correctly classified 88.24% 44

45 HDMA Example (cont.). probit deny pi_rat Iteration 0: log likelihood = Iteration 1: log likelihood = Iteration 2: log likelihood = Iteration 3: log likelihood = Probit regression Number of obs = 2380 LR chi2(1) = Prob > chi2 = Log likelihood = Pseudo R2 = deny Coef. Std. Err. z P> z [95% Conf. Interval] pi_rat _cons Note: 0 failures and 1 success completely determined.. logit deny pi_rat Iteration 0: log likelihood = Iteration 1: log likelihood = Iteration 2: log likelihood = Iteration 3: log likelihood = Iteration 4: log likelihood = Logistic regression Number of obs = 2380 LR chi2(1) = Prob > chi2 = Log likelihood = Pseudo R2 = deny Coef. Std. Err. z P> z [95% Conf. Interval] pi_rat _cons

46 Prediction For example, what is the probability of being denied a loan when the payment-to-income ratio is 60%? No race distinction in the regression. prvalue, x(pi_rat = 0.6) logit: Predictions for deny Confidence intervals by delta method 95% Conf. Interval Pr(y=1 x): [ , ] Pr(y=0 x): [ , ] pi_rat x=.6 46

47 Now including black. logit deny pi_rat black Iteration 0: log likelihood = Iteration 1: log likelihood = Iteration 2: log likelihood = Iteration 3: log likelihood = Iteration 4: log likelihood = Logistic regression Number of obs = 2380 LR chi2(2) = Prob > chi2 = Log likelihood = Pseudo R2 = deny Coef. Std. Err. z P> z [95% Conf. Interval] pi_rat black _cons prvalue, x(pi_rat = 0.6 black = 0) logit: Predictions for deny Confidence intervals by delta method 95% Conf. Interval Pr(y=1 x): [ , ] Pr(y=0 x): [ , ]. prvalue, x(pi_rat = 0.6 black = 1) logit: Predictions for deny Confidence intervals by delta method 95% Conf. Interval Pr(y=1 x): [ , ] Pr(y=0 x): [ , ] pi_rat black pi_rat black x=.6 0 x=

48 Butitdepends depends Same exercise, but now the payment to income ratio is 10%. prvalue, x(pi_rat = 0.1 black = 0) logit: Predictions for deny Confidence intervals by delta method 95% Conf. Interval Pr(y=1 x): [ , ] Pr(y=0 x): [ , ] pi_rat black x=.1 0. prvalue, x(pi_rat = 0.1 black = 1) logit: Predictions for deny Confidence intervals by delta method 95% Conf. Interval Pr(y=1 x): [ , ] Pr(y=0 x): [ , ] pi_rat black x=.1 1 Probably blacks have less income, hence the results when black is excluded. d The rejection rate is twice as high at high payment/income ratios but almost 4 times at low payment/income ratios 48

49 MLE, GMM and the Information Matrix Equality Latent variable models are often estimated with MLE Let me use GMM results to explain the statistical properties of MLE in general settings. Specifically, understanding the information matrix equality is helpful in understanding Quasi-MLE and other related estimation principles 49

50 GMM and MLE In a random sample, the density for the i th observation is: f(y i jx i ; ) Since this is a proper density, it must integrate t to 1 Z A f(y i jx i ; )dy =1 Assuming integrals and derivatives are interchangeable and taking the derivative on both sides: Z i jx i ; ) dy 50

51 GMM and MLE (cont.) Multiplying and dividing by f: Z A That i jx i ; 1 f(y i jx i ; ) f(y ijx i ; )dy =0 Z log f(yi jx i ; ) f(y i jx i ; )dy A h(y i ; )f(y i jx i ; )dy =0 h is called the score and the previous result implies that E[h(y i ; )jx i ]=0! 1 n nx h(y i ; ) =0 i=11 51

52 GMM and MLE Recall that the log-likelihood function for the sample is: L( ) = nx log f(yi jx i ; FOC! 1 X n i j i; ) h(yi i ; ) =0 n i=1 I=1 So if we set up MLE as a GMM problem based on the condition that the expected value of the score is zero (and assuming all the right assumptions are satisfied), and using the optimal weighting matrix, then p d n( ^ )! N(0; (G 0 Ð 1 G) 1 ) 52

53 GMM and MLE Notice that: And G = 1 n nx 2 log f(y i jx i 0 Ð=E[h(y i ; )h(y i ; ) 0 ]! ^Ð = 1 n However, recall that Z log f(y i jx i ; ) f(y i jx i ; )dy nx h(y i ; )h(y i ; ) 0 i=11 A A h(y i ; )f(y i jx i ; )dy =0 And assuming that integration and differentiation are exchangeable 53

54 GMM and MLE Z A then Z i i ; ) f(y i jx i ; )dy f(y i jx i ; )dy + Z A h(y i ; ) Z A h(y i ; ijx i 0 dy =0 1 f(y i jx i ; i jx i ) f(y i jx i ; )dy 0 Z i ; ) f(y i jx i ; )dy Z A h(y i ; ) h 0 (y i ; ) f(y i jx i ; )dy So that 54

55 The Information Matrix Equality Recall Z log f(y i jx i ; f(y i jx i ; )dy = Z A h(y i ; )f(y i jx i ; )dy =0 and from the previous slide Z ; ) f(y i jx i ; )dy A Implying: h(y i ; ) h 0 (y i ; ) f(y i jx i ; ) log f(y ijx i ; log f(y i jx i ; = E log f(y i jx @ 0 This is called the information matrix equality 55

56 Remarks When the likelihood function is correctly specified, the information matrix equality holds and an estimate t of the asymptotic covariance matrix of the parameters estimated can be obtained with either the outerproduct of the scores, or the second-derivative estimate. When the likelihood is incorrectly specified, the information matrix does not hold. However, as long as the score has mean zero, extremum estimation results can be used to show that the estimator is consistent and with asymptotic covariance matrix given by the sandwich c estimator 56

57 Remarks (cont.) Moreover, the information matrix equality is the basis for the BHHH algorithm at the true parameter values, the outer product of the gradients is equivalent to the Hessian. 57

58 Recap 58 FOC of MLE as a moment condition: L( ) = GMM: with G = 1 n nx i=1 nx i=1 log f(y i jx i ; ) FOC! g n ( ) = 1 n nx h(y i ; ) =0 i=1 p n( ^ ) d! N(0; (G 0 Ð 1 G) 1 2 log f(y i jx i 0 n Ð=E[h(y i ; )h(y i ; ) 0 ]! ^Ð = 1 nx h(y i ; )h(y i ; ) 0 n f(yi 2 i ) log ijx i ; log f(y i jx i ; = log f(y ijx @ 0

59 The Newton Method for MLE FOC for MLE And 1 n nx log f(y i jx i ; = 1 n 2 log f(y i jx i ; ) nx h(y i ; ) =0 i=1 G = 1 n 0 The Newton step is (using the information matrix equality): ^ (j+1) = ^ (j) ^G 1 (j)^h (j) ¼ ^ (j) +(^h (j)^h0 (j) ) 1^h(j) 59

Multivariate Regression: Part I

Multivariate Regression: Part I Topic 1 Multivariate Regression: Part I ARE/ECN 240 A Graduate Econometrics Professor: Òscar Jordà Outline of this topic Statement of the objective: we want to explain the behavior of one variable as a

More information

ECON Introductory Econometrics. Lecture 11: Binary dependent variables

ECON Introductory Econometrics. Lecture 11: Binary dependent variables ECON4150 - Introductory Econometrics Lecture 11: Binary dependent variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 11 Lecture Outline 2 The linear probability model Nonlinear probability

More information

Chapter 11. Regression with a Binary Dependent Variable

Chapter 11. Regression with a Binary Dependent Variable Chapter 11 Regression with a Binary Dependent Variable 2 Regression with a Binary Dependent Variable (SW Chapter 11) So far the dependent variable (Y) has been continuous: district-wide average test score

More information

Applied Economics. Regression with a Binary Dependent Variable. Department of Economics Universidad Carlos III de Madrid

Applied Economics. Regression with a Binary Dependent Variable. Department of Economics Universidad Carlos III de Madrid Applied Economics Regression with a Binary Dependent Variable Department of Economics Universidad Carlos III de Madrid See Stock and Watson (chapter 11) 1 / 28 Binary Dependent Variables: What is Different?

More information

2. We care about proportion for categorical variable, but average for numerical one.

2. We care about proportion for categorical variable, but average for numerical one. Probit Model 1. We apply Probit model to Bank data. The dependent variable is deny, a dummy variable equaling one if a mortgage application is denied, and equaling zero if accepted. The key regressor is

More information

Binary Dependent Variable. Regression with a

Binary Dependent Variable. Regression with a Beykent University Faculty of Business and Economics Department of Economics Econometrics II Yrd.Doç.Dr. Özgür Ömer Ersin Regression with a Binary Dependent Variable (SW Chapter 11) SW Ch. 11 1/59 Regression

More information

ECON 594: Lecture #6

ECON 594: Lecture #6 ECON 594: Lecture #6 Thomas Lemieux Vancouver School of Economics, UBC May 2018 1 Limited dependent variables: introduction Up to now, we have been implicitly assuming that the dependent variable, y, was

More information

Binary Dependent Variables

Binary Dependent Variables Binary Dependent Variables In some cases the outcome of interest rather than one of the right hand side variables - is discrete rather than continuous Binary Dependent Variables In some cases the outcome

More information

Extensions to the Basic Framework I

Extensions to the Basic Framework I Topic 5 Extensions to the Basic Framework I ARE/ECN 240 A Graduate Econometrics Professor: Òscar Jordà Outline of this topic Heteroskedasticity: reminder of OLS results and White (1980) corrected standard

More information

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria SOLUTION TO FINAL EXAM Friday, April 12, 2013. From 9:00-12:00 (3 hours) INSTRUCTIONS:

More information

Regression with a Binary Dependent Variable (SW Ch. 9)

Regression with a Binary Dependent Variable (SW Ch. 9) Regression with a Binary Dependent Variable (SW Ch. 9) So far the dependent variable (Y) has been continuous: district-wide average test score traffic fatality rate But we might want to understand the

More information

7/28/15. Review Homework. Overview. Lecture 6: Logistic Regression Analysis

7/28/15. Review Homework. Overview. Lecture 6: Logistic Regression Analysis Lecture 6: Logistic Regression Analysis Christopher S. Hollenbeak, PhD Jane R. Schubart, PhD The Outcomes Research Toolbox Review Homework 2 Overview Logistic regression model conceptually Logistic regression

More information

Econometrics Lecture 5: Limited Dependent Variable Models: Logit and Probit

Econometrics Lecture 5: Limited Dependent Variable Models: Logit and Probit Econometrics Lecture 5: Limited Dependent Variable Models: Logit and Probit R. G. Pierse 1 Introduction In lecture 5 of last semester s course, we looked at the reasons for including dichotomous variables

More information

Homework Solutions Applied Logistic Regression

Homework Solutions Applied Logistic Regression Homework Solutions Applied Logistic Regression WEEK 6 Exercise 1 From the ICU data, use as the outcome variable vital status (STA) and CPR prior to ICU admission (CPR) as a covariate. (a) Demonstrate that

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information

Nonlinear Regression Functions

Nonlinear Regression Functions Nonlinear Regression Functions (SW Chapter 8) Outline 1. Nonlinear regression functions general comments 2. Nonlinear functions of one variable 3. Nonlinear functions of two variables: interactions 4.

More information

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals (SW Chapter 5) Outline. The standard error of ˆ. Hypothesis tests concerning β 3. Confidence intervals for β 4. Regression

More information

Discrete Dependent Variable Models

Discrete Dependent Variable Models Discrete Dependent Variable Models James J. Heckman University of Chicago This draft, April 10, 2006 Here s the general approach of this lecture: Economic model Decision rule (e.g. utility maximization)

More information

Lecture 3.1 Basic Logistic LDA

Lecture 3.1 Basic Logistic LDA y Lecture.1 Basic Logistic LDA 0.2.4.6.8 1 Outline Quick Refresher on Ordinary Logistic Regression and Stata Women s employment example Cross-Over Trial LDA Example -100-50 0 50 100 -- Longitudinal Data

More information

Graduate Econometrics Lecture 4: Heteroskedasticity

Graduate Econometrics Lecture 4: Heteroskedasticity Graduate Econometrics Lecture 4: Heteroskedasticity Department of Economics University of Gothenburg November 30, 2014 1/43 and Autocorrelation Consequences for OLS Estimator Begin from the linear model

More information

Linear Regression With Special Variables

Linear Regression With Special Variables Linear Regression With Special Variables Junhui Qian December 21, 2014 Outline Standardized Scores Quadratic Terms Interaction Terms Binary Explanatory Variables Binary Choice Models Standardized Scores:

More information

Maximum Likelihood (ML) Estimation

Maximum Likelihood (ML) Estimation Econometrics 2 Fall 2004 Maximum Likelihood (ML) Estimation Heino Bohn Nielsen 1of32 Outline of the Lecture (1) Introduction. (2) ML estimation defined. (3) ExampleI:Binomialtrials. (4) Example II: Linear

More information

Marginal Effects for Continuous Variables Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised January 20, 2018

Marginal Effects for Continuous Variables Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised January 20, 2018 Marginal Effects for Continuous Variables Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised January 20, 2018 References: Long 1997, Long and Freese 2003 & 2006 & 2014,

More information

Control Function and Related Methods: Nonlinear Models

Control Function and Related Methods: Nonlinear Models Control Function and Related Methods: Nonlinear Models Jeff Wooldridge Michigan State University Programme Evaluation for Policy Analysis Institute for Fiscal Studies June 2012 1. General Approach 2. Nonlinear

More information

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests ECON4150 - Introductory Econometrics Lecture 5: OLS with One Regressor: Hypothesis Tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 5 Lecture outline 2 Testing Hypotheses about one

More information

Applied Statistics and Econometrics

Applied Statistics and Econometrics Applied Statistics and Econometrics Lecture 6 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 53 Outline of Lecture 6 1 Omitted variable bias (SW 6.1) 2 Multiple

More information

Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018

Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018 Econometrics I KS Module 1: Bivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: March 12, 2018 Alexander Ahammer (JKU) Module 1: Bivariate

More information

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter

More information

Regression #8: Loose Ends

Regression #8: Loose Ends Regression #8: Loose Ends Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #8 1 / 30 In this lecture we investigate a variety of topics that you are probably familiar with, but need to touch

More information

Using the same data as before, here is part of the output we get in Stata when we do a logistic regression of Grade on Gpa, Tuce and Psi.

Using the same data as before, here is part of the output we get in Stata when we do a logistic regression of Grade on Gpa, Tuce and Psi. Logistic Regression, Part III: Hypothesis Testing, Comparisons to OLS Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised January 14, 2018 This handout steals heavily

More information

5. Let W follow a normal distribution with mean of μ and the variance of 1. Then, the pdf of W is

5. Let W follow a normal distribution with mean of μ and the variance of 1. Then, the pdf of W is Practice Final Exam Last Name:, First Name:. Please write LEGIBLY. Answer all questions on this exam in the space provided (you may use the back of any page if you need more space). Show all work but do

More information

CRE METHODS FOR UNBALANCED PANELS Correlated Random Effects Panel Data Models IZA Summer School in Labor Economics May 13-19, 2013 Jeffrey M.

CRE METHODS FOR UNBALANCED PANELS Correlated Random Effects Panel Data Models IZA Summer School in Labor Economics May 13-19, 2013 Jeffrey M. CRE METHODS FOR UNBALANCED PANELS Correlated Random Effects Panel Data Models IZA Summer School in Labor Economics May 13-19, 2013 Jeffrey M. Wooldridge Michigan State University 1. Introduction 2. Linear

More information

Final Exam. Question 1 (20 points) 2 (25 points) 3 (30 points) 4 (25 points) 5 (10 points) 6 (40 points) Total (150 points) Bonus question (10)

Final Exam. Question 1 (20 points) 2 (25 points) 3 (30 points) 4 (25 points) 5 (10 points) 6 (40 points) Total (150 points) Bonus question (10) Name Economics 170 Spring 2004 Honor pledge: I have neither given nor received aid on this exam including the preparation of my one page formula list and the preparation of the Stata assignment for the

More information

Applied Statistics and Econometrics

Applied Statistics and Econometrics Applied Statistics and Econometrics Lecture 5 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 44 Outline of Lecture 5 Now that we know the sampling distribution

More information

Linear Regression with Multiple Regressors

Linear Regression with Multiple Regressors Linear Regression with Multiple Regressors (SW Chapter 6) Outline 1. Omitted variable bias 2. Causality and regression analysis 3. Multiple regression and OLS 4. Measures of fit 5. Sampling distribution

More information

EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7

EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7 Introduction to Generalized Univariate Models: Models for Binary Outcomes EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7 EPSY 905: Intro to Generalized In This Lecture A short review

More information

Nonlinear Econometric Analysis (ECO 722) : Homework 2 Answers. (1 θ) if y i = 0. which can be written in an analytically more convenient way as

Nonlinear Econometric Analysis (ECO 722) : Homework 2 Answers. (1 θ) if y i = 0. which can be written in an analytically more convenient way as Nonlinear Econometric Analysis (ECO 722) : Homework 2 Answers 1. Consider a binary random variable y i that describes a Bernoulli trial in which the probability of observing y i = 1 in any draw is given

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

Econometrics II Tutorial Problems No. 1

Econometrics II Tutorial Problems No. 1 Econometrics II Tutorial Problems No. 1 Lennart Hoogerheide & Agnieszka Borowska 15.02.2017 1 Summary Binary Response Model: A model for a binary (or dummy, i.e. with two possible outcomes 0 and 1) dependent

More information

Practice exam questions

Practice exam questions Practice exam questions Nathaniel Higgins nhiggins@jhu.edu, nhiggins@ers.usda.gov 1. The following question is based on the model y = β 0 + β 1 x 1 + β 2 x 2 + β 3 x 3 + u. Discuss the following two hypotheses.

More information

Ninth ARTNeT Capacity Building Workshop for Trade Research "Trade Flows and Trade Policy Analysis"

Ninth ARTNeT Capacity Building Workshop for Trade Research Trade Flows and Trade Policy Analysis Ninth ARTNeT Capacity Building Workshop for Trade Research "Trade Flows and Trade Policy Analysis" June 2013 Bangkok, Thailand Cosimo Beverelli and Rainer Lanz (World Trade Organization) 1 Selected econometric

More information

Lecture 12: Effect modification, and confounding in logistic regression

Lecture 12: Effect modification, and confounding in logistic regression Lecture 12: Effect modification, and confounding in logistic regression Ani Manichaikul amanicha@jhsph.edu 4 May 2007 Today Categorical predictor create dummy variables just like for linear regression

More information

Gibbs Sampling in Latent Variable Models #1

Gibbs Sampling in Latent Variable Models #1 Gibbs Sampling in Latent Variable Models #1 Econ 690 Purdue University Outline 1 Data augmentation 2 Probit Model Probit Application A Panel Probit Panel Probit 3 The Tobit Model Example: Female Labor

More information

A simple alternative to the linear probability model for binary choice models with endogenous regressors

A simple alternative to the linear probability model for binary choice models with endogenous regressors A simple alternative to the linear probability model for binary choice models with endogenous regressors Christopher F Baum, Yingying Dong, Arthur Lewbel, Tao Yang Boston College/DIW Berlin, U.Cal Irvine,

More information

Binary Logistic Regression

Binary Logistic Regression The coefficients of the multiple regression model are estimated using sample data with k independent variables Estimated (or predicted) value of Y Estimated intercept Estimated slope coefficients Ŷ = b

More information

Binary Choice Models Probit & Logit. = 0 with Pr = 0 = 1. decision-making purchase of durable consumer products unemployment

Binary Choice Models Probit & Logit. = 0 with Pr = 0 = 1. decision-making purchase of durable consumer products unemployment BINARY CHOICE MODELS Y ( Y ) ( Y ) 1 with Pr = 1 = P = 0 with Pr = 0 = 1 P Examples: decision-making purchase of durable consumer products unemployment Estimation with OLS? Yi = Xiβ + εi Problems: nonsense

More information

i (x i x) 2 1 N i x i(y i y) Var(x) = P (x 1 x) Var(x)

i (x i x) 2 1 N i x i(y i y) Var(x) = P (x 1 x) Var(x) ECO 6375 Prof Millimet Problem Set #2: Answer Key Stata problem 2 Q 3 Q (a) The sample average of the individual-specific marginal effects is 0039 for educw and -0054 for white Thus, on average, an extra

More information

At this point, if you ve done everything correctly, you should have data that looks something like:

At this point, if you ve done everything correctly, you should have data that looks something like: This homework is due on July 19 th. Economics 375: Introduction to Econometrics Homework #4 1. One tool to aid in understanding econometrics is the Monte Carlo experiment. A Monte Carlo experiment allows

More information

Recent Developments in Multilevel Modeling

Recent Developments in Multilevel Modeling Recent Developments in Multilevel Modeling Roberto G. Gutierrez Director of Statistics StataCorp LP 2007 North American Stata Users Group Meeting, Boston R. Gutierrez (StataCorp) Multilevel Modeling August

More information

ESTIMATING AVERAGE TREATMENT EFFECTS: REGRESSION DISCONTINUITY DESIGNS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics

ESTIMATING AVERAGE TREATMENT EFFECTS: REGRESSION DISCONTINUITY DESIGNS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics ESTIMATING AVERAGE TREATMENT EFFECTS: REGRESSION DISCONTINUITY DESIGNS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics July 2009 1. Introduction 2. The Sharp RD Design 3.

More information

Consider Table 1 (Note connection to start-stop process).

Consider Table 1 (Note connection to start-stop process). Discrete-Time Data and Models Discretized duration data are still duration data! Consider Table 1 (Note connection to start-stop process). Table 1: Example of Discrete-Time Event History Data Case Event

More information

Generalized Linear Models. Last time: Background & motivation for moving beyond linear

Generalized Linear Models. Last time: Background & motivation for moving beyond linear Generalized Linear Models Last time: Background & motivation for moving beyond linear regression - non-normal/non-linear cases, binary, categorical data Today s class: 1. Examples of count and ordered

More information

ECON Introductory Econometrics. Lecture 17: Experiments

ECON Introductory Econometrics. Lecture 17: Experiments ECON4150 - Introductory Econometrics Lecture 17: Experiments Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 13 Lecture outline 2 Why study experiments? The potential outcome framework.

More information

Lecture 5. In the last lecture, we covered. This lecture introduces you to

Lecture 5. In the last lecture, we covered. This lecture introduces you to Lecture 5 In the last lecture, we covered. homework 2. The linear regression model (4.) 3. Estimating the coefficients (4.2) This lecture introduces you to. Measures of Fit (4.3) 2. The Least Square Assumptions

More information

Week 7: Binary Outcomes (Scott Long Chapter 3 Part 2)

Week 7: Binary Outcomes (Scott Long Chapter 3 Part 2) Week 7: (Scott Long Chapter 3 Part 2) Tsun-Feng Chiang* *School of Economics, Henan University, Kaifeng, China April 29, 2014 1 / 38 ML Estimation for Probit and Logit ML Estimation for Probit and Logit

More information

Chapter 9 Regression with a Binary Dependent Variable. Multiple Choice. 1) The binary dependent variable model is an example of a

Chapter 9 Regression with a Binary Dependent Variable. Multiple Choice. 1) The binary dependent variable model is an example of a Chapter 9 Regression with a Binary Dependent Variable Multiple Choice ) The binary dependent variable model is an example of a a. regression model, which has as a regressor, among others, a binary variable.

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Limited Dependent Variable Models II

Limited Dependent Variable Models II Limited Dependent Variable Models II Fall 2008 Environmental Econometrics (GR03) LDV Fall 2008 1 / 15 Models with Multiple Choices The binary response model was dealing with a decision problem with two

More information

4 Instrumental Variables Single endogenous variable One continuous instrument. 2

4 Instrumental Variables Single endogenous variable One continuous instrument. 2 Econ 495 - Econometric Review 1 Contents 4 Instrumental Variables 2 4.1 Single endogenous variable One continuous instrument. 2 4.2 Single endogenous variable more than one continuous instrument..........................

More information

Logistic & Tobit Regression

Logistic & Tobit Regression Logistic & Tobit Regression Different Types of Regression Binary Regression (D) Logistic transformation + e P( y x) = 1 + e! " x! + " x " P( y x) % ln$ ' = ( + ) x # 1! P( y x) & logit of P(y x){ P(y

More information

Problem Set 10: Panel Data

Problem Set 10: Panel Data Problem Set 10: Panel Data 1. Read in the data set, e11panel1.dta from the course website. This contains data on a sample or 1252 men and women who were asked about their hourly wage in two years, 2005

More information

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N 1) [4 points] Let Econ 836 Final Exam Y Xβ+ ε, X w+ u, w N w~ N(, σi ), u N u~ N(, σi ), ε N ε~ Nu ( γσ, I ), where X is a just one column. Let denote the OLS estimator, and define residuals e as e Y X.

More information

4 Instrumental Variables Single endogenous variable One continuous instrument. 2

4 Instrumental Variables Single endogenous variable One continuous instrument. 2 Econ 495 - Econometric Review 1 Contents 4 Instrumental Variables 2 4.1 Single endogenous variable One continuous instrument. 2 4.2 Single endogenous variable more than one continuous instrument..........................

More information

Modeling Binary Outcomes: Logit and Probit Models

Modeling Binary Outcomes: Logit and Probit Models Modeling Binary Outcomes: Logit and Probit Models Eric Zivot December 5, 2009 Motivating Example: Women s labor force participation y i = 1 if married woman is in labor force = 0 otherwise x i k 1 = observed

More information

Interpreting coefficients for transformed variables

Interpreting coefficients for transformed variables Interpreting coefficients for transformed variables! Recall that when both independent and dependent variables are untransformed, an estimated coefficient represents the change in the dependent variable

More information

Section Least Squares Regression

Section Least Squares Regression Section 2.3 - Least Squares Regression Statistics 104 Autumn 2004 Copyright c 2004 by Mark E. Irwin Regression Correlation gives us a strength of a linear relationship is, but it doesn t tell us what it

More information

Sociology 362 Data Exercise 6 Logistic Regression 2

Sociology 362 Data Exercise 6 Logistic Regression 2 Sociology 362 Data Exercise 6 Logistic Regression 2 The questions below refer to the data and output beginning on the next page. Although the raw data are given there, you do not have to do any Stata runs

More information

Single-level Models for Binary Responses

Single-level Models for Binary Responses Single-level Models for Binary Responses Distribution of Binary Data y i response for individual i (i = 1,..., n), coded 0 or 1 Denote by r the number in the sample with y = 1 Mean and variance E(y) =

More information

Machine Learning Lecture 7

Machine Learning Lecture 7 Course Outline Machine Learning Lecture 7 Fundamentals (2 weeks) Bayes Decision Theory Probability Density Estimation Statistical Learning Theory 23.05.2016 Discriminative Approaches (5 weeks) Linear Discriminant

More information

1 Outline. 1. Motivation. 2. SUR model. 3. Simultaneous equations. 4. Estimation

1 Outline. 1. Motivation. 2. SUR model. 3. Simultaneous equations. 4. Estimation 1 Outline. 1. Motivation 2. SUR model 3. Simultaneous equations 4. Estimation 2 Motivation. In this chapter, we will study simultaneous systems of econometric equations. Systems of simultaneous equations

More information

Section I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem -

Section I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem - First Exam: Economics 388, Econometrics Spring 006 in R. Butler s class YOUR NAME: Section I (30 points) Questions 1-10 (3 points each) Section II (40 points) Questions 11-15 (10 points each) Section III

More information

ECON3150/4150 Spring 2015

ECON3150/4150 Spring 2015 ECON3150/4150 Spring 2015 Lecture 3&4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo January 29, 2015 1 / 67 Chapter 4 in S&W Section 17.1 in S&W (extended OLS assumptions) 2

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,

More information

Lab 07 Introduction to Econometrics

Lab 07 Introduction to Econometrics Lab 07 Introduction to Econometrics Learning outcomes for this lab: Introduce the different typologies of data and the econometric models that can be used Understand the rationale behind econometrics Understand

More information

Instrumental Variables, Simultaneous and Systems of Equations

Instrumental Variables, Simultaneous and Systems of Equations Chapter 6 Instrumental Variables, Simultaneous and Systems of Equations 61 Instrumental variables In the linear regression model y i = x iβ + ε i (61) we have been assuming that bf x i and ε i are uncorrelated

More information

GMM Estimation in Stata

GMM Estimation in Stata GMM Estimation in Stata Econometrics I Department of Economics Universidad Carlos III de Madrid Master in Industrial Economics and Markets 1 Outline Motivation 1 Motivation 2 3 4 2 Motivation 3 Stata and

More information

raise Coef. Std. Err. z P> z [95% Conf. Interval]

raise Coef. Std. Err. z P> z [95% Conf. Interval] 1 We will use real-world data, but a very simple and naive model to keep the example easy to understand. What is interesting about the example is that the outcome of interest, perhaps the probability or

More information

Assessing the Calibration of Dichotomous Outcome Models with the Calibration Belt

Assessing the Calibration of Dichotomous Outcome Models with the Calibration Belt Assessing the Calibration of Dichotomous Outcome Models with the Calibration Belt Giovanni Nattino The Ohio Colleges of Medicine Government Resource Center The Ohio State University Stata Conference -

More information

Lab 11 - Heteroskedasticity

Lab 11 - Heteroskedasticity Lab 11 - Heteroskedasticity Spring 2017 Contents 1 Introduction 2 2 Heteroskedasticity 2 3 Addressing heteroskedasticity in Stata 3 4 Testing for heteroskedasticity 4 5 A simple example 5 1 1 Introduction

More information

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han Econometrics Honor s Exam Review Session Spring 2012 Eunice Han Topics 1. OLS The Assumptions Omitted Variable Bias Conditional Mean Independence Hypothesis Testing and Confidence Intervals Homoskedasticity

More information

Lecture 2: Poisson and logistic regression

Lecture 2: Poisson and logistic regression Dankmar Böhning Southampton Statistical Sciences Research Institute University of Southampton, UK S 3 RI, 11-12 December 2014 introduction to Poisson regression application to the BELCAP study introduction

More information

Lecture 10: Alternatives to OLS with limited dependent variables. PEA vs APE Logit/Probit Poisson

Lecture 10: Alternatives to OLS with limited dependent variables. PEA vs APE Logit/Probit Poisson Lecture 10: Alternatives to OLS with limited dependent variables PEA vs APE Logit/Probit Poisson PEA vs APE PEA: partial effect at the average The effect of some x on y for a hypothetical case with sample

More information

Introduction to Econometrics

Introduction to Econometrics Introduction to Econometrics T H I R D E D I T I O N Global Edition James H. Stock Harvard University Mark W. Watson Princeton University Boston Columbus Indianapolis New York San Francisco Upper Saddle

More information

Econometrics Midterm Examination Answers

Econometrics Midterm Examination Answers Econometrics Midterm Examination Answers March 4, 204. Question (35 points) Answer the following short questions. (i) De ne what is an unbiased estimator. Show that X is an unbiased estimator for E(X i

More information

Comparing groups using predicted probabilities

Comparing groups using predicted probabilities Comparing groups using predicted probabilities J. Scott Long Indiana University May 9, 2006 MAPSS - May 9, 2006 - Page 1 The problem Allison (1999): Di erences in the estimated coe cients tell us nothing

More information

Lecture 5: Poisson and logistic regression

Lecture 5: Poisson and logistic regression Dankmar Böhning Southampton Statistical Sciences Research Institute University of Southampton, UK S 3 RI, 3-5 March 2014 introduction to Poisson regression application to the BELCAP study introduction

More information

Generalized Linear Models for Non-Normal Data

Generalized Linear Models for Non-Normal Data Generalized Linear Models for Non-Normal Data Today s Class: 3 parts of a generalized model Models for binary outcomes Complications for generalized multivariate or multilevel models SPLH 861: Lecture

More information

Instrumental Variable Regression

Instrumental Variable Regression Topic 6 Instrumental Variable Regression ARE/ECN 240 A Graduate Econometrics Professor: Òscar Jordà Outline of this topic Randomized Experiments, natural experiments and causation Instrumental variables:

More information

Essential of Simple regression

Essential of Simple regression Essential of Simple regression We use simple regression when we are interested in the relationship between two variables (e.g., x is class size, and y is student s GPA). For simplicity we assume the relationship

More information

Logistic regression: Why we often can do what we think we can do. Maarten Buis 19 th UK Stata Users Group meeting, 10 Sept. 2015

Logistic regression: Why we often can do what we think we can do. Maarten Buis 19 th UK Stata Users Group meeting, 10 Sept. 2015 Logistic regression: Why we often can do what we think we can do Maarten Buis 19 th UK Stata Users Group meeting, 10 Sept. 2015 1 Introduction Introduction - In 2010 Carina Mood published an overview article

More information

Non-linear panel data modeling

Non-linear panel data modeling Non-linear panel data modeling Laura Magazzini University of Verona laura.magazzini@univr.it http://dse.univr.it/magazzini May 2010 Laura Magazzini (@univr.it) Non-linear panel data modeling May 2010 1

More information

ECONOMETRICS HONOR S EXAM REVIEW SESSION

ECONOMETRICS HONOR S EXAM REVIEW SESSION ECONOMETRICS HONOR S EXAM REVIEW SESSION Eunice Han ehan@fas.harvard.edu March 26 th, 2013 Harvard University Information 2 Exam: April 3 rd 3-6pm @ Emerson 105 Bring a calculator and extra pens. Notes

More information

Practical Econometrics. for. Finance and Economics. (Econometrics 2)

Practical Econometrics. for. Finance and Economics. (Econometrics 2) Practical Econometrics for Finance and Economics (Econometrics 2) Seppo Pynnönen and Bernd Pape Department of Mathematics and Statistics, University of Vaasa 1. Introduction 1.1 Econometrics Econometrics

More information

ECONOMICS AND ECONOMIC METHODS PRELIM EXAM Statistics and Econometrics August 2013

ECONOMICS AND ECONOMIC METHODS PRELIM EXAM Statistics and Econometrics August 2013 ECONOMICS AND ECONOMIC METHODS PRELIM EXAM Statistics and Econometrics August 2013 Instructions: Answer all six (6) questions. Point totals for each question are given in parentheses. The parts within

More information

LOGISTIC REGRESSION Joseph M. Hilbe

LOGISTIC REGRESSION Joseph M. Hilbe LOGISTIC REGRESSION Joseph M. Hilbe Arizona State University Logistic regression is the most common method used to model binary response data. When the response is binary, it typically takes the form of

More information

fhetprob: A fast QMLE Stata routine for fractional probit models with multiplicative heteroskedasticity

fhetprob: A fast QMLE Stata routine for fractional probit models with multiplicative heteroskedasticity fhetprob: A fast QMLE Stata routine for fractional probit models with multiplicative heteroskedasticity Richard Bluhm May 26, 2013 Introduction Stata can easily estimate a binary response probit models

More information

A Course in Applied Econometrics Lecture 14: Control Functions and Related Methods. Jeff Wooldridge IRP Lectures, UW Madison, August 2008

A Course in Applied Econometrics Lecture 14: Control Functions and Related Methods. Jeff Wooldridge IRP Lectures, UW Madison, August 2008 A Course in Applied Econometrics Lecture 14: Control Functions and Related Methods Jeff Wooldridge IRP Lectures, UW Madison, August 2008 1. Linear-in-Parameters Models: IV versus Control Functions 2. Correlated

More information

Measurement Error. Often a data set will contain imperfect measures of the data we would ideally like.

Measurement Error. Often a data set will contain imperfect measures of the data we would ideally like. Measurement Error Often a data set will contain imperfect measures of the data we would ideally like. Aggregate Data: (GDP, Consumption, Investment are only best guesses of theoretical counterparts and

More information

Greene, Econometric Analysis (6th ed, 2008)

Greene, Econometric Analysis (6th ed, 2008) EC771: Econometrics, Spring 2010 Greene, Econometric Analysis (6th ed, 2008) Chapter 17: Maximum Likelihood Estimation The preferred estimator in a wide variety of econometric settings is that derived

More information

Generalized Linear Models Introduction

Generalized Linear Models Introduction Generalized Linear Models Introduction Statistics 135 Autumn 2005 Copyright c 2005 by Mark E. Irwin Generalized Linear Models For many problems, standard linear regression approaches don t work. Sometimes,

More information