Link to Paper. The latest iteration can be found at:
|
|
- Nora Douglas
- 6 years ago
- Views:
Transcription
1 Link to Paper Introduction The latest iteration can be found at:
2 BKW dignostics in GRETL: Interpretation and Performance Oklahoma State University 5th Gretl Conference, 2017
3 Table of Contents Introduction 1 Introduction 2 3 Example Simulation Analysis Example Simulation Analysis 4
4 Consequences of Weak Identification Parameters estimated imprecisely Estimated covariance nonsingular or not positive definite Excessive numbers of iterations or numerical algorithms don t converge Distorts the distribution of the estimator (e.g., IV estimator with weak instruments) Use the BKW diagnostics to diagnose identification problems in nonlinear models.
5 BKW Table Introduction Condition Variance Proportions of OLS Index var(b 1 ) var(b 2 ) var(b k ) η 1 φ 11 φ 12 φ 1k η 1 φ 21 φ 22 φ 2k... η k φ k1 φ k2 φ kk Table: Matrix of Variance Proportions In linear models, the Table summarizes much of what we can learn about collinearity in data.
6 Condition Index Introduction The BKW diagnostics use the orthonormal eigenvectors, V, and eigenvalues from the SVD of Cov( ˆβ) 1 = UΛV T, where the columns of the Cov( ˆβ) have been scaled to unit length. The condition index is η l = where η 1 = 1 < η 2 <... < η l. ( λ1 λ l ) 1 2.
7 Variance Decomposition Define φ jk = v 2 jk λ k, and let φ j be the variance of b j, apart from the error variance, σ 2. Var(b j ) = σ 2 φ j = ( v 2 j1 + v j v jk 2 λ 1 λ 2 λ k The proportion of the variance of b j associated with the k th eigenvalue λ k is φ jk φ j. )
8 Recipe Introduction Check the condition numbers. Anything larger than 30 indicates a severe problem. 1 condition number is large: read variance decompositions across its row of the table. Multiple φ jk > 50% indicate weakness. 2 large condition numbers: sum the variance decomp s φ jk across j φ jk and repeat. Check to see if any parameters are well identified. Single large condition numbers with variance < 50% is good. For multiple large η, sum φ jk
9 Introduction The BKW diagnostics can be used in any estimator whose covariance can be estimated. The s/n diagnostics can be used for estimators that are normally, or approximately normally distributed. 1 2 Endogenous
10 Introduction Following (Greene, 2012, p ) the ordered probit model is treated as a latent variable, y that depends on a linear index, x T β. y = x T β + e The latent variable y is unobserved. Instead we observe integers, y, such that y=0 if y 0 y=1 if 0 y τ 1 y=2 if τ 1 y τ 2. y=j. if τ J 1 y
11 The β and the τs are unknown parameters. If e is assumed to be normally distributed across observations. Since the overall variance is unidentified by the model it is conventional to let σ = 1 in order to identify β. Then, Prob(y = 0 x) = Φ( x T β) Prob(y = 1 x) = Φ(τ 1 x T β) Φ( x T β) Prob(y = 2 x) = Φ(τ 2 x T β) Φ(τ 1 x T β). Prob(y = J x) = 1 Φ(τ J 1 x T β) with 0 < τ 1 < τ 2 <... < τ J 1 and Φ(t) is the standard normal cdf evaluated at t.
12 example Model 2:, using observations Dependent variable: kidsl6 Standard errors based on Hessian Coefficient Std. Error z p-value educ exper age cut cut cut
13 BKW table from GRETL The vif command in GRETL produces the BKW table as follows: --- variance proportions --- lambda cond educ exper age
14 BKW table from GRETL, cont. The RHS of the vif command in GRETL: --- variance proportions --- lambda cond cut1 cut2 cut
15 Design Introduction There are three regressors in the model. X = {const, age, educ, exper}. (1) is decomposed using SVD. X = UDV, where D is a diagonal matrix containing the eigenvalues of the original data. These are replaced using the desired ones Λ as in. X = UΛV
16 Collinear Regressors Introduction Five sets of eigenvalues for the SVD are considered. Eigenvalues Λ = {1, 1, 1, 1} Λ = {10, 7, 4, 1} Λ = {10, 1, 1, 0.1} Λ = {10, 10, 0.1, 0.1} Λ = {10, 0.1, 0.05, 0.05} Collinearity None Moderate Moderate Moderately Severe Severe (2)
17 Setting cutoff points, τ j The cutoff parameters are set by taking the mean of Xβ and adding c such that P(X < c) = p with the parameter p chosen from Cumulative Prob Bin Size p = {0.1, 0.5, 0.9} Wide (3) p = {0.1, 0.3, 0.5} Moderate p = {0.1, 0.2, 0.3} Narrow The parameters from the regression are set to β is set to { 0.1, 0.05, 0.02}, which mimic the values of the parameters when the original data are used.
18 Table: Average condition numbers for the various collinearity designs. Large bins, p = {0.1, 0.5, 0.9}. Collinearity among variables Cond {1,1,1,1} {10,1,1,0.1} {10,0,0.1,0.1} {10,.1,.05,.05} η η η η η η l
19 Monte Carlo Summary Statistics β 1 β 2 β 3 τ 1 τ 2 τ 3 True Mean Std Err t-ratio Averages of the Variance Proportion Table η l β 1 β 2 β 3 τ 1 τ 2 τ Standard Errors of the Variance Proportion Table Table: Orthogonal Regressors: Λ = {1, 1, 1, 1}, with Wide Bins: p={0.1, 0.5, 0.9}. Results based on 100 Monte Carlo samples.
20 Monte Carlo Summary Statistics β 1 β 2 β 3 τ 1 τ 2 τ 3 True Mean Std Err t-ratio Averages of the Variance Proportion Table η l β 1 β 2 β 3 τ 1 τ 2 τ Standard Errors of the Variance Proportion Table Table: Moderate Collinearity: Λ = {10, 1, 1, 0.1}, with Wide Bins: p={0.1, 0.5, 0.9}. Results based on 100 Monte Carlo samples.
21 Monte Carlo Summary Statistics β 1 β 2 β 3 τ 1 τ 2 τ 3 True Mean Std Err t-ratio Averages of the Variance Proportion Table η l β 1 β 2 β 3 τ 1 τ 2 τ Standard Errors of the Variance Proportion Table Table: Severe Collinearity: Λ = {10, 10, 0.1, 0.1}, with Wide Bins: p={0.1, 0.5, 0.9}. Results based on 100 Monte Carlo samples.
22 Monte Carlo Summary Statistics β 1 β 2 β 3 τ 1 τ 2 τ 3 True Mean Std Err t-ratio Averages of the Variance Proportion Table η l β 1 β 2 β 3 τ 1 τ 2 τ Standard Errors of the Variance Proportion Table Table: Severe Collinearity: Λ = {10, 0.1, 0.05, 0.05}, with Wide Bins: p={0.1, 0.5, 0.9}. Results based on 100 Monte Carlo samples.
23 General Thoughts on Estimation Bias low. Orthogonality of regressors does not imply no weakness Variance proportions measured fairly precisely.
24 Table: Various bin sizes for a mildly collinear data in. Λ = 10, 7, 4, 1. There is only 1 relatively large condition number in this scenario. Wide β 1 β 2 β 3 τ 1 τ 2 τ 3 Std Err η l φ l1 φ l2 φ l3 φ l4 φ l5 φ l Narrow β 1 β 2 β 3 τ 1 τ 2 τ 3 Std Err η l φ l1 φ l2 φ l3 φ l4 φ l5 φ l Really narrow β 1 β 2 β 3 τ 1 τ 2 τ 3 Std Err η l φ l1 φ l2 φ l3 φ l4 φ l5 φ l
25 1 Identification depends on collinear variables as well as functional form and parameters. 2 BKW diagnostics useful in diagnosing problems, both as collinearity worsens and as bins size shrinks. 3 Proportion of BKW diagnostics variability due to parameter estimation is minimal.
26 : Selection Equation z i = w T i γ + u i, i = 1,..., N (4) The continuous latent variable z is not directly observed. Instead we only know { 1 z z i = i > 0 0 otherwise
27 : Regression Equation y i = x T i β + e i, i = 1,..., n, N > n (5) The random disturbances of the two equations are normally distributed and correlated [ ] [( ) ( )] ui 0 1 ρ N, e i 0 ρ σe 2 Notice that the selection equation has more observations than the regression equation. (6)
28 Correlation between u i and e i alters the conditional mean of the the linear regression (5). E[y i z i > 0] = xi T β + ρσ e λ i where λ i = φ(w i γ)/φ(w i γ) is the inverse Mill s ratio, φ( ) and Φ( ) are standard normal pdf and cdf, respectively. Also, it is conventional to define β λ = ρσ e
29 The log-likelhihood (Greene, 2012, p. 878) is ln L = [ exp( (1/2)e 2 i /σ 2 ( e) ρei /σ e + w i Φ γ )] + [1 ln Φ(w z=1 σ e 2π 1 ρ 2 i γ)] z=0 where e i = y i x i β and i = 1, 2,..., n. Note: Fewer observations are available for the regression part of this likelihood. That affects identification in an important way.
30 Dependent variable: l wage, Selection variable: lfp Coefficient Std. Error z p-value const educ exper lambda Selection equation const educ exper kidsl Total observations: 753 Censored observations: 325 (43.2%)
31 Regression part of the BKW table --- variance proportions --- cond const educ exper lambda
32 BKW table, cont. Selection --- variance proportions --- cond const educ exper kidsl
33 Collinear variables Introduction X = {const, educ, exper, kidsl6}. is decomposed using the SVD. X = UDV and the diagonal matrix containing the eigenvalues, D, is replaced using the desired ones, X = UΛV. Since n < N the sample collinearity in regression and selection equations will be different, and not necessarily orthogonal in the design.
34 Four sets of eigenvalues are considered as replacements for D in the SVD. Eigenvalues Λ = {1, 1, 1, 1} Λ = {10, 1, 1, 0.1} Λ = {10, 10, 0.1, 0.1} Λ = {10, 0.1, 0.1, 0.05} Collinearity None Moderate High Severe
35 Censoring and Variance The selection degree is chosen from ρ = {0, 0.5, 0.99} and the variance from σ e = {1, 2} Doubling σ e doubles β λ as well as its standard error. It also doubles the measured standard errors of the regression equation, having no effect on the selection equation.
36 Choosing regressors: same or different subsets Two sets of regressors are chosen based on the collinear induced X obtained above. These are x i = 1, educ i, exper i and. w i = 1, educ i, exper i x i = 1, educ i, exper i w i = 1, educ i, kidsl6 i
37 Notes and issues Introduction 400 simulated samples per design. β estimated using the original data and rounded γ also estimated using the original data and rounded β λ = ρσ e n < N, which affects collinearity within each sample.
38 ρ β 1 β 2 β 3 β λ γ 1 γ 2 γ True Mean Std Err η l φ l1 φ l2 φ l3 φ l4 φ l5 φ l6 φ l ρ β 1 β 2 β 3 β λ γ 1 γ 2 γ True Mean Std Err η l φ l1 φ l2 φ l3 φ l4 φ l5 φ l6 φ l ρ β 1 β 2 β 3 β λ γ 1 γ 2 γ True Mean Std Err η l φ l1 φ l2 φ l3 φ l4 φ l5 φ l6 φ l ρ β 1 β 2 β 3 β λ γ 1 γ 2 γ True Mean Std Err η l φ l1 φ l2 φ l3 φ l4 φ l5 φ l6 φ l ρ β 1 β 2 β 3 β λ γ 1 γ 2 γ True Mean Std Err η l φ l1 φ l2 φ l3 φ l4 φ l5 φ l6 φ l
39 Variable choice. Orthogonal variables. Regression Selection λ cond cconst ceduc cexper β λ cconst ceduc cexper ckidsl λ cond cconst ceduc cexper β λ cconst ceduc cexper ckidsl λ cond cconst ceduc cexper β λ cconst ceduc cexper ckidsl
40 Variable choice. Orthogonal variables, con t. Regression Selection λ cond cconst ceduc cexper β λ cconst ceduc cexper ckidsl λ cond cconst ceduc cexper β λ cconst ceduc cexper ckidsl Note: cexper in the selection equation helps to identify cexper in the regression.
41 Monte Carlo Summary Statistics β 1 β 2 β 3 β λ γ 1 γ 2 γ 3 True Mean Std Err t-ratio Averages of the Variance Proportion Table η l β 1 β 2 β 3 β λ γ 1 γ 2 γ Standard Errors of the Variance Proportion Table Table: Orthogonal Regressors: Λ = {1, 1, 1, 1}, with moderate selectivity: ρ = 0.5. σ e = 1 and x i2 w i2. Failure proportion=0.
42 β 1 β 2 β 3 β λ γ 1 γ 2 γ 3 True Mean Std Err t-ratio Averages of the Variance Proportion Table η l β 1 β 2 β 3 β λ γ 1 γ 2 γ Standard Errors of the Variance Proportion Table Table: Severe Collinearity: Λ = {10, 0.1, 0.1, 0.05}, with moderate selectivity: ρ = 0.5. σ e = 1 and x i2 w i2. Failure proportion=0.02.
43 Monte Carlo Summary Statistics β 1 β 2 β 3 β λ γ 1 γ 2 γ 3 True Mean Std Err t-ratio Averages of the Variance Proportion Table η l β 1 β 2 β 3 β λ γ 1 γ 2 γ Standard Errors of the Variance Proportion Table Table: Orthogonal Regressors: Λ = {1, 1, 1, 1}, with high selectivity: ρ = σ e = 1 and x i2 w i2. Failure proportion=0.005.
44 Monte Carlo Summary Statistics β 1 β 2 β 3 β λ γ 1 γ 2 γ 3 True Mean Std Err t-ratio Averages of the Variance Proportion Table η l β 1 β 2 β 3 β λ γ 1 γ 2 γ Standard Errors of the Variance Proportion Table Table: Severe Collinearity: Λ = {10, 0.1, 0.1, 0.05}, with high selectivity: ρ = σ e = 1 and x i2 w i2. Failure proportion=0.008.
45 MLE badly biased for β λ when w = x. Bias smaller when w x. As sample endogeneity gets worse (ρ 1), the identification of β λ improves. Overall identification improves (a little). BKW diagnostics helpful in detecting issues with collinear variables and the identification of selectivity parameter.
46 Introduction The BKW diagnostics appear to work quite nicely and as expected in the nonlinear settings examined here. Not only can collinearity be detected, but issues with parameters and functional form can also be revealed. Thanks Allin for adding them to GRETL. :-)
47 Link to Paper Introduction For more information, results, and code see:
48 Greene, William H. (2012), Econometric Analysis, 7th edn, Prentice-Hall, Upper Saddle River, NJ.
AGEC 661 Note Fourteen
AGEC 661 Note Fourteen Ximing Wu 1 Selection bias 1.1 Heckman s two-step model Consider the model in Heckman (1979) Y i = X iβ + ε i, D i = I {Z iγ + η i > 0}. For a random sample from the population,
More informationOKLAHOMA STATE UNIVERSITY
2015 OKSWP1506 Economics Working Paper Series Department of Economics OKLAHOMA STATE UNIVERSITY http://spears.okstate.edu/ecls/ Collinearity Diagnostics in gretl Lee C. Adkins Oklahoma State University
More informationShort Questions (Do two out of three) 15 points each
Econometrics Short Questions Do two out of three) 5 points each ) Let y = Xβ + u and Z be a set of instruments for X When we estimate β with OLS we project y onto the space spanned by X along a path orthogonal
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationMulticollinearity and A Ridge Parameter Estimation Approach
Journal of Modern Applied Statistical Methods Volume 15 Issue Article 5 11-1-016 Multicollinearity and A Ridge Parameter Estimation Approach Ghadban Khalaf King Khalid University, albadran50@yahoo.com
More informationThe Multiple Regression Model Estimation
Lesson 5 The Multiple Regression Model Estimation Pilar González and Susan Orbe Dpt Applied Econometrics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 5 Regression model:
More informationApplied Health Economics (for B.Sc.)
Applied Health Economics (for B.Sc.) Helmut Farbmacher Department of Economics University of Mannheim Autumn Semester 2017 Outlook 1 Linear models (OLS, Omitted variables, 2SLS) 2 Limited and qualitative
More informationSTA Homework 1. Due back at the beginning of class on Oct 20, 2008
STA 410 - Homework 1 Due back at the beginning of class on Oct 20, 2008 The following is split into two parts. Part I should be solved by all. Part II is mandatory only for the graduate students in Statistics.
More informationTruncation and Censoring
Truncation and Censoring Laura Magazzini laura.magazzini@univr.it Laura Magazzini (@univr.it) Truncation and Censoring 1 / 35 Truncation and censoring Truncation: sample data are drawn from a subset of
More information5. Erroneous Selection of Exogenous Variables (Violation of Assumption #A1)
5. Erroneous Selection of Exogenous Variables (Violation of Assumption #A1) Assumption #A1: Our regression model does not lack of any further relevant exogenous variables beyond x 1i, x 2i,..., x Ki and
More informationBOOTSTRAP INFERENCES IN HETEROSCEDASTIC SAMPLE SELECTION MODELS: A MONTE CARLO INVESTIGATION
BOOTSTRAP INFERENCES IN HETEROSCEDASTIC SAMPLE SELECTION MODELS: A MONTE CARLO INVESTIGATION LEE C. ADKINS AND R. CARTER HILL Abstract. Several methods of estimating heteroscedastic sample selection models
More informationEconometrics Master in Business and Quantitative Methods
Econometrics Master in Business and Quantitative Methods Helena Veiga Universidad Carlos III de Madrid This chapter deals with truncation and censoring. Truncation occurs when the sample data are drawn
More informationEconomics 583: Econometric Theory I A Primer on Asymptotics
Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:
More informationCase of single exogenous (iv) variable (with single or multiple mediators) iv à med à dv. = β 0. iv i. med i + α 1
Mediation Analysis: OLS vs. SUR vs. ISUR vs. 3SLS vs. SEM Note by Hubert Gatignon July 7, 2013, updated November 15, 2013, April 11, 2014, May 21, 2016 and August 10, 2016 In Chap. 11 of Statistical Analysis
More informationEconometrics. 8) Instrumental variables
30C00200 Econometrics 8) Instrumental variables Timo Kuosmanen Professor, Ph.D. http://nomepre.net/index.php/timokuosmanen Today s topics Thery of IV regression Overidentification Two-stage least squates
More informationLecture 4: Multivariate Regression, Part 2
Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above
More informationMatrix Factorizations
1 Stat 540, Matrix Factorizations Matrix Factorizations LU Factorization Definition... Given a square k k matrix S, the LU factorization (or decomposition) represents S as the product of two triangular
More informationcomponent risk analysis
273: Urban Systems Modeling Lec. 3 component risk analysis instructor: Matteo Pozzi 273: Urban Systems Modeling Lec. 3 component reliability outline risk analysis for components uncertain demand and uncertain
More informationNinth ARTNeT Capacity Building Workshop for Trade Research "Trade Flows and Trade Policy Analysis"
Ninth ARTNeT Capacity Building Workshop for Trade Research "Trade Flows and Trade Policy Analysis" June 2013 Bangkok, Thailand Cosimo Beverelli and Rainer Lanz (World Trade Organization) 1 Selected econometric
More informationSample Selection. In a UW Econ Ph.D. Model. April 25, Kevin Kasberg and Alex Stevens ECON 5360 Dr. Aadland
Sample Selection In a UW Econ Ph.D. Model Kevin Kasberg and Alex Stevens ECON 5360 Dr. Aadland April 25, 2013 Sample Selection Bias Sample Selection Bias (Incidental Truncation): When estimates from a
More informationQuantitative Methods I: Regression diagnostics
Quantitative Methods I: Regression University College Dublin 10 December 2014 1 Assumptions and errors 2 3 4 Outline Assumptions and errors 1 Assumptions and errors 2 3 4 Assumptions: specification Linear
More informationProblem set - Selection and Diff-in-Diff
Problem set - Selection and Diff-in-Diff 1. You want to model the wage equation for women You consider estimating the model: ln wage = α + β 1 educ + β 2 exper + β 3 exper 2 + ɛ (1) Read the data into
More informationLecture 4: Multivariate Regression, Part 2
Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above
More informationApplied Economics. Regression with a Binary Dependent Variable. Department of Economics Universidad Carlos III de Madrid
Applied Economics Regression with a Binary Dependent Variable Department of Economics Universidad Carlos III de Madrid See Stock and Watson (chapter 11) 1 / 28 Binary Dependent Variables: What is Different?
More information8. Nonstandard standard error issues 8.1. The bias of robust standard errors
8.1. The bias of robust standard errors Bias Robust standard errors are now easily obtained using e.g. Stata option robust Robust standard errors are preferable to normal standard errors when residuals
More informationLeast Squares Estimation-Finite-Sample Properties
Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions
More informationAsymptotic Properties and simulation in gretl
Asymptotic Properties and simulation in gretl Quantitative Microeconomics R. Mora Department of Economics Universidad Carlos III de Madrid Outline 1 Asymptotic Results for OLS 2 3 4 5 Classical Assumptions
More informationThe Simple Regression Model. Simple Regression Model 1
The Simple Regression Model Simple Regression Model 1 Simple regression model: Objectives Given the model: - where y is earnings and x years of education - Or y is sales and x is spending in advertising
More informationOrdinary Least Squares Regression
Ordinary Least Squares Regression Goals for this unit More on notation and terminology OLS scalar versus matrix derivation Some Preliminaries In this class we will be learning to analyze Cross Section
More informationEconometric Analysis of Cross Section and Panel Data
Econometric Analysis of Cross Section and Panel Data Jeffrey M. Wooldridge / The MIT Press Cambridge, Massachusetts London, England Contents Preface Acknowledgments xvii xxiii I INTRODUCTION AND BACKGROUND
More informationIV estimators and forbidden regressions
Economics 8379 Spring 2016 Ben Williams IV estimators and forbidden regressions Preliminary results Consider the triangular model with first stage given by x i2 = γ 1X i1 + γ 2 Z i + ν i and second stage
More informationSection I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem -
First Exam: Economics 388, Econometrics Spring 006 in R. Butler s class YOUR NAME: Section I (30 points) Questions 1-10 (3 points each) Section II (40 points) Questions 11-15 (10 points each) Section III
More informationLecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)
Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16) 1 2 Model Consider a system of two regressions y 1 = β 1 y 2 + u 1 (1) y 2 = β 2 y 1 + u 2 (2) This is a simultaneous equation model
More informationA Short Note on Resolving Singularity Problems in Covariance Matrices
International Journal of Statistics and Probability; Vol. 1, No. 2; 2012 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education A Short Note on Resolving Singularity Problems
More informationStatistics II. Management Degree Management Statistics IIDegree. Statistics II. 2 nd Sem. 2013/2014. Management Degree. Simple Linear Regression
Model 1 2 Ordinary Least Squares 3 4 Non-linearities 5 of the coefficients and their to the model We saw that econometrics studies E (Y x). More generally, we shall study regression analysis. : The regression
More informationMediation Analysis: OLS vs. SUR vs. 3SLS Note by Hubert Gatignon July 7, 2013, updated November 15, 2013
Mediation Analysis: OLS vs. SUR vs. 3SLS Note by Hubert Gatignon July 7, 2013, updated November 15, 2013 In Chap. 11 of Statistical Analysis of Management Data (Gatignon, 2014), tests of mediation are
More informationL7: Multicollinearity
L7: Multicollinearity Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Introduction ï Example Whats wrong with it? Assume we have this data Y
More informationECON Introductory Econometrics. Lecture 11: Binary dependent variables
ECON4150 - Introductory Econometrics Lecture 11: Binary dependent variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 11 Lecture Outline 2 The linear probability model Nonlinear probability
More informationEconometrics Lecture 5: Limited Dependent Variable Models: Logit and Probit
Econometrics Lecture 5: Limited Dependent Variable Models: Logit and Probit R. G. Pierse 1 Introduction In lecture 5 of last semester s course, we looked at the reasons for including dichotomous variables
More informationRidge Regression Revisited
Ridge Regression Revisited Paul M.C. de Boer Christian M. Hafner Econometric Institute Report EI 2005-29 In general ridge (GR) regression p ridge parameters have to be determined, whereas simple ridge
More information8 - Continuous random vectors
8-1 Continuous random vectors S. Lall, Stanford 2011.01.25.01 8 - Continuous random vectors Mean-square deviation Mean-variance decomposition Gaussian random vectors The Gamma function The χ 2 distribution
More informationECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors
ECON4150 - Introductory Econometrics Lecture 6: OLS with Multiple Regressors Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 6 Lecture outline 2 Violation of first Least Squares assumption
More informationLINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception
LINEAR MODELS FOR CLASSIFICATION Classification: Problem Statement 2 In regression, we are modeling the relationship between a continuous input variable x and a continuous target variable t. In classification,
More informationPh.D. Qualifying Exam Friday Saturday, January 6 7, 2017
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a
More informationLecture 11/12. Roy Model, MTE, Structural Estimation
Lecture 11/12. Roy Model, MTE, Structural Estimation Economics 2123 George Washington University Instructor: Prof. Ben Williams Roy model The Roy model is a model of comparative advantage: Potential earnings
More informationLogistic & Tobit Regression
Logistic & Tobit Regression Different Types of Regression Binary Regression (D) Logistic transformation + e P( y x) = 1 + e! " x! + " x " P( y x) % ln$ ' = ( + ) x # 1! P( y x) & logit of P(y x){ P(y
More informationMaximum Likelihood and. Limited Dependent Variable Models
Maximum Likelihood and Limited Dependent Variable Models Michele Pellizzari IGIER-Bocconi, IZA and frdb May 24, 2010 These notes are largely based on the textbook by Jeffrey M. Wooldridge. 2002. Econometric
More informationESTIMATING AVERAGE TREATMENT EFFECTS: REGRESSION DISCONTINUITY DESIGNS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics
ESTIMATING AVERAGE TREATMENT EFFECTS: REGRESSION DISCONTINUITY DESIGNS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics July 2009 1. Introduction 2. The Sharp RD Design 3.
More informationSemiparametric Estimation with Mismeasured Dependent Variables: An Application to Duration Models for Unemployment Spells
Semiparametric Estimation with Mismeasured Dependent Variables: An Application to Duration Models for Unemployment Spells Jason Abrevaya University of Chicago Graduate School of Business Chicago, IL 60637,
More information1 Outline. 1. Motivation. 2. SUR model. 3. Simultaneous equations. 4. Estimation
1 Outline. 1. Motivation 2. SUR model 3. Simultaneous equations 4. Estimation 2 Motivation. In this chapter, we will study simultaneous systems of econometric equations. Systems of simultaneous equations
More informationUniversity of California at Berkeley Fall Introductory Applied Econometrics Final examination. Scores add up to 125 points
EEP 118 / IAS 118 Elisabeth Sadoulet and Kelly Jones University of California at Berkeley Fall 2008 Introductory Applied Econometrics Final examination Scores add up to 125 points Your name: SID: 1 1.
More informationBinary choice 3.3 Maximum likelihood estimation
Binary choice 3.3 Maximum likelihood estimation Michel Bierlaire Output of the estimation We explain here the various outputs from the maximum likelihood estimation procedure. Solution of the maximum likelihood
More informationChapter 11. Regression with a Binary Dependent Variable
Chapter 11 Regression with a Binary Dependent Variable 2 Regression with a Binary Dependent Variable (SW Chapter 11) So far the dependent variable (Y) has been continuous: district-wide average test score
More informationHomoskedasticity. Var (u X) = σ 2. (23)
Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This
More informationChapter 6: Endogeneity and Instrumental Variables (IV) estimator
Chapter 6: Endogeneity and Instrumental Variables (IV) estimator Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans December 15, 2013 Christophe Hurlin (University of Orléans)
More informationEssential of Simple regression
Essential of Simple regression We use simple regression when we are interested in the relationship between two variables (e.g., x is class size, and y is student s GPA). For simplicity we assume the relationship
More informationECON 3150/4150, Spring term Lecture 6
ECON 3150/4150, Spring term 2013. Lecture 6 Review of theoretical statistics for econometric modelling (II) Ragnar Nymoen University of Oslo 31 January 2013 1 / 25 References to Lecture 3 and 6 Lecture
More informationAn explanation of Two Stage Least Squares
Introduction Introduction to Econometrics An explanation of Two Stage Least Squares When we get an endogenous variable we know that OLS estimator will be inconsistent. In addition OLS regressors will also
More informationEconometrics of Panel Data
Econometrics of Panel Data Jakub Mućk Meeting # 2 Jakub Mućk Econometrics of Panel Data Meeting # 2 1 / 26 Outline 1 Fixed effects model The Least Squares Dummy Variable Estimator The Fixed Effect (Within
More informationPRINCIPAL COMPONENTS ANALYSIS
121 CHAPTER 11 PRINCIPAL COMPONENTS ANALYSIS We now have the tools necessary to discuss one of the most important concepts in mathematical statistics: Principal Components Analysis (PCA). PCA involves
More informationEcon 836 Final Exam. 2 w N 2 u N 2. 2 v N
1) [4 points] Let Econ 836 Final Exam Y Xβ+ ε, X w+ u, w N w~ N(, σi ), u N u~ N(, σi ), ε N ε~ Nu ( γσ, I ), where X is a just one column. Let denote the OLS estimator, and define residuals e as e Y X.
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1
MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical
More informationRegression #8: Loose Ends
Regression #8: Loose Ends Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #8 1 / 30 In this lecture we investigate a variety of topics that you are probably familiar with, but need to touch
More informationLecture 12: Application of Maximum Likelihood Estimation:Truncation, Censoring, and Corner Solutions
Econ 513, USC, Department of Economics Lecture 12: Application of Maximum Likelihood Estimation:Truncation, Censoring, and Corner Solutions I Introduction Here we look at a set of complications with the
More informationChapter 2: simple regression model
Chapter 2: simple regression model Goal: understand how to estimate and more importantly interpret the simple regression Reading: chapter 2 of the textbook Advice: this chapter is foundation of econometrics.
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions
More informationInstrumental Variables, Simultaneous and Systems of Equations
Chapter 6 Instrumental Variables, Simultaneous and Systems of Equations 61 Instrumental variables In the linear regression model y i = x iβ + ε i (61) we have been assuming that bf x i and ε i are uncorrelated
More informationApplied Statistics and Econometrics
Applied Statistics and Econometrics Lecture 6 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 53 Outline of Lecture 6 1 Omitted variable bias (SW 6.1) 2 Multiple
More information4 Instrumental Variables Single endogenous variable One continuous instrument. 2
Econ 495 - Econometric Review 1 Contents 4 Instrumental Variables 2 4.1 Single endogenous variable One continuous instrument. 2 4.2 Single endogenous variable more than one continuous instrument..........................
More informationECON3150/4150 Spring 2016
ECON3150/4150 Spring 2016 Lecture 4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo Last updated: January 26, 2016 1 / 49 Overview These lecture slides covers: The linear regression
More informationECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47
ECON2228 Notes 2 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 2 2014 2015 1 / 47 Chapter 2: The simple regression model Most of this course will be concerned with
More informationA Primer on Asymptotics
A Primer on Asymptotics Eric Zivot Department of Economics University of Washington September 30, 2003 Revised: October 7, 2009 Introduction The two main concepts in asymptotic theory covered in these
More informationSimulating Uniform- and Triangular- Based Double Power Method Distributions
Journal of Statistical and Econometric Methods, vol.6, no.1, 2017, 1-44 ISSN: 1792-6602 (print), 1792-6939 (online) Scienpress Ltd, 2017 Simulating Uniform- and Triangular- Based Double Power Method Distributions
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationRecent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data
Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More informationRegression. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh.
Regression Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh September 24 (All of the slides in this course have been adapted from previous versions
More informationMultivariate Regression: Part I
Topic 1 Multivariate Regression: Part I ARE/ECN 240 A Graduate Econometrics Professor: Òscar Jordà Outline of this topic Statement of the objective: we want to explain the behavior of one variable as a
More informationLecture 8: Instrumental Variables Estimation
Lecture Notes on Advanced Econometrics Lecture 8: Instrumental Variables Estimation Endogenous Variables Consider a population model: y α y + β + β x + β x +... + β x + u i i i i k ik i Takashi Yamano
More informationMultiple Linear Regression CIVL 7012/8012
Multiple Linear Regression CIVL 7012/8012 2 Multiple Regression Analysis (MLR) Allows us to explicitly control for many factors those simultaneously affect the dependent variable This is important for
More informationEconomics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects
Economics 113 Simple Regression Models Simple Regression Assumptions Simple Regression Derivation Changing Units of Measurement Nonlinear effects OLS and unbiased estimates Variance of the OLS estimates
More informationLeast Squares Estimation of a Panel Data Model with Multifactor Error Structure and Endogenous Covariates
Least Squares Estimation of a Panel Data Model with Multifactor Error Structure and Endogenous Covariates Matthew Harding and Carlos Lamarche January 12, 2011 Abstract We propose a method for estimating
More information1. The OLS Estimator. 1.1 Population model and notation
1. The OLS Estimator OLS stands for Ordinary Least Squares. There are 6 assumptions ordinarily made, and the method of fitting a line through data is by least-squares. OLS is a common estimation methodology
More informationChapter 4: Factor Analysis
Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.
More informationECON Introductory Econometrics. Lecture 16: Instrumental variables
ECON4150 - Introductory Econometrics Lecture 16: Instrumental variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 12 Lecture outline 2 OLS assumptions and when they are violated Instrumental
More information[POLS 8500] Review of Linear Algebra, Probability and Information Theory
[POLS 8500] Review of Linear Algebra, Probability and Information Theory Professor Jason Anastasopoulos ljanastas@uga.edu January 12, 2017 For today... Basic linear algebra. Basic probability. Programming
More informationTHE MULTIVARIATE LINEAR REGRESSION MODEL
THE MULTIVARIATE LINEAR REGRESSION MODEL Why multiple regression analysis? Model with more than 1 independent variable: y 0 1x1 2x2 u It allows : -Controlling for other factors, and get a ceteris paribus
More information4 Instrumental Variables Single endogenous variable One continuous instrument. 2
Econ 495 - Econometric Review 1 Contents 4 Instrumental Variables 2 4.1 Single endogenous variable One continuous instrument. 2 4.2 Single endogenous variable more than one continuous instrument..........................
More informationSTK-IN4300 Statistical Learning Methods in Data Science
Outline of the lecture STK-I4300 Statistical Learning Methods in Data Science Riccardo De Bin debin@math.uio.no Model Assessment and Selection Cross-Validation Bootstrap Methods Methods using Derived Input
More informationLinear Regression. Junhui Qian. October 27, 2014
Linear Regression Junhui Qian October 27, 2014 Outline The Model Estimation Ordinary Least Square Method of Moments Maximum Likelihood Estimation Properties of OLS Estimator Unbiasedness Consistency Efficiency
More informationMultiple Regression Analysis. Part III. Multiple Regression Analysis
Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationDiscrete Choice Modeling
[Part 4] 1/43 Discrete Choice Modeling 0 Introduction 1 Summary 2 Binary Choice 3 Panel Data 4 Bivariate Probit 5 Ordered Choice 6 Count Data 7 Multinomial Choice 8 Nested Logit 9 Heterogeneity 10 Latent
More informationECON 4551 Econometrics II Memorial University of Newfoundland. Panel Data Models. Adapted from Vera Tabakova s notes
ECON 4551 Econometrics II Memorial University of Newfoundland Panel Data Models Adapted from Vera Tabakova s notes 15.1 Grunfeld s Investment Data 15.2 Sets of Regression Equations 15.3 Seemingly Unrelated
More informationNonresponse weighting adjustment using estimated response probability
Nonresponse weighting adjustment using estimated response probability Jae-kwang Kim Yonsei University, Seoul, Korea December 26, 2006 Introduction Nonresponse Unit nonresponse Item nonresponse Basic strategy
More informationMA Advanced Econometrics: Applying Least Squares to Time Series
MA Advanced Econometrics: Applying Least Squares to Time Series Karl Whelan School of Economics, UCD February 15, 2011 Karl Whelan (UCD) Time Series February 15, 2011 1 / 24 Part I Time Series: Standard
More informationEconomics 326 Methods of Empirical Research in Economics. Lecture 14: Hypothesis testing in the multiple regression model, Part 2
Economics 326 Methods of Empirical Research in Economics Lecture 14: Hypothesis testing in the multiple regression model, Part 2 Vadim Marmer University of British Columbia May 5, 2010 Multiple restrictions
More informationIntroduction to Econometrics Final Examination Fall 2006 Answer Sheet
Introduction to Econometrics Final Examination Fall 2006 Answer Sheet Please answer all of the questions and show your work. If you think a question is ambiguous, clearly state how you interpret it before
More informationA Robust Approach to Estimating Production Functions: Replication of the ACF procedure
A Robust Approach to Estimating Production Functions: Replication of the ACF procedure Kyoo il Kim Michigan State University Yao Luo University of Toronto Yingjun Su IESR, Jinan University August 2018
More informationHeteroskedasticity. Part VII. Heteroskedasticity
Part VII Heteroskedasticity As of Oct 15, 2015 1 Heteroskedasticity Consequences Heteroskedasticity-robust inference Testing for Heteroskedasticity Weighted Least Squares (WLS) Feasible generalized Least
More information