Estimating Semi-parametric Panel Multinomial Choice Models

Similar documents
Estimating Semi-parametric Panel Multinomial Choice Models using Cyclic Monotonicity

Estimating Semi-parametric Panel Multinomial Choice Models using Cyclic Monotonicity

Informational Content in Static and Dynamic Discrete Response Panel Data Models

Exclusion Restrictions in Dynamic Binary Choice Panel Data Models

Identification in Discrete Choice Models with Fixed Effects

Partial Identification and Inference in Binary Choice and Duration Panel Data Models

IDENTIFICATION OF THE BINARY CHOICE MODEL WITH MISCLASSIFICATION

Panel Data Seminar. Discrete Response Models. Crest-Insee. 11 April 2008

Non-linear panel data modeling

Semiparametric Identification in Panel Data Discrete Response Models

An Overview of the Special Regressor Method

Working Paper No Maximum score type estimators

Quantile Regression for Panel Data Models with Fixed Effects and Small T : Identification and Estimation

Identification and Estimation of Marginal Effects in Nonlinear Panel Models 1

Nonseparable multinomial choice models in cross-section and panel data

Binary Choice Models with Discrete Regressors: Identification and Misspecification

Lecture 32: Asymptotic confidence sets and likelihoods

What s New in Econometrics? Lecture 14 Quantile Methods

Syllabus. By Joan Llull. Microeconometrics. IDEA PhD Program. Fall Chapter 1: Introduction and a Brief Review of Relevant Tools

Identification in Nonparametric Limited Dependent Variable Models with Simultaneity and Unobserved Heterogeneity

Nonparametric panel data models with interactive fixed effects

Random Coefficients on Endogenous Variables in Simultaneous Equations Models

UNIVERSITY OF CALIFORNIA Spring Economics 241A Econometrics

Rank Estimation of Partially Linear Index Models

Partial Rank Estimation of Transformation Models with General forms of Censoring

Unobserved Preference Heterogeneity in Demand Using Generalized Random Coefficients

Econometric Analysis of Games 1

Identification of Panel Data Models with Endogenous Censoring

Active Learning: Disagreement Coefficient

Econ 273B Advanced Econometrics Spring

Nonparametric Identification and Estimation of a Transformation Model

Discrete Time Duration Models with Group level Heterogeneity

Comments on: Panel Data Analysis Advantages and Challenges. Manuel Arellano CEMFI, Madrid November 2006

A characterization of combinatorial demand

Unobserved Preference Heterogeneity in Demand Using Generalized Random Coefficients

Econometrics of Panel Data

Identification and Estimation of Nonlinear Dynamic Panel Data. Models with Unobserved Covariates

Data Models. Dalia A. Ghanem. May 8, Abstract. Recent work on nonparametric identification of average partial effects (APEs) from panel

A Semiparametric Network Formation Model with Multiple Linear Fixed Effects

CALIFORNIA INSTITUTE OF TECHNOLOGY

Random Projection Estimation of Discrete-Choice Models with Large Choice Sets

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity

FLEXIBLE CORRELATED RANDOM EFFECTS ESTIMATION IN PANEL MODELS WITH UNOBSERVED HETEROGENEITY

The Identification Zoo - Meanings of Identification in Econometrics: PART 3

Chapter 6. Hypothesis Tests Lecture 20: UMP tests and Neyman-Pearson lemma

Estimation of Structural Parameters and Marginal Effects in Binary Choice Panel Data Models with Fixed Effects

Random Projection Estimation of Discrete-Choice Models with Large Choice Sets

Identification and Estimation of Nonlinear Dynamic Panel Data. Models with Unobserved Covariates

Random Projection Estimation of Discrete-Choice Models with Large Choice Sets

A Discontinuity Test for Identification in Nonparametric Models with Endogeneity

Identifying Effects of Multivalued Treatments

Lecture 6: Discrete Choice: Qualitative Response

CRE METHODS FOR UNBALANCED PANELS Correlated Random Effects Panel Data Models IZA Summer School in Labor Economics May 13-19, 2013 Jeffrey M.

Random coefficients on endogenous variables in simultaneous equations models

Simple Estimators for Monotone Index Models

An instrumental variable random-coefficients model for binary outcomes

A dynamic model for binary panel data with unobserved heterogeneity admitting a n-consistent conditional estimator

ESTIMATION OF NONPARAMETRIC MODELS WITH SIMULTANEITY

Social Interactions under Incomplete Information with Heterogeneous Expectations

Heterogeneity. Krishna Pendakur. May 24, Krishna Pendakur () Heterogeneity May 24, / 21

Missing dependent variables in panel data models

Chapter 4: Asymptotic Properties of the MLE

17/003. Alternative moment conditions and an efficient GMM estimator for dynamic panel data models. January, 2017

New Developments in Econometrics Lecture 16: Quantile Estimation

A Note on Binary Choice Duration Models

1 Estimation of Persistent Dynamic Panel Data. Motivation

L p Functions. Given a measure space (X, µ) and a real number p [1, ), recall that the L p -norm of a measurable function f : X R is defined by

Econometric Analysis of Cross Section and Panel Data

Nonparametric identification and estimation of random coefficients in multinomial choice models

Kernel Logistic Regression and the Import Vector Machine

Parametric identification of multiplicative exponential heteroskedasticity ALYSSA CARLSON

Semiparametric Identification and Fisher Information

Identification of Consumer Switching Behavior with Market Level Data

Nonparametric identification of endogenous and heterogeneous aggregate demand models: complements, bundles and the market level

IDENTIFICATION IN DIFFERENTIATED PRODUCTS MARKETS USING MARKET LEVEL DATA. Steven T. Berry and Philip A. Haile. January 2010 Updated February 2010

Lecture 21: Convergence of transformations and generating a random variable

On the Empirical Content of the Beckerian Marriage Model

Chapter 1 - Preference and choice

Simple Estimators for Semiparametric Multinomial Choice Models

Statistical Properties of Numerical Derivatives

Lecture 8 Inequality Testing and Moment Inequality Models

An instrumental variable model of multiple discrete choice

Identification and Estimation of Partially Linear Censored Regression Models with Unknown Heteroscedasticity

Limited Dependent Variables and Panel Data

16/018. Efficiency Gains in Rank-ordered Multinomial Logit Models. June 13, 2016

LIMITED DEPENDENT VARIABLE CORRELATED RANDOM COEFFICIENT PANEL DATA MODELS. A Dissertation ZHONGWEN LIANG

Estimation of the Conditional Variance in Paired Experiments

Flexible Estimation of Treatment Effect Parameters

Optimality Conditions for Nonsmooth Convex Optimization

IDENTIFICATION IN NONPARAMETRIC SIMULTANEOUS EQUATIONS

Nonlinear Panel Data Analysis

Quantile Regression for Dynamic Panel Data

An instrumental variable random coefficients model for binary outcomes

NBER WORKING PAPER SERIES NONPARAMETRIC IDENTIFICATION OF MULTINOMIAL CHOICE DEMAND MODELS WITH HETEROGENEOUS CONSUMERS

Ninth ARTNeT Capacity Building Workshop for Trade Research "Trade Flows and Trade Policy Analysis"

Semiparametric Estimation of Invertible Models

Random Coefficients on Endogenous Variables in Simultaneous Equations Models

A Course in Applied Econometrics Lecture 18: Missing Data. Jeff Wooldridge IRP Lectures, UW Madison, August Linear model with IVs: y i x i u i,

Random Projection Estimation of Discrete-Choice Models with Large Choice Sets

Stat 5101 Lecture Notes

Transcription:

Estimating Semi-parametric Panel Multinomial Choice Models Xiaoxia Shi, Matthew Shum, Wei Song UW-Madison, Caltech, UW-Madison September 15, 2016 1 / 31

Introduction We consider the panel multinomial choice model: Y k it = 1{β X k it + A k i + u k it β X k it + A k i + ɛ k it k = 0,..., K}. A k i : individual/choice-specific fixed effect with arbitrary dependence on {Xit k}t t=1. ɛ k it : idiosyncratic error with arbitrary dependence and heterogeneity across k. Short panel: T = 2. Normalization: X 0 it = 0, A0 i = 0; β = 1; no constant in X k it. Examples: Transportation mode choices; demand for differentiated products. 2 / 31

Contributions Derive identifying moment inequalities where the fixed effects are eliminated. Give easy-to-verify point identification conditions for β. Propose a consistent estimator. 3 / 31

Contributions Derive identifying moment inequalities where the fixed effects are eliminated. Give easy-to-verify point identification conditions for β. Propose a consistent estimator. Not yet done: the asymptotic distribution of the estimator. 3 / 31

Literature Panel data multinomial choice. Chamberlain (1989, logit), Arellano and Bonhomme (2009, long panel), Pakes and Porter (2015), Khan, Tamer, and Ouyang (2016) Cross-sectional data multinomial choice. McFadden (1973, 1978, 1989, logit), Manski (1975, iid error), Matzkin (1993, 2012), Lewbel (2000), Fox (2007, exchangeable error), Yan (2012), Powell and Ruud (2008), Ichimura, Powell, and Ruud (2015, rank correlation) Panel data binary choice. Chamberlain (1984,2001, logit), Manski (1987, maximum score), Honoré and Kyriazidou (2000), Honoré and Lewbel (2002, special regressor), Abrevaya (2000), Hoderlein and White (2012). Cross-sectional data binary choice. Logit, Probit, maximum score, rank correlation (Han, 1987), maximum likelihood, least square, etc. Cyclic monotonicity. Rochet (1987, implementability), Browning (1989, SREH), Ashlagi, Braverman, Hassidim, Monderer (2010). 4 / 31

Roadmap Cyclic monotonicity and our model Identification Estimation and Monte Carlo Conclusion 5 / 31

Cyclic Monotonicity Cyclic monotonicity is a generalization of monotonicity from R R to R K R K. Define a cycle of length M in B R k as: x 1, x 2,..., x M, x M+1 = x 1. f : B R K is cyclic monotone wrt the above cycle iff M (x m x m+1 ) f (x m ) 0. m=1 Special case: if M = 2: (x 2 x 1 ) (f (x 2 ) f (x 1 )) 0. Let F : B R be a continuously differentiable function. Then F is convex if and only if its gradient F : B R K is cyclic monotone. ref. Rockafellar (1970, Ch. 24). 6 / 31

in a Multinomial Choice Model Consider the model Y k = 1{β X k + u k β X k + u k for all k = 0,..., K} Normalization: X 0 = 0. Let u = (u 0,..., u K ), X = (X 1,..., X K ), and Y = (Y 1,..., Y K ). Define the social surplus function: { } G(β X) = E max k=0,...,k [β X k + u k ] β X. 7 / 31

in a Multinomial Choice Model { } G(β X) = E max k=0,...,k [β X k + u k ] β X. Proposition Suppose that u is independent of X, and the cdf of u is continuous. Then (a) G( ) is convex and differentiable, (b) G( ) = p( ), where p( ) E(Y β X = ), and (c) p( ) is cyclic monotone. Proof is simple and parallels McFadden (1981). 8 / 31

Panel Multionoial Choice Model Consider the panel model Y k it = 1{β X k it + A k i + ɛ k it β X k it + A k i + ɛ k it Define the conditional choice probability p(v, a) = E[Y it X itβ = v, A i = a]. for all k = 0,..., K}. Assumption 3.1 (a) ɛ it (X i1, X i2 ) A i, (b) ɛ i1 ɛ i2 A i, and (c) the conditional cdf of ɛ it given A i is continuous. Lemma The function p(, a) is cyclic monotone for every a under Assumption 3.1. 9 / 31

From CM to Identification p(v, a) = E[Y it X it β = v, A i = a] is cyclic monotone in v. Thus, for any x 1, x 2 supp(x it ), and any a supp(a i ), (p(x 1β, a) p(x 2β, a)) (x 1β x 2β) 0. 10 / 31

From CM to Identification p(v, a) = E[Y it X it β = v, A i = a] is cyclic monotone in v. Thus, for any x 1, x 2 supp(x it ), and any a supp(a i ), (p(x 1β, a) p(x 2β, a)) (x 1β x 2β) 0. With panel data: p(x β, a) = E[Y it X it = x, X i = a] t. Thus, ( E[Y i1 X i1 = x 1, A i = a ] E[Y i2 X i2 = x 2, A i = a ] ) (x 1β x 2β) 0, for any x 1, x 2 supp(x it ), and any a supp(a i ), 10 / 31

From CM to Identification Then, ( E[Y i1 X i1, X i2, A i ] E[Y i2 X i1, X i2, A i ] ) (X i1β X i2β) 0 Take conditional expectation given X i1,..., X i2, and we have where Z i = Z i2 Z i1. E[ Y i X i1, X i2 ] X iβ 0 This reduces to the rank-based identification conditions in Manski (1987) and Abrevaya (2000) when K = 1 (binary choice). 11 / 31

Point Identification of β Our identification inequality is E[ Y i X i1, X i2 ] X iβ 0 pointwise. G supp( X i E[ Y i X i1, X i2 ]) is important for identification. It is important up to its convex conic hull: cc(g) = {λ 1 g 1 + λ 2 g 2 R dx : g 1, g 2 G, λ 1, λ 2 0}. The identified set B 0 of β is B 0 = {b R dx : b = 1, b g 0 g G} = {b R dx : b = 1, b g 0 g cc(g)}. 12 / 31

cc(g) and the Identification of β A A A D E O C O C O C F G B B B Panel (i) Panel (ii) Panel (iii) Figure : Point identification of β: geometric intuition OC, OECF, OACB stand for cc(g) in panels (i), (ii), (iii), respectively. OACB, ODCG, OC stand for ID set of β in panels (i), (ii), (iii), rspctvly. 13 / 31

Point Identification goal Recall that G supp( X i E[ Y i X i1, X i2 ]). We would like to find conditions on model priminitives (X it, A it and ɛ it ) that guarantee that G is rich enough to ensure point ID, i.e. to guarantee the following high-level condition. Assumption 3.2 The set cc(g) is a halfspace of R K. 14 / 31

Point ID Conditions on Unobservables First, we impose a condition on the unobservables: Assumption 4.1. (a) Pr(supp(ɛ it A i ) = R K ) > 0. (b) The conditional CDF of A i given (X i1, X i2 ) = (x 1, x 2 ) is continuous in (x 1, x 2 ). Last, we need a condition on the observable X i. 15 / 31

Point ID Intuition from the Binary Case To get intuition, consider the binary case (K = 1): E[ Y i X i1, X i2 ] is a scalar. Then cc(g) = cc(supp( X i E[ Y i X i1, X i2 ])) This form makes the task easier. = cc (supp( X i sign(e[ Y i X i1, X i2 ]))) = cc ( supp( X i sign( X iβ)) ) 16 / 31

A Simple Sufficient Condition in the Binary Case For example, a sufficient condition when d x = 2: supp( X i ) supp( X i ) occupies all polar angles. And, for general d x, a sufficent condition is: cone (supp( X i ) supp( X i )) = R dx, where cone(g) = {λg : λ 0, g G}. 17 / 31

Restoring the Simple Form when K 2 X i E[ Y i X i1, X i2 ] = K k=1 X k i E[ Y k i X i1, X i2 ] Two ways to reduce the summation to the simple form: Conditional on the event Xi k = Xi 1 for all k. Then X i E[ Y i X i1, X i2 ] = X 1 i E[ Y 0 i X i1, X i2 ]. For a given k, conditional on the event X k i = 0. Then X i E[ Y i X i1, X i2 ] = X k i E[ Y k i X i1, X i2 ]. 18 / 31

Let G I be the union (over k) of supp( Xi k X k i = 0) supp( Xi k X k i = 0), where X k i = ( X 1 i,..., X k 1 Let G II be i, X k+1 i,..., X K i ). supp( Xi 1 Xi k = Xi 1 k) supp( Xi 1 Xi k = Xi 1 k) Let G = G I G II. 19 / 31

Point ID Our identification conditions will be imposed on the set G. Two sets of such conditions are given. Each is sufficient by itself. The first set allows all covariates to be bounded, but essentially requires all covariates to be continuous. The second set allows discrete regressors most generally, but requires one regressor with large support. 20 / 31

Point Identification Bounded Covariates The first sufficient uniform point identification allows (but does not require) all covariates to be bounded. Assumption 4.2 The set cone(g) = R dx. Theorem 4.1 Under Assumptions 3.1 and 4.1-2, Assumption 3.2 holds, and thus (a) B 0 = {β}, (b) Q(β) = 0, Q(b) > 0 for all b {b R dx : b = 1} and b β, where Q(b) = E min(0, E[ Y i X i1, X i2 ] X ib). Verifying Assumption 4.2 is fairly easy. Let s see a few examples. 21 / 31

Verifying Assumption 4.2 Examples For simplicity, let K = 2 (3-choices), d x = 2. Example 1. supp((x k it ) t=1,2;k=1,2) = [0, 1] 8. Then Then, given X 1 i = 0, supp(( X k i ) k=1,2 ) = [ 1, 1] 4. supp( X 2 i ) = [ 1, 1] 2. The cone generated by [ 1, 1] 2 is R 2, and [ 1, 1] 2 G I G. Thus, Assumption 4.2 is satisfied. 22 / 31

Verifying Assumption 4.2 Examples For simplicity, let K = 2 (3-choices), d x = 2. Example 2. Suppose that the covariates do not vary across k: X k it = X it for k = 1, 2, and supp((x it ) t=1,2 ) = [0, 1] 4. Thes, G II = supp( X i ) = [ 1, 1] 2. Assumption 4.2 is verified by noting that cone([ 1, 1] 2 ) = R 2, and that G II G. 23 / 31

Verifying Assumption 4.2 Examples For simplicity, let K = 2 (3-choices), d x = 2. Example 3. supp((xit 1) t=1,2) = [0, 1] 4, supp((xit 2) t=1,2) = {0, 1} 4, and the joint support is the Cartesian product. Then, supp( X 1 i X 2 i = 0) = [ 1, 1] 2. Note that cone([ 1, 1] 2 ) = R 2, and [ 1, 1] 2 G I G. Thus, Assumption 4.2 holds. 24 / 31

Point Identification Discrete Covariates Assumption 4.2 allows all covariates to be bounded, but only allows discrete covariates to a limited extent. Allowing for discrete covariates more generally requires the presence of an unbounded covariate. (Chamberlain (2001), our Theorem B.1). Those are impossibility results of uniform ID. The CM inequalities may still point ID β, depending on the true value of it. We next present an alternative to Assumption 4.2 that allows discrete covariates generally. 25 / 31

Point Identification Discrete Covariates Let g j = (g 1,..., g j 1, g j+1,..., g dx ). Let G j be the projection of G on the space of g j. Let G j (g j ) = {g j R : (g 1,..., g dx ) G}. Assumption 4.3 There exists a j {1,..., d x } such that (a) G j (g j ) = R for all g j in a subset G 0 j of G j, (b) G 0 j is symmetric about the origin, and is not contained in a proper linear subspace of R dx 1, and (c) the jth element of β, denoted by β j, is nonzero. Theorem 4.2 Under Assumptions 3.1, 4.1, and 4.3, Assumption 3.2 holds, and thus (a) B 0 = {β}, (b) Q(β) = 0, Q(b) > 0 for all b {b R dx : b = 1} and b β. 26 / 31

Verifying Assumption 4.3 An Example Let K = 2 (3-choices), d x = 2, and suppose that β 2 0. Example 1. Suppose that the first covariate is a time dummy: X1,it k = t for all k, t, and the second covariate has unbounded support: supp(x2,it k ) t=1,2;k=1,2) = (c, ) 4 for some c R. 27 / 31

Verifying Assumption 4.3 An Example Let K = 2 (3-choices), d x = 2, and suppose that β 2 0. Example 1. Suppose that the first covariate is a time dummy: X1,it k = t for all k, t, and the second covariate has unbounded support: supp(x2,it k ) t=1,2;k=1,2) = (c, ) 4 for some c R. Then supp ( ( X1,i, 1 X1,i, 2 X2,i, 1 X2,i) 2 ) = {1} 2 R 2. Then, supp ( Xi 1 Xi 1 = Xi 2 ) = {1} R. Then, G G II = { 1, 1} R. Let the special j in Assumption 4.3 be 2, and let G 2 0 = { 1, 1}. Assumption 4.3(b) obviously holds. Assumption 4.3(a) also holds because G 1 ( 1) = G 1 (1) = R. Assumption 4.3(c) holds since β 2 0. 27 / 31

A Consistent Estimator Let ˆp(X i1, X i2 ) be an estimator of p(x i1, X i2 ) E[ Y i X i1, X i2 ]. Let β be the argmin of the following criterion function, subject to normalization, Q n (b) = n [ b X i ˆp(X i1, X i2 ) ]. i=1 Assumption 5.1 (a) sup x1,x 2 X ˆp(x 1, x 2 ) p(x 1, x 2 ) p 0 as n, (b) E[ X it ] <. Theorem 5.1 Under Assumptions 3.1, 3.2, and 5.1, β p β as n. 28 / 31

Monte Carlo Exercise K = 2 (Trinary choice), T = 3. Let Xit k R3. Let X k across k, j, and t. j,it be Uniform[0, 1] and let it be independent Let A k i = (ωi k + 3 j=1 X j,i1 k )/4 where ωk i across k and of other model primitives. is Uniform[0, 1], independent Let ɛ k it = Ak i (uk it u0 it ) for k = 1, 2, where u 0 it 0 1 0 0 uit 1 N 0, 0 1 0.5 uit 2 0 0 0.5 1 Let the u s be independent across t. Let β = (1, 0.5, 0). 29 / 31

Monte Carlo Results Table : Monte Carlo Results (5000 Repetitions), β 1 normalized to 1 n β 2 β3 BIAS SD rmse BIAS SD rmse Length-2 cycles only 250 -.1052.1077.1506 -.0096.0996.1000 500 -.0800.0781.1118 -.0110.0738.0746 1000 -.0624.0580.0852 -.0119.0545.0558 2000 -.0509.0418.0659 -.0107.0387.0402 All cycles 250 -.1080.1058.1512 -.0095.0982.0986 500 -.0817.0778.1128 -.0108.0735.0742 1000 -.0637.0580.0861 -.0117.0543.0556 2000 -.0516.0418.0664 -.0110.0390.0405 30 / 31

Conclusion and Beyond We derived conditional moment inequalies for a semi-parametric panel multinomial choice model using the cyclic monotonicity of the choice-probability function. We show point identification of the coefficient parameter in the linear utility index. We propose a consistent estimator based on the conditional moment inequalities. We are still developing an asymptotic distribution theory for the estimator (or an alternative consistent estimator.) Further future: marginal effects. 31 / 31