Direction Estimation in a General Regression Model with Discrete Predictors

Size: px
Start display at page:

Download "Direction Estimation in a General Regression Model with Discrete Predictors"

Transcription

1 Direction Estimation in a General Regression Model with Discrete Predictors Yuexiao Dong and Zhou Yu Abstract Consider a general regression model, where the response Y depends on discrete predictors X only through the index β T X. It is well-known that the ordinary least squares (OLS) estimator can recover the underlying direction β exactly if the link function between Y and X is linear. Li and Duan (989) showed that the OLS estimator can recover β proportionally if the predictors satisfy the linear conditional mean (LCM) condition. For discrete predictors, we demonstrate that the LCM condition generally does not hold. To improve the OLS estimator in the presence of discrete predictors, we model the conditional mean E(X β T X) as a polynomial function of β T X and use the central solution space (CSS) estimator. The superior performances of the CSS estimators are confirmed through numerical studies. Introduction Consider univariate response Y R and q-dimensional predictor X R q. For β R q, let θ = β T X be a linear combination of the predictor X. Suppose the probability distribution function of Y depends on X only through θ, such that Y F θ (Y ), where θ = β T X. () Model () allows Y to be continuous or discrete, and we will refer to it as the general regression model. Many popular regression models can be formulated under this framework, such as linear regression, single index model, transformation models, and generalized linear regression. Yuexiao Dong Temple University, Philadelphia, USA, 9. ydong@temple.edu Zhou Yu East China Normal University, Shanghai, China, zyu@stat.ecnu.edu.cn

2 Yuexiao Dong and Zhou Yu As a special case of the general regression model (), single index model (SIM) considers Y = g(β T X) + ε for continuous response Y and unknown link function g : R R. SIM is a semiparametric model that allows for flexibility through the nonparametric link function and mitigates the curse of dimensionality through the parametric index structure. The estimation problem in SIM involves estimation of the index β and the link function g( ) simultaneously. Some classical references include Powell et al. (989), Härdle et al. (993), Ichimura (993), Horowitz and Härdle (996), Carroll et al. (997), etc. Cui et al. (0) considered a more general model, where the response can be continuous or discrete, and the single index appears in both the mean component and the variance component. Since nonparametric smoothing is involved in all the aforementioned methods, theses methods require that the link function g( ) is smooth and at least one predictor is continuous. To avoid nonparametric smoothing, methods based on distance covariance and Hilbert-Schmidt Independence Criterion are proposed, respectively, in Sheng and Yin (03) and Zhang and Yin (05). In this paper, we aim to estimate the index θ = β T X in the general regression model (), where all the predictors are discrete. To the best of our knowledge, Cook and Li (009) and Sheng and Yin (03) are the only methods in the literature that allow all the predictors to be discrete. Let X = (X,...,X q ) T. Cook and Li (009) assumed that the conditional distribution of X i Y follows a one-parameter exponential family distribution for i =,..., q, and used the maximum likelihood estimator of the inverse regression model to recover the direction β in the original forward regression. On the other hand, Sheng and Yin (03) argued that β can be recovered through maximizing the distance covariance between η T X and Y over η R p. The estimator in Sheng and Yin (03) is desirable as its asymptotic normality has been established under some technical conditions. We provide an alternative method for direction recovery with discrete predictors. Denote var(x) = Σ and assume E(X) = 0 without loss of generality. The classical ordinary least squares (OLS) estimator from the linear regression model is β OLS = Σ E(XY ). A surprising fact revealed by Li and Duan (989) is that β OLS is proportional to the true direction β in model () under the following linear conditional mean (LCM) condition E(X β T X) is a linear function of β T X. () For continuous X, it is well-known that if () holds for all β R, then X has to follow an elliptically-contoured distribution. Variable transformation and elliptical trimming (Cook and Nachtsheim, 994) are thus useful data preprocessing tools for continuous predictors. For discrete predictors, variable transformation and elliptical trimming no longer work, and new direction recovery method is needed when the LCM condition () is violated. This motivates us to consider the OLS estimator based on the central solution space (CSS). Although the CSS estimators have been systematically studied in Li and Dong (009), Dong and Li (00), and Dong and Yu (0), we are the first to study CSS estimators in the presence of exclusively discrete predictors.

3 Direction Estimation in a General Regression Model with Discrete Predictors 3 The rest of the paper is organized as follows. In Section, we examine the OLS estimator with discrete predictors and examine the role of the LCM condition. In Section 3, we demonstrate that the LCM condition is oftentimes violated with discrete predictors, and introduce the central solution space estimator as an extension of OLS. Numerical studies are performed in Section 4 and we conclude the paper with some discussions in Section 5. Two scenarios that OLS works We explore the two scenarios when β OLS can be used to recover the true direction β. In the first scenario, the response Y and the discrete predictor X have a linear relationship. In the second scenario, the discrete predictor X satisfies the LCM condition (). We assume E(X) = 0 unless specified otherwise.. OLS with linear link function In the case with continuous Y, suppose Y and X follow the linear model Y = α + β T X+ε, where α R, β R q, and ε is independent of X. Then obviously we have β OLS = Σ E(XY ) = ασ E(X) + Σ E(XX T )β + Σ E(X)E(ε) = β, where the last equality holds because E(X) = 0 and E(XX T ) = Σ. In the case with discrete Y, β OLS can be a good estimator of β when Y and X satisfy Y = α + β T X, which will be referred to as the no-error linear model. To fix the idea, consider the following two examples. We will see that the response and the predictors exactly follow the no-error linear model in Example, while the no-error linear model is a good approximation in Example. Example. Consider the zoo data from the UCI machine learning repository htt p : //archive.ics.uci.edu/ml/datasets/zoo. This data set contains 0 animals with 6 attributes. Except for the number of legs, all the other attributes (aquatic or not; feathers or not; milk or not; etc.) are binary. The response is the type of the animal, with mammals (4 cases) and birds (0 cases) being the two dominant types. We collapse the other 40 cases into the third category, which includes insects, fish, reptiles, molluscs, and amphibians. The sample OLS estimator has only two coefficients that are significantly different from zero. Not surprisingly, the corresponding attributes are feathers and milk. For this particular data set, all the animals that milk are mammals, and all the animals with feathers are birds. Example. Consider the congressional voting records from the 98th Congress, nd session in 984, which is available at the UCI machine learning repository htt p : //archive.ics.uci.edu/ml/datasets/congressional + Voting + Records. The data set consists of votes from 435 U.S. House of Representatives Congressmen on

4 4 Yuexiao Dong and Zhou Yu 6 subjects, with each vote being one of the three categories: yea, nay, and unknown disposition. The response here is whether the Congressman is democrat or republican. From the sample OLS estimator, we see that the most dominant coefficient corresponds to a bill related to physician fee freeze. It turns out that 45 of 67 democrats voted against this bill, while 63 of 68 republicans voted for the same bill.. OLS with the LCM condition For nonlinear link functions, Li and Duan (989) showed that β OLS is proportional to the true direction β in model () under the LCM condition (). For a,b R q, the Σ-inner product between a and b is a,b Σ = a T Σb. Let P Σ (β) = Σβ(β T Σβ) β T be the projection matrix onto the column space of β under the Σ-inner product. The LCM condition () then implies that E(X β T X) = P Σ (β)x. Under model (), we have E(XY ) = E{XE(Y X)} = E{XE(Y β T X)} = E{E(X β T X)Y }. (3) Plug in E(X β T X) = P Σ (β)x and pre-multiply Σ, we get β OLS = Σ E(XY ) = Σ P Σ (β)e(xy ) = β(β T Σβ) β T E(XY ) β, where means proportional to. REMARK. A related concept is sufficient dimension reduction (SDR) (Cook, 998). SDR aims to find B = {β,...,β d } R q d with the smallest possible column space, such that Y X B T X, (4) where means statistical independence. Correspondingly, the smallest column space is called the central space for the regression between Y and X, and is denoted as S Y X. (4) implies that Y depends on X only through B T X. Under the general regression model (), Y is independent of X given the index θ = β T X, which guarantees the second equality in (3). Model () can thus be viewed as a special case of (4) with the number of indices d =. Using the SDR terminology, the finding in Li and Duan (989) becomes that β OLS S Y X. REMARK. Sheng and Yin (03) proposed direction recovery in model () through distance covariance. They require a technical condition that P Σ (β)x Q Σ (β)x, (5) where Q Σ (β) = I q P Σ (β) with I q being the q q dimensional identity matrix. Note that Y Q Σ (β)x P Σ (β)x under model (). Together with (5), we have {Y,P Σ (β)x} Q Σ (β)x. It follows that

5 Direction Estimation in a General Regression Model with Discrete Predictors 5 E(XY ) = E{P Σ (β)xy } + E{Q Σ (β)xy } = E{P Σ (β)xy } = P Σ (β)e(xy ), where the second equality holds because E{Q Σ (β)xy } = Q Σ (β)e(x)e(y ) = 0. Under the same technical condition (5) as in Sheng and Yin (03), we have shown that β OLS β. Next we present an example with discrete predictors and demonstrate that β OLS β when the LCM condition () is satisfied. The square link function is used for the ease of demonstration. Other types of link functions lead to similar results. We do not require E(X) = 0, and the OLS estimator β OLS = Σ E{(X E(X))Y } is used in this example. Example 3. Consider X = (X,X ) T, β = (,) T, and Y = (β T X). Case i: X Poisson(l ), X Poisson(l ), and X X. It is easy to see X (X + X = m) Binomial(m,ρ) with ρ = l /(l + l ). It follows that E(X X + X ) = ρ(x + X ) is linear in X + X. Similarly E(X X + X ) is linear in X + X. Together the LCM condition () holds. On the other hand, we have l 0 l (l Σ = and E{(X E(X))Y } = + l + ). 0 l l (l + l + ) It follows that β OLS = Σ E{(X E(X))Y } = l + l + l + l + = β. Case ii: X Binomial(n, p), X Binomial(n, p), and X X. Then X (X + X = m) Hypergeometric(n + n,m,n ). It is easy to see that the LCM condition is satisfied. We also have n p 0 n p( p)( p + n Σ = and E{(X E(X))Y } = p + n p). 0 n p n p( p)( p + n p + n p) It follows that β OLS = Σ E{(X E(X))Y } = ( p)( p + n p + n p) ( p)( p + n p + n p) Case iii: (X,X,W) T Multinomial(n,(p, p, p 3 )). We have X (X + X = m) Binomial(m,π) with π = p /(p + p ). = β. Thus the LCM condition is satisfied. The explicit calculation of β OLS shows that it is proportional to β, and the details are provided in the Appendix.

6 6 Yuexiao Dong and Zhou Yu 3 Central solution space with discrete predictors We have seen in Section that the LCM condition plays an important role for β OLS to recover β in model (). We will see that the LCM condition is more often than not violated with discrete predictors. Transformation or trimming, which are popular methods for continuous predictors, are not applicable for discrete predictors. This motivates us to consider the central solution space estimator (Li and Dong, 009), which does not require the LCM condition. 3. Beyond the LCM condition Consider X = (X,X ) T, X Binomial(n =, p ), X Binomial(n = 3, p ), and X X. We are interested in E(X β T X) as a function of β T X for different combinations of p, p and β. The explicit form of E(X β T X) may not be easily derived. Instead, we generate N = 0 6 random samples and calculate the sample conditional mean E N (X β T X) as an approximation of E(X β T X). Specifically, let {X (i),i =,...,N} be an i.i.d. sample of X, where X (i) = (X (i) (i),x )T. Suppose Ω is the support of β T X. For any ω Ω, let I(β T X = ω) be the indicator function of β T X = ω. Then E(X β T X = ω) can be approximated by E N (X β T X = ω) = E N{X I(β T X = ω)} E N {I(β T X = ω)} = N i= X (i) I(β T X (i) = ω) N i= I(β T X (i) = ω) We plot E N (X β T X) versus β T X in Figure. The linear, quadratic and cubic fit between E N (X β T X) and β T X are also included. The upper left panel confirms the finding in case ii of Example 3. Namely, E(X β T X) is a linear function of β T X and the LCM condition is satisfied when p = p and β = (,) T. In the first row of Figure, we fix β = (,) T and modify p and p. It is clear that E(X β T X) is a nonlinear function of β T X when p p. In the second row of Figure, we fix p = p =. and modify β. We see that the LCM condition will not hold when β is not proportional to (,) T. Similar results hold true for cases i and iii of Example 3, which are not reported here. We see from Figure that the LCM condition could easily be violated with discrete predictors. The next example demonstrates that without the LCM condition, β OLS may not be able to recover β. Example 4. Let X = (X,X ) T, β = (,) T, and Y = (β T X) 3. Suppose X Bernoulli(p ), X Bernoulli(p ), and X X. Similar to Figure, one can check that the LCM condition is violated if p p. On the other hand, we have ( p 0 p p Σ = and E{(X E(X))Y } = + 6p p 6p p ) 0 p p p + 6p p 6p p..

7 Direction Estimation in a General Regression Model with Discrete Predictors 7 p =., p =., β T =(,) p =., p =.5, β T =(,) p =., p =.9, β T =(,) E N(X X + X) linear quadratic cubic E N(X X + X) linear quadratic cubic E N(X X + X) linear quadratic cubic X + X X + X X + X p =., p =., β T =(,) p =., p =., β T =(,) p =., p =., β T =(5,) E N(X X + X) linear quadratic cubic E N(X X + X) linear quadratic cubic E N(X 5X + X) linear quadratic cubic X + X X + X 5X + X Fig. E N (X β T X) v.s. β T X. N = 0 6, X Binomial (, p ), X Binomial (3, p ), and X independent of X. It follows that β OLS = Σ p + 6p E{(X E(X))Y } = 6p p, p + 6p 6p p which is not proportional to β = (,) T unless p = p. 3. The central solution space estimator The OLS-based central solution space (CSS) estimator was first proposed in Li and Dong (009), which does not require the LCM condition. Although Li and Dong (009) focused on CSS with continuous predictors, the CSS method can be applied to discrete predictors as well. Suppose E(X) = 0. Under model (), we observe the following equality E(XY ) = E{E(X β T X)Y }. (6)

8 8 Yuexiao Dong and Zhou Yu Denote the solution of equation (6) as β CSS. The following result is a special case of Theorem 4. in Li and Dong (009). Proposition. Suppose Pr(E{E(X η T X)Y } = E{E(X β T X)Y }) > 0 whenever η is not proportional to β. Then under model (), the solution of equation (6) is unique up to a scalar multiplication, and satisfies β CSS β. Furthermore, if the LCM condition () holds, then β CSS β OLS. The first part of Proposition states that one can use β CSS to recover β without the LCM condition (). The second part reveals that β CSS coincides with β OLS when the LCM condition is true. Ichimura (993) provided an example that the parameters in the index β are unidentifiable when the predictors are all discrete. Our assumption in Proposition eliminates such pathological situations, and guarantees that the CSS estimator is unique even with exclusively discrete predictors. For a = (a,...,a q ) T R q, let a = q i= a i. To find the solution of equation (6), Li and Dong (009) suggested solving the following optimization problem minimize L(η) over η R q, where L(η) = E{XY E(X η T X)Y }. (7) Denote the minimizer of L(η) as η 0. As long as all the moments involved exist, Theorem 4. in Li and Dong (009) states that η 0 is proportional to β CSS under the same condition as in Proposition. The sample level CSS estimator is based on minimizing L N (η), the sample objective function of L(η) in (7). Motivated from the observations in Section 3., we model E(X η T X) as a polynomial function of η T X. As we have seen in Figure, for discrete predictors, using quadratic and cubic functions of η T X may be more suitable than using linear functions of η T X. For any function r(x,y ) and an i.i.d. sample {(X (i),y (i) ),i =,...,N}, we denote the sample estimator of E{r(X,Y )} by E N {r(x,y )} = N N i= r(x(i),y (i) ). The step-by-step algorithm is as follows.. Center Y (i) and X (i) as Ŷ (i) = Y (i) E N (Y ), ˆX (i) = X (i) E N (X).. Set k = or k = 3. Denote G(η T ˆX (i) ) as {,η T ˆX (i),...,(η T ˆX (i) ) k } T. The sample estimator of E( ˆX η T ˆX) becomes E N ( ˆX η T ˆX) = E N { ˆXG T (η T ˆX)} ( E N {G(η T ˆX)G T (η T ˆX)} ) G(η T ˆX). 3. Estimate β by solving the following optimization problem minimize L N (η) over η R q, where L N (η) = E N { ˆXŶ E N ( ˆX η T ˆX)Ŷ }, where the minimizer is denoted as ˆβ (k) CSS. A Newton-Raphson type estimator can be implemented to find the minimizer in step 3 of the above algorithm. See, for example, Cook and Li (009), Dong and Zhu (0).

9 Direction Estimation in a General Regression Model with Discrete Predictors 9 4 Numerical studies In this section, we demonstrate the effectiveness of CSS estimator with discrete predictors via simulation studies. For X = (X,...,X q ) T and β R q, consider three settings of β T X as follows, Case (i): β = (,,,0,...,0) T,X Binomial(,.),X Binomial(3,.9), X 3,...,X q Bernoulli(.5), and X i X j for i j. Case (ii): β = (,,3,0,...,0) T,X Poisson(),X Poisson(.5), X 3,...,X q Bernoulli(.5), and X i X j for i j. Case (iii): β = (,0,,0,...,0) T,(X,X,W) T Multinomial(4,(.,.4,.4)), X 3,...,X q Bernoulli(.5), and X i X j for i j except for X and X. For each combination of β and X, consider the following models Model I: Y =.exp(β T X + ) +.ε, Model II: Y =.(β T X) 3 +.ε, { if β T X 0, Model III: Y = 0 if β T X < 0. In Model I and II, we have continuous response with error ε N(0,) and ε X. The response in Model III is discrete. All three models are special cases of the general regression model (). Denote Σ N as the sample covariance matrix of X. Then the classical ordinary least squares estimator is ˆβ OLS = Σ N E N{(X E N (X))Y }. Two () (3) CSS estimators ˆβ CSS and ˆβ CSS are also included for the comparison, where we estimate E(X β T X) as a quadratic or a cubic function of β T X. Let R ( ˆβ,β) be the squared sample Pearson correlation between β T X and ˆβ T X, which will be used to measure the performances of different estimators. We fix predictor dimension q = 0, sample size N = 00, and report the simulation results in Table based on 00 repetitions. Each entry of Table is formatted as a(b), where a is the average of R ( ˆβ,β) across the 00 repetitions, and b is the standard error of the average. We see that settings, while ˆβ () CSS outperforms ˆβ OLS in six out of nine ˆβ (3) CSS consistently outperforms ˆβ OLS in all nine settings. ˆβ (3) CSS enjoys the best overall performances, as it is better than CSS in all but one setting. This simulation study confirms our intuition in Section 3. Namely, when the LCM condition is violated and E(X β T X) can not be modeled by linear functions of β T X, the CCS estimators can improve the OLS estimator as long as we can model E(X β T X) properly. ˆβ ()

10 0 Yuexiao Dong and Zhou Yu Table Comparison of OLS and CSS estimators based on 00 repetitions. Case Model R ( ˆβ OLS,β ) R ( ˆβ () CSS,β) R ( ˆβ (3) CSS,β) I.96 (.003).956 (.00).97 (.008) (i) II.964 (.00).9984 (.000).9994 (.000) III.8038 (.0084).7976 (.044).8700 (.0099) I.97 (.005).9547 (.000).9803 (.003) (ii) II.8997 (.0035).980 (.0034).9430 (.0040) III.9593 (.000).9543 (.00).9803 (.0009) I.960 (.008).9855 (.000).9977 (.0004) (iii) II.9389 (.003).9647 (.005).9880 (.0008) III.90 (.009).9846 (.0009).9794 (.005) 5 Discussion To recover the direction β in the general regression model () with discrete predictors, we examine the OLS and the CSS-based OLS estimators in this paper. In the case when the link function is linear, or when the predictor distribution satisfies the LCM condition, OLS can recover the true direction up to a multiplicative scalar. If the link function is nonlinear and the LCM condition is not satisfied, OLS is no longer suitable. By relaxing the LCM condition and modeling E(X β T X) as polynomial functions of β T X, the CSS-based estimators improve OLS and lead to better accuracy at recovering β. Our development addresses a gap in the existing CSS literature, which mainly focuses on continuous predictors (Li and Dong, 009; Dong and Li,00; Dong and Yu, 0). In the case of continuous predictors, Theorem 6. in Li and Dong (009) provides the consistency results of the CSS estimators. From Figure, it is obvious that E(X β T X) can not be consistently estimated by polynomial functions of β T X. Finding consistent CSS estimators with discrete predictors is a topic worth future investigation. Appendix Proof of β OLS β in Example 3, case iii. For (X,X,W) T Multinomial (n,(p, p, p 3 )), it is well-known that X i Binomial(n, p i ) for i =,, and cov(x,x ) = np p. It follows that E(X X ) = n(n )p p. It can be shown that E(X 3 ) = np + 3n(n )p + n(n )(n )p 3. Next we denote k 3 = n k k and calculate E(X X ) as follows

11 Direction Estimation in a General Regression Model with Discrete Predictors E(X X ) = k n! k k,k k!k!k 3! pk pk pk 3 3 (n )! = n(n )p p (k + ) k,k (k )!(k )!k 3! pk p k p k 3 3 = n(n )(n )p (n 3)! p k,k (k )!(k )!k 3! pk p k p k 3 3 (n )! + n(n )p p k,k (k )!(k )!k 3! pk p k p k 3 3 = n(n )p p ( + np p ). For Y = (X + X ), we thus have E{(X E(X ))Y } = E(X 3 ) E(X )E(X ) + E(X X ) + E(X X ) E(X )E(X ) + E(X X ) E(X )E(X X ) = np { + (n 3)p (n )p + p (n 3 (n )p (4n 4)p )}. Similarly, E{(X E(X ))Y } is equal to np { + (n 3)p (n )p + p (n 3 (n )p (4n 4)p )}. Recall that X = (X,X ) T and var(x) = Σ. Let Σ be the determinant of Σ. Since the first row of Σ is Σ {np ( p ),np p }, the first component of β OLS becomes Σ [np ( p )E{(X E(X ))Y } + np p E{(X E(X ))Y }] = Σ n p p { + (n 3)(p + p ) (n )(p + p )}. Due to the symmetry between p and p in the expression above, the second component of β OLS is exactly the same as the first component. Thus we have β OLS (,) T = β. PROOF OF PROPOSITION. Assume E{E(X η T X)Y } = E{E(X β T X)Y } with probability for some η, such that η is not proportional to β. Then both η and β will satisfy equation (6), which means the solution of (6) is not unique up to a s- calar multiplication. Thus under the assumption that Pr(E{E(X η T X)Y } = E{E(X β T X)Y }) > 0 whenever η is not proportional to β, the solution of (6) is unique. Because β satisfies (6) and the solution of (6) is unique up to a scalar multiplication, we have β CSS β. Under the additional LCM condition (), β OLS is also proportional to β. Consequently we have β CSS β OLS.

12 Yuexiao Dong and Zhou Yu References. Carroll, R. J., Fan, J., Gijbels, I., Wand, M. P. (997). Generalized partially linear single-index models. Journal of the American Statistical Association, 9, Cook, R. D. (998). Regression Graphics. New York: Wiley. 3. Cook, R. D., Li, L. (009). Dimension reduction in regressions with exponential family predictors. Journal of Computational and Graphical Statistics, 8, Cook, R. D., Nachtsheim, C. (994). Reweighting to achieve elliptically contoured covariates in regression. Journal of the American Statistical Association, 89, Cui, X., Härdle, W., Zhu, L. X. (0). The EFM approach for single-index models. The Annals of Statistics, 39, Dong, Y., Li, B. (00). Dimension reduction for non-elliptically distributed predictors: second order methods. Biometrika, 97, Dong, Y., Yu, Z. (0). Dimension reduction for the conditional kth moment via central solution space. Journal of Multivariate Analysis,, Dong, Y., Zhu, L. P. (0). A note on sliced inverse regression with missing predictors. Statistical Analysis and Data Mining, 5, Härdle, W., Hall, P., Ichimura, H. (993). Optimal smoothing in single-index models. The Annals of Statistics,, Horowitz, J. L., Härdle, W. (996). Direct semiparametric estimation of single-index models with discrete covariates. Journal of the American Statistical Association, 9, Ichimura, H. (993). Semiparametric least square (SLS) and weighted SLS estimation of single-index models. Journal of Econometrics, 58, Li, B., Dong, Y. (009). Dimension reduction for non-elliptically distributed predictors. The Annals of Statistics, 37, Li, K. C., Duan, N. (989). Regression analysis under link violation. The Annals of Statistics, 7, Powell, J. L., Stock, J. M., Stoker, T. M. (989). Semiparametric estimation of index coefficients. Econometrica, 57, Sheng, W., Yin, X. (03). Direction estimation in single-index models via distance covariance. Journal of Multivariate Analysis,, Zhang, N., Yin, X. (05). Direction estimation in single-index regressions via Hilbert- Schmidt independence criterion. Statistica Sinica, 5,

SEMIPARAMETRIC ESTIMATION OF CONDITIONAL HETEROSCEDASTICITY VIA SINGLE-INDEX MODELING

SEMIPARAMETRIC ESTIMATION OF CONDITIONAL HETEROSCEDASTICITY VIA SINGLE-INDEX MODELING Statistica Sinica 3 (013), 135-155 doi:http://dx.doi.org/10.5705/ss.01.075 SEMIPARAMERIC ESIMAION OF CONDIIONAL HEEROSCEDASICIY VIA SINGLE-INDEX MODELING Liping Zhu, Yuexiao Dong and Runze Li Shanghai

More information

Regression Graphics. R. D. Cook Department of Applied Statistics University of Minnesota St. Paul, MN 55108

Regression Graphics. R. D. Cook Department of Applied Statistics University of Minnesota St. Paul, MN 55108 Regression Graphics R. D. Cook Department of Applied Statistics University of Minnesota St. Paul, MN 55108 Abstract This article, which is based on an Interface tutorial, presents an overview of regression

More information

Regression Graphics. 1 Introduction. 2 The Central Subspace. R. D. Cook Department of Applied Statistics University of Minnesota St.

Regression Graphics. 1 Introduction. 2 The Central Subspace. R. D. Cook Department of Applied Statistics University of Minnesota St. Regression Graphics R. D. Cook Department of Applied Statistics University of Minnesota St. Paul, MN 55108 Abstract This article, which is based on an Interface tutorial, presents an overview of regression

More information

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas 0 0 5 Motivation: Regression discontinuity (Angrist&Pischke) Outcome.5 1 1.5 A. Linear E[Y 0i X i] 0.2.4.6.8 1 X Outcome.5 1 1.5 B. Nonlinear E[Y 0i X i] i 0.2.4.6.8 1 X utcome.5 1 1.5 C. Nonlinearity

More information

Sufficient Dimension Reduction using Support Vector Machine and it s variants

Sufficient Dimension Reduction using Support Vector Machine and it s variants Sufficient Dimension Reduction using Support Vector Machine and it s variants Andreas Artemiou School of Mathematics, Cardiff University @AG DANK/BCS Meeting 2013 SDR PSVM Real Data Current Research and

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;

More information

Final Exam # 3. Sta 230: Probability. December 16, 2012

Final Exam # 3. Sta 230: Probability. December 16, 2012 Final Exam # 3 Sta 230: Probability December 16, 2012 This is a closed-book exam so do not refer to your notes, the text, or any other books (please put them on the floor). You may use the extra sheets

More information

COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS. Abstract

COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS. Abstract Far East J. Theo. Stat. 0() (006), 179-196 COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS Department of Statistics University of Manitoba Winnipeg, Manitoba, Canada R3T

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Practice Examination # 3

Practice Examination # 3 Practice Examination # 3 Sta 23: Probability December 13, 212 This is a closed-book exam so do not refer to your notes, the text, or any other books (please put them on the floor). You may use a single

More information

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS020) p.3863 Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Jinfang Wang and

More information

Nonparametric Econometrics

Nonparametric Econometrics Applied Microeconometrics with Stata Nonparametric Econometrics Spring Term 2011 1 / 37 Contents Introduction The histogram estimator The kernel density estimator Nonparametric regression estimators Semi-

More information

A review on Sliced Inverse Regression

A review on Sliced Inverse Regression A review on Sliced Inverse Regression Kevin Li To cite this version: Kevin Li. A review on Sliced Inverse Regression. 2013. HAL Id: hal-00803698 https://hal.archives-ouvertes.fr/hal-00803698v1

More information

Econometric Analysis of Cross Section and Panel Data

Econometric Analysis of Cross Section and Panel Data Econometric Analysis of Cross Section and Panel Data Jeffrey M. Wooldridge / The MIT Press Cambridge, Massachusetts London, England Contents Preface Acknowledgments xvii xxiii I INTRODUCTION AND BACKGROUND

More information

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception LINEAR MODELS FOR CLASSIFICATION Classification: Problem Statement 2 In regression, we are modeling the relationship between a continuous input variable x and a continuous target variable t. In classification,

More information

ECO Class 6 Nonparametric Econometrics

ECO Class 6 Nonparametric Econometrics ECO 523 - Class 6 Nonparametric Econometrics Carolina Caetano Contents 1 Nonparametric instrumental variable regression 1 2 Nonparametric Estimation of Average Treatment Effects 3 2.1 Asymptotic results................................

More information

Summer School in Statistics for Astronomers V June 1 - June 6, Regression. Mosuk Chow Statistics Department Penn State University.

Summer School in Statistics for Astronomers V June 1 - June 6, Regression. Mosuk Chow Statistics Department Penn State University. Summer School in Statistics for Astronomers V June 1 - June 6, 2009 Regression Mosuk Chow Statistics Department Penn State University. Adapted from notes prepared by RL Karandikar Mean and variance Recall

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

Flexible Estimation of Treatment Effect Parameters

Flexible Estimation of Treatment Effect Parameters Flexible Estimation of Treatment Effect Parameters Thomas MaCurdy a and Xiaohong Chen b and Han Hong c Introduction Many empirical studies of program evaluations are complicated by the presence of both

More information

A review of some semiparametric regression models with application to scoring

A review of some semiparametric regression models with application to scoring A review of some semiparametric regression models with application to scoring Jean-Loïc Berthet 1 and Valentin Patilea 2 1 ENSAI Campus de Ker-Lann Rue Blaise Pascal - BP 37203 35172 Bruz cedex, France

More information

Binary choice 3.3 Maximum likelihood estimation

Binary choice 3.3 Maximum likelihood estimation Binary choice 3.3 Maximum likelihood estimation Michel Bierlaire Output of the estimation We explain here the various outputs from the maximum likelihood estimation procedure. Solution of the maximum likelihood

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

CS281A/Stat241A Lecture 17

CS281A/Stat241A Lecture 17 CS281A/Stat241A Lecture 17 p. 1/4 CS281A/Stat241A Lecture 17 Factor Analysis and State Space Models Peter Bartlett CS281A/Stat241A Lecture 17 p. 2/4 Key ideas of this lecture Factor Analysis. Recall: Gaussian

More information

Interaction effects for continuous predictors in regression modeling

Interaction effects for continuous predictors in regression modeling Interaction effects for continuous predictors in regression modeling Testing for interactions The linear regression model is undoubtedly the most commonly-used statistical model, and has the advantage

More information

Linear model A linear model assumes Y X N(µ(X),σ 2 I), And IE(Y X) = µ(x) = X β, 2/52

Linear model A linear model assumes Y X N(µ(X),σ 2 I), And IE(Y X) = µ(x) = X β, 2/52 Statistics for Applications Chapter 10: Generalized Linear Models (GLMs) 1/52 Linear model A linear model assumes Y X N(µ(X),σ 2 I), And IE(Y X) = µ(x) = X β, 2/52 Components of a linear model The two

More information

TESTING SERIAL CORRELATION IN SEMIPARAMETRIC VARYING COEFFICIENT PARTIALLY LINEAR ERRORS-IN-VARIABLES MODEL

TESTING SERIAL CORRELATION IN SEMIPARAMETRIC VARYING COEFFICIENT PARTIALLY LINEAR ERRORS-IN-VARIABLES MODEL Jrl Syst Sci & Complexity (2009) 22: 483 494 TESTIG SERIAL CORRELATIO I SEMIPARAMETRIC VARYIG COEFFICIET PARTIALLY LIEAR ERRORS-I-VARIABLES MODEL Xuemei HU Feng LIU Zhizhong WAG Received: 19 September

More information

Introduction to Econometrics Final Examination Fall 2006 Answer Sheet

Introduction to Econometrics Final Examination Fall 2006 Answer Sheet Introduction to Econometrics Final Examination Fall 2006 Answer Sheet Please answer all of the questions and show your work. If you think a question is ambiguous, clearly state how you interpret it before

More information

Linear regression COMS 4771

Linear regression COMS 4771 Linear regression COMS 4771 1. Old Faithful and prediction functions Prediction problem: Old Faithful geyser (Yellowstone) Task: Predict time of next eruption. 1 / 40 Statistical model for time between

More information

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite

More information

SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES. Liping Zhu, Mian Huang, & Runze Li. The Pennsylvania State University

SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES. Liping Zhu, Mian Huang, & Runze Li. The Pennsylvania State University SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES Liping Zhu, Mian Huang, & Runze Li The Pennsylvania State University Technical Report Series #10-104 College of Health and Human Development

More information

Chapter 5. Chapter 5 sections

Chapter 5. Chapter 5 sections 1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Revision: Probability and Linear Algebra Week 1, Lecture 2

MA 575 Linear Models: Cedric E. Ginestet, Boston University Revision: Probability and Linear Algebra Week 1, Lecture 2 MA 575 Linear Models: Cedric E Ginestet, Boston University Revision: Probability and Linear Algebra Week 1, Lecture 2 1 Revision: Probability Theory 11 Random Variables A real-valued random variable is

More information

A Selective Review of Sufficient Dimension Reduction

A Selective Review of Sufficient Dimension Reduction A Selective Review of Sufficient Dimension Reduction Lexin Li Department of Statistics North Carolina State University Lexin Li (NCSU) Sufficient Dimension Reduction 1 / 19 Outline 1 General Framework

More information

Sliced Regression for Dimension Reduction

Sliced Regression for Dimension Reduction Sliced Regression for Dimension Reduction Hansheng Wang and Yingcun Xia Peking University & National University of Singapore Current Version: December 22, 2008. Abstract By slicing the region of the response

More information

A Note on Visualizing Response Transformations in Regression

A Note on Visualizing Response Transformations in Regression Southern Illinois University Carbondale OpenSIUC Articles and Preprints Department of Mathematics 11-2001 A Note on Visualizing Response Transformations in Regression R. Dennis Cook University of Minnesota

More information

The Multivariate Gaussian Distribution

The Multivariate Gaussian Distribution The Multivariate Gaussian Distribution Chuong B. Do October, 8 A vector-valued random variable X = T X X n is said to have a multivariate normal or Gaussian) distribution with mean µ R n and covariance

More information

Nonparametric Principal Components Regression

Nonparametric Principal Components Regression Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session CPS031) p.4574 Nonparametric Principal Components Regression Barrios, Erniel University of the Philippines Diliman,

More information

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Jianqing Fan Department of Statistics Chinese University of Hong Kong AND Department of Statistics

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

High-dimensional Ordinary Least-squares Projection for Screening Variables

High-dimensional Ordinary Least-squares Projection for Screening Variables 1 / 38 High-dimensional Ordinary Least-squares Projection for Screening Variables Chenlei Leng Joint with Xiangyu Wang (Duke) Conference on Nonparametric Statistics for Big Data and Celebration to Honor

More information

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract

More information

Ultra High Dimensional Variable Selection with Endogenous Variables

Ultra High Dimensional Variable Selection with Endogenous Variables 1 / 39 Ultra High Dimensional Variable Selection with Endogenous Variables Yuan Liao Princeton University Joint work with Jianqing Fan Job Market Talk January, 2012 2 / 39 Outline 1 Examples of Ultra High

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis. 401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis

More information

Linear Models in Machine Learning

Linear Models in Machine Learning CS540 Intro to AI Linear Models in Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu We briefly go over two linear models frequently used in machine learning: linear regression for, well, regression,

More information

Tobit Model Estimation and Sliced Inverse Regression

Tobit Model Estimation and Sliced Inverse Regression Tobit Model Estimation and Sliced Inverse Regression Lexin Li Department of Statistics North Carolina State University E-mail: li@stat.ncsu.edu Jeffrey S. Simonoff Leonard N. Stern School of Business New

More information

Lecture 3: Statistical Decision Theory (Part II)

Lecture 3: Statistical Decision Theory (Part II) Lecture 3: Statistical Decision Theory (Part II) Hao Helen Zhang Hao Helen Zhang Lecture 3: Statistical Decision Theory (Part II) 1 / 27 Outline of This Note Part I: Statistics Decision Theory (Classical

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

Dealing With Endogeneity

Dealing With Endogeneity Dealing With Endogeneity Junhui Qian December 22, 2014 Outline Introduction Instrumental Variable Instrumental Variable Estimation Two-Stage Least Square Estimation Panel Data Endogeneity in Econometrics

More information

An Introduction to Nonstationary Time Series Analysis

An Introduction to Nonstationary Time Series Analysis An Introduction to Analysis Ting Zhang 1 tingz@bu.edu Department of Mathematics and Statistics Boston University August 15, 2016 Boston University/Keio University Workshop 2016 A Presentation Friendly

More information

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28 Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:

More information

Simple Estimators for Semiparametric Multinomial Choice Models

Simple Estimators for Semiparametric Multinomial Choice Models Simple Estimators for Semiparametric Multinomial Choice Models James L. Powell and Paul A. Ruud University of California, Berkeley March 2008 Preliminary and Incomplete Comments Welcome Abstract This paper

More information

Efficiency loss and the linearity condition in dimension reduction

Efficiency loss and the linearity condition in dimension reduction Biometrika (203), 00,2,pp. 37 383 doi: 0.093/biomet/ass075 C 203 Biometrika Trust Advance Access publication 24 February 203 Printed in Great Britain Efficiency loss and the linearity condition in dimension

More information

Statistics 135 Fall 2008 Final Exam

Statistics 135 Fall 2008 Final Exam Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations

More information

A Note on Bootstraps and Robustness. Tony Lancaster, Brown University, December 2003.

A Note on Bootstraps and Robustness. Tony Lancaster, Brown University, December 2003. A Note on Bootstraps and Robustness Tony Lancaster, Brown University, December 2003. In this note we consider several versions of the bootstrap and argue that it is helpful in explaining and thinking about

More information

STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song

STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song Presenter: Jiwei Zhao Department of Statistics University of Wisconsin Madison April

More information

Testing Restrictions and Comparing Models

Testing Restrictions and Comparing Models Econ. 513, Time Series Econometrics Fall 00 Chris Sims Testing Restrictions and Comparing Models 1. THE PROBLEM We consider here the problem of comparing two parametric models for the data X, defined by

More information

Linear Regression With Special Variables

Linear Regression With Special Variables Linear Regression With Special Variables Junhui Qian December 21, 2014 Outline Standardized Scores Quadratic Terms Interaction Terms Binary Explanatory Variables Binary Choice Models Standardized Scores:

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

Advanced Statistics I : Gaussian Linear Model (and beyond)

Advanced Statistics I : Gaussian Linear Model (and beyond) Advanced Statistics I : Gaussian Linear Model (and beyond) Aurélien Garivier CNRS / Telecom ParisTech Centrale Outline One and Two-Sample Statistics Linear Gaussian Model Model Reduction and model Selection

More information

Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty

Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty Journal of Data Science 9(2011), 549-564 Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty Masaru Kanba and Kanta Naito Shimane University Abstract: This paper discusses the

More information

Sliced Inverse Regression

Sliced Inverse Regression Sliced Inverse Regression Ge Zhao gzz13@psu.edu Department of Statistics The Pennsylvania State University Outline Background of Sliced Inverse Regression (SIR) Dimension Reduction Definition of SIR Inversed

More information

Binary Models with Endogenous Explanatory Variables

Binary Models with Endogenous Explanatory Variables Binary Models with Endogenous Explanatory Variables Class otes Manuel Arellano ovember 7, 2007 Revised: January 21, 2008 1 Introduction In Part I we considered linear and non-linear models with additive

More information

BTRY 4090: Spring 2009 Theory of Statistics

BTRY 4090: Spring 2009 Theory of Statistics BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown.

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown. Weighting We have seen that if E(Y) = Xβ and V (Y) = σ 2 G, where G is known, the model can be rewritten as a linear model. This is known as generalized least squares or, if G is diagonal, with trace(g)

More information

Simulating Realistic Ecological Count Data

Simulating Realistic Ecological Count Data 1 / 76 Simulating Realistic Ecological Count Data Lisa Madsen Dave Birkes Oregon State University Statistics Department Seminar May 2, 2011 2 / 76 Outline 1 Motivation Example: Weed Counts 2 Pearson Correlation

More information

Using Estimating Equations for Spatially Correlated A

Using Estimating Equations for Spatially Correlated A Using Estimating Equations for Spatially Correlated Areal Data December 8, 2009 Introduction GEEs Spatial Estimating Equations Implementation Simulation Conclusion Typical Problem Assess the relationship

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University JSM, 2015 E. Christou, M. G. Akritas (PSU) SIQR JSM, 2015

More information

Identification and Estimation of Partially Linear Censored Regression Models with Unknown Heteroscedasticity

Identification and Estimation of Partially Linear Censored Regression Models with Unknown Heteroscedasticity Identification and Estimation of Partially Linear Censored Regression Models with Unknown Heteroscedasticity Zhengyu Zhang School of Economics Shanghai University of Finance and Economics zy.zhang@mail.shufe.edu.cn

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Advanced Methods for Data Analysis (36-402/36-608 Spring 2014 1 Generalized linear models 1.1 Introduction: two regressions So far we ve seen two canonical settings for regression.

More information

Rejoinder. 1 Phase I and Phase II Profile Monitoring. Peihua Qiu 1, Changliang Zou 2 and Zhaojun Wang 2

Rejoinder. 1 Phase I and Phase II Profile Monitoring. Peihua Qiu 1, Changliang Zou 2 and Zhaojun Wang 2 Rejoinder Peihua Qiu 1, Changliang Zou 2 and Zhaojun Wang 2 1 School of Statistics, University of Minnesota 2 LPMC and Department of Statistics, Nankai University, China We thank the editor Professor David

More information

Random variables (discrete)

Random variables (discrete) Random variables (discrete) Saad Mneimneh 1 Introducing random variables A random variable is a mapping from the sample space to the real line. We usually denote the random variable by X, and a value that

More information

Comment on Article by Scutari

Comment on Article by Scutari Bayesian Analysis (2013) 8, Number 3, pp. 543 548 Comment on Article by Scutari Hao Wang Scutari s paper studies properties of the distribution of graphs ppgq. This is an interesting angle because it differs

More information

Learning discrete graphical models via generalized inverse covariance matrices

Learning discrete graphical models via generalized inverse covariance matrices Learning discrete graphical models via generalized inverse covariance matrices Duzhe Wang, Yiming Lv, Yongjoon Kim, Young Lee Department of Statistics University of Wisconsin-Madison {dwang282, lv23, ykim676,

More information

Regression. ECO 312 Fall 2013 Chris Sims. January 12, 2014

Regression. ECO 312 Fall 2013 Chris Sims. January 12, 2014 ECO 312 Fall 2013 Chris Sims Regression January 12, 2014 c 2014 by Christopher A. Sims. This document is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported License What

More information

Logistic Regression. Seungjin Choi

Logistic Regression. Seungjin Choi Logistic Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

STAT 540: Data Analysis and Regression

STAT 540: Data Analysis and Regression STAT 540: Data Analysis and Regression Wen Zhou http://www.stat.colostate.edu/~riczw/ Email: riczw@stat.colostate.edu Department of Statistics Colorado State University Fall 205 W. Zhou (Colorado State

More information

Multinomial Logistic Regression Models

Multinomial Logistic Regression Models Stat 544, Lecture 19 1 Multinomial Logistic Regression Models Polytomous responses. Logistic regression can be extended to handle responses that are polytomous, i.e. taking r>2 categories. (Note: The word

More information

Cross-Validation with Confidence

Cross-Validation with Confidence Cross-Validation with Confidence Jing Lei Department of Statistics, Carnegie Mellon University WHOA-PSI Workshop, St Louis, 2017 Quotes from Day 1 and Day 2 Good model or pure model? Occam s razor We really

More information

LOGISTIC REGRESSION Joseph M. Hilbe

LOGISTIC REGRESSION Joseph M. Hilbe LOGISTIC REGRESSION Joseph M. Hilbe Arizona State University Logistic regression is the most common method used to model binary response data. When the response is binary, it typically takes the form of

More information

TUTORIAL 8 SOLUTIONS #

TUTORIAL 8 SOLUTIONS # TUTORIAL 8 SOLUTIONS #9.11.21 Suppose that a single observation X is taken from a uniform density on [0,θ], and consider testing H 0 : θ = 1 versus H 1 : θ =2. (a) Find a test that has significance level

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

A Simple Plot for Model Assessment

A Simple Plot for Model Assessment A Simple Plot for Model Assessment David J. Olive Southern Illinois University September 16, 2005 Abstract Regression is the study of the conditional distribution y x of the response y given the predictors

More information

Outline of GLMs. Definitions

Outline of GLMs. Definitions Outline of GLMs Definitions This is a short outline of GLM details, adapted from the book Nonparametric Regression and Generalized Linear Models, by Green and Silverman. The responses Y i have density

More information

Calibration Estimation of Semiparametric Copula Models with Data Missing at Random

Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Shigeyuki Hamori 1 Kaiji Motegi 1 Zheng Zhang 2 1 Kobe University 2 Renmin University of China Econometrics Workshop UNC

More information

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality.

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality. 88 Chapter 5 Distribution Theory In this chapter, we summarize the distributions related to the normal distribution that occur in linear models. Before turning to this general problem that assumes normal

More information

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition Last updated: Oct 22, 2012 LINEAR CLASSIFIERS Problems 2 Please do Problem 8.3 in the textbook. We will discuss this in class. Classification: Problem Statement 3 In regression, we are modeling the relationship

More information

5 CORRELATION. Expectation of the Binomial Distribution I The Binomial distribution can be defined as: P(X = r) = p r q n r where p + q = 1 and 0 r n

5 CORRELATION. Expectation of the Binomial Distribution I The Binomial distribution can be defined as: P(X = r) = p r q n r where p + q = 1 and 0 r n 5 CORRELATION The covariance of two random variables gives some measure of their independence. A second way of assessing the measure of independence will be discussed shortly but first the expectation

More information

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables. Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just

More information

Stat 5101 Notes: Brand Name Distributions

Stat 5101 Notes: Brand Name Distributions Stat 5101 Notes: Brand Name Distributions Charles J. Geyer September 5, 2012 Contents 1 Discrete Uniform Distribution 2 2 General Discrete Uniform Distribution 2 3 Uniform Distribution 3 4 General Uniform

More information

STAT 4385 Topic 01: Introduction & Review

STAT 4385 Topic 01: Introduction & Review STAT 4385 Topic 01: Introduction & Review Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso xsu@utep.edu Spring, 2016 Outline Welcome What is Regression Analysis? Basics

More information

On a Nonparametric Notion of Residual and its Applications

On a Nonparametric Notion of Residual and its Applications On a Nonparametric Notion of Residual and its Applications Bodhisattva Sen and Gábor Székely arxiv:1409.3886v1 [stat.me] 12 Sep 2014 Columbia University and National Science Foundation September 16, 2014

More information

Cross-Validation with Confidence

Cross-Validation with Confidence Cross-Validation with Confidence Jing Lei Department of Statistics, Carnegie Mellon University UMN Statistics Seminar, Mar 30, 2017 Overview Parameter est. Model selection Point est. MLE, M-est.,... Cross-validation

More information

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models University of Illinois Fall 2016 Department of Economics Roger Koenker Economics 536 Lecture 7 Introduction to Specification Testing in Dynamic Econometric Models In this lecture I want to briefly describe

More information

This paper is not to be removed from the Examination Halls

This paper is not to be removed from the Examination Halls ~~ST104B ZA d0 This paper is not to be removed from the Examination Halls UNIVERSITY OF LONDON ST104B ZB BSc degrees and Diplomas for Graduates in Economics, Management, Finance and the Social Sciences,

More information

Well-developed and understood properties

Well-developed and understood properties 1 INTRODUCTION TO LINEAR MODELS 1 THE CLASSICAL LINEAR MODEL Most commonly used statistical models Flexible models Well-developed and understood properties Ease of interpretation Building block for more

More information

Econometrics I Lecture 3: The Simple Linear Regression Model

Econometrics I Lecture 3: The Simple Linear Regression Model Econometrics I Lecture 3: The Simple Linear Regression Model Mohammad Vesal Graduate School of Management and Economics Sharif University of Technology 44716 Fall 1397 1 / 32 Outline Introduction Estimating

More information