Estimating subgroup specific treatment effects via concave fusion

Size: px
Start display at page:

Download "Estimating subgroup specific treatment effects via concave fusion"

Transcription

1 Estimating subgroup specific treatment effects via concave fusion Jian Huang University of Iowa April 6, 2016

2 Outline 1 Motivation and the problem 2 The proposed model and approach Concave pairwise fusion Computational algorithm: ADMM Initial value and solution path 3 Statistical properties 4 Numerical studies Simulations ACTG175 data 5 Concluding remarks

3 Motivation: Precision medicine Most medical treatments have been designed for the average patient. As a result of this one-size-ts-all approach, treatments can be very successful for some patients but not for others. Precision medicine is an approach to disease treatment and prevention that seeks to maximize eectiveness by taking into account individual variability in genes, environment, and lifestyle. However, it does not mean the creation of drugs or medical devices that are unique to a patient, but rather the ability to classify individuals into subpopulations that dier in their susceptibility to a particular disease, in the genetic factors of a diseases, or in their response to a specic treatment.

4 Motivation: subgroup analysis Subgroup analysis: subgrouping (clustering) with respect to how a clinical outcome is related to individual characteristics, including possibly unobserved ones. Estimation of subgroup specic treatment eects: subgrouping (clustering) with respect to heterogeneous treatment eects. Estimation of treatment assignment rules: this may need to take into account heterogeneity in the target patient population.

5 A simulated example Example 1. Consider a regression model with heterogeneous treatment eects : y i = z T i η + x i β i +ε i, i = 1,..., n, (1) where z i IR 5. We randomly assign the treatment coecients to two groups with equal probabilities, so that β i = 2 for i G 1 and β i = 2 for i G 2. Consider the two approaches: Least squares regression without taking into account heterogeneity. The proposed method.

6 Example Figure 1 : Simulated example, the two solid black lines represent y = 2x and y = 2x x

7 Some existing approaches Mixture model analysis (Gaussian mixture model): used widely for data clustering and classication (Baneld and Raftery (1993); Hastie and Tibshirani (1996); McNicholas (2010); Wei and Kosorok (2013), Shen and He (2015)). This approach requires specifying the number of subgroups in the population and a parametric model assumption. Methods of estimating homogeneity eects of covariates (Tibshirani et al. (2005); Bondell and Reich (2008); Shen and Huang (2010); Ke, Fan and Wu (2013), among others). These works consider grouping covariates, not observations.

8 Model and approach We consider the model y i = z T i η + x T i β i + ε i, i = 1,..., n. (2) Heterogeneous treatment eects: let G = (G 1,..., G K ) be a partition of {1,..., n}. Assume β i = α k for all i G k, where α k is the common value for the β i 's from group G k. Goal: estimate K and identify the subgroups; estimate (α 1,..., α K ) and η. Method: a concave pairwise fusion penalized least squares approach. Algorithm: an alternating direction method of multipliers (ADMM, Boyd et al. 2011). Challenge: information of subgroups are unknown (the number of subgroups, which subjects belong to which subgroups, etc.)

9 Subgroup Analysis via Concave Pairwise Fusion Consider the concave pairwise fusion penalized least squares criterion Q n (η, β; λ) = 1 2 n (y i z T i η x T i β i ) 2 + i=1 1 i<j n p( β i β j, λ), (3) where p(, λ) is a penalty function with a tuning parameter λ 0. Let ( η(λ), β(λ)) = argmin Q n (η, β; λ). (4) η IR q, β IR np We compute (ˆη(λ), ˆβ(λ)) for λ [λ min, λ max ], where λ max is the value that forces a constant ˆβ solution and λ min is a small positive number. We are particularly interested in the path {ˆβ(λ) : λ [λ min, λ max ]}.

10 Concave Pairwise Fusion The penalty shrinks some of the pairs β j β k to zero. Based on this, we can partition the sample into subgroups. Let { α 1,..., α K } be the distinct values of β. Let Ĝ k = {i : ˆβ i = α k, 1 i n}, 1 k K. Then {Ĝ1,..., Ĝ K } constitutes a partition of {1,..., n}.

11 Penalty function L 1 penalty: p γ (t, λ) = λt, leads to biased estimates; In our numerical studies, the L 1 penalty tends to either yield a large number of subgroups or no subgroups on the solution path. A penalty which can produce nearly unbiased estimates is more appealing. The SCAD penalty (Fan and Li 2001): p γ (t, λ) = λ t The MCP (Zhang 2010): 0 p γ (t, λ) = λ min{1, (γ x/λ) + /(γ 1)}dx, γ > 2 t 0 (1 x/(γλ)) + dx, γ > 1

12 ADMM Algorithm Introduce a new set of parameters δ ij = β i β j. The minimization of (3) is equivalent to minimizing L 0 (η, β, δ) = 1 2 n where δ = {δ T ij, i < j} T. i=1 (y i z T i η x T i β i ) 2 + i<j p γ( δ ij, λ), subject to β i β j δ ij = 0, (5)

13 ADMM The augmented Lagrangian is L(η, β, δ, υ) = 1 2 n (y i z T i η x T i β i ) 2 + p γ( δ jk, λ) (6) j<k i=1 + j<k υjk, β j β k δ jk + ϑ 2 β j β k δ jk 2. j<k For a given (δ m, υ m ) at step m, the iteration goes as follows: (η m+1, β m+1 ) = argmin L(η, β, δ m, υ m ), (7) η,β δ m+1 = argmin L(η m+1, β m+1, δ, υ m ), (8) δ υ m+1 ij = υ m ij + ϑ(βi m+1 β m+1 j δ m+1 ij ). (9)

14 ADMM Step (7) is a quadratic minimization problem. Step (8) involves minimizing ϑ 2 ζm jk δ jk 2 + p γ ( δ jk, λ) (10) with respect to δ jk, where ζ m jk = βm j β m k + ϑ 1 υ m jk. This is a thresholding operator corresponding to p γ. For the L 1 penalty, δ m+1 jk = S(ζ m jk, λ/ϑ), (11) where S(z, t) = (1 t/ z ) + z is the groupwise soft thresholding operator. Here (x) + = x if x > 0 and = 0, otherwise.

15 ADMM MCP with γ > 1/ϑ, δ m+1 ij = { S(ζ m ij,λ/ϑ) 1 1/(γϑ) ζ ij if ζ m ij γλ, if ζ m ij > γλ. (12) SCAD penalty with γ > 1/ϑ + 1, δ m+1 ij = S(ζ m ij, λ/ϑ) S(ζ m ij,γλ/((γ 1)ϑ)) 1 1/((γ 1)ϑ) ζ m ij if ζ m ij λ + λ/ϑ, if λ + λ/ϑ < ζ m ij γλ, if ζ m ij > γλ. (13)

16 ADMM initial value To start the ADMM algorithm, it is important to nd a reasonable initial value. We consider the ridge fusion criterion given by L R (η, β) = 1 2 Zη Xβ y 2 + λ 2 1 j<k n β j β k 2, where λ is the tuning parameter having a small value. We use λ = in our analysis.

17 ADMM solution path To compute the solution path of η and β along the λ values, we use the warm start and continuation strategy to update the solutions. Let [λ min, λ max ] be the interval on which we compute the solution path. Let λ min = λ 0 < λ 1 < < λ K λ max be a grid of λ values in [λ min, λ max ]. Compute the ridge fusion solution ( η(λ 0 ), ˆβ(λ 0 )) and use it as the initial value. Compute ( η(λ k ), ˆβ(λ k )) using ( η(λ k 1 ), ˆβ(λ k 1 )) as the initial value for k = 1,..., K. Note that we start from the smallest λ value in computing the solution path.

18 Statistical Properties Let W= {w ik } be an n K matrix with w ik = 1 for i G k w ik = 0 otherwise. Let W = W I p. Let and M G = {β IR np : β i = β j, for any i, j G k, 1 k K}. For each β M G, it can be written as β = Wα, where α = (α T 1,..., αt K )T and α k is a p 1 vector of the kth subgroup-specic parameter for k = 1,..., K. Denote the minimum and maximum group sizes by G min = min 1 k K G k and G max = max 1 k K G k, respectively. Let X = XW and U = (Z, XW).

19 Statistical properties If the underlying groups G 1,..., G K were known, the oracle estimator of (η, β) is ( η or, ˆβ or ) = 1 argmin η IR q, β M G 2 y Zη Xβ 2, (14) and correspondingly, the oracle estimators for the common coecient α and the coecients η are ( η or, α or 1 2 ) = argmin y Zη Xα η IR q, α IR Kp 2 = (U T U) 1 U T y. Let α 0 k be the true common coecient vector for group G k, k = 1,..., K and α 0 = ((α 0 k )T, k = 1,..., K) T. Of course, oracle estimators are not real estimators, they are theoretical constructions useful for stating the properties of the proposed estimators.

20 (C1) The noise vector ε = (ε 1,..., ε n ) T has sub-gaussian tails such that P( a T ε > a x) 2 exp( c 1 x 2 ) for any vector a IR n and x > 0, where 0 < c 1 <. (C2) Let ρ(t) = λ 1 p γ (t, λ). Suppose ρ(t) is a symmetric function of t and is non-decreasing and concave on [0, ). Also, ρ(t) is a constant for t aλ for some constant a > 0, and ρ(0) = 0. In addition, ρ (t) exists and is continuous except for a nite number of t and ρ (0+) = 1. (C3) Assume n i=1 z2 ij = n for 1 k q, and n i=1 x2 ij 1{i G k } = G k for 1 j p, λ min (U T U) C 1 G min, sup i x i C 2 p and sup i z i C 3 q for some constants 0 < C1 <, 0 < C 2 < and 0 < C 3 <.

21 Let and b n = φ n = c 1/2 1 C 1 1 q + Kp Gmin 1 n log n. (15) min i G k,j G k,k k β0 i β 0 j = min k k α0 k α 0 k be the minimal dierence of the common values between two groups. Theorem Suppose (C1)-(C3) hold, Kp = o(n), q = o(n), and G min (q + Kp)n log n. If b n > aλ and λ φ n, for some constant a > 0, where φ n is given in (15), then there exists a local minimizer ( η(λ), β(λ)) of the objective function Q n (η, β; λ) given in (3) satisfying ( P ( η(λ), β(λ)) = ( η or, β or ) ) 1.

22 Simulation Studies We use the modied Bayes Information Criterion (BIC) (Schwarz, 1978; Wang, Li and Tsai, 2007) for high-dimensional data settings to select the tuning parameter by minimizing BIC(λ) = log[ n (y i z T i=1 i η(λ) x T i β i (λ)) 2 log n /n]+c n n ( K(λ)p+q), (16) where C n is a positive number which can depend on n. We use C n = log(np + q). We select λ by minimizing the modied BIC.

23 Example 1 Example 1 (One treatment variable). Consider where y i = z T i η + x i β i +ε i, i = 1,..., n, (17) z i = (z i1, z i2,..., z i5 ) T with z i1 = 1 and (z i2,..., z i5 ) T generated from multivariate normal with mean 0, variance 1 and an exchangeable correlation ρ = 0.3, x i is simulated from N(0, 1). ε i are i.i.d. N(0, ). η = (η 1,..., η 5 ) T with η k simulated from Uniform[1, 2] for k = 1,..., 5. We randomly assign the treatment coecients to two groups with equal probabilities, i.e., p(i G 1 ) = p(i G 2 ) = 1/2, so that β i = α 1 for i G 1 and β i = α 2 for i G 2, where α 1 = 2 and α 2 = 2. We consider n = 100, 200.

24 Example 1 β β β Figure 2 : Fusiongram: Solution paths for ( ˆβ 1 (λ),..., ˆβ n (λ)) against λ with n = 200 for data from Example 1. MCP SCAD LASSO λ λ λ

25 Example 1 Table 1 : The sample mean, median and standard deviation (s.d.) of K and the percentage (per) of K equaling the true number of subgroups by MCP and SCAD based on 100 replications with n = 100 and 200 in Example 1. n = 100 n = 200 mean median s.d. per mean median s.d. per MCP SCAD

26 Table 2 : The sample mean, median and asymptotic standard deviation (ASD) of the estimators α 1 and α 2 by MCP and SCAD and oracle estimators α 1 or and αor 2 based on 100 replications with n = 100, 200 in Example 1. n = 100 n = 200 mean median ASD mean median ASD α 1 MCP SCAD α 1 or α 2 MCP SCAD α 2 or

27 Figure 3 : The boxplots of the MSEs of η using MCP and SCAD, respectively, with n = 100 (white) and n = 200 (grey) in Example 1. MSE(η) MCP MCP SCAD SCAD

28 Example 2 (Multiple treatment variables). We simulated data from the heterogeneous model with multiple treatment variables: where y i = z T i η + x T i β i +ε i, i = 1,..., n, (18) z i, ε i and η are simulated in the same way as in Example 1. Let x i = (x i1, x i2, x i3 ) T in which x i1 is simulated from standard normal and (x i2, x i3 ) T are from centered and standardized binomial with probability 0.7 for one outcome. We randomly assign the responses to two groups with equal probabilities, i.e., we let p(i G 1 ) = p(i G 2 ) = 1/2, so that β i = α 1 for i G 1 and β i = α 2 for i G 2, where α 1 = (α 11, α 12, α 13 ) and α 2 = (α 21, α 22, α 23 ). Let α 1j = α and α 2j = α for j = 1, 2, 3. We let α = 1, 2 for dierent signal-noise ratios. Let n = 200.

29 Figure 4 : Fusiongram for (β 11,..., β 1n ), the rst component in β i 's in Example 2. MCP SCAD LASSO β β β λ λ λ

30 Table 3 : The sample mean, median and standard deviation (s.d.) of K and the percentage (per) that K equals to the true number of subgroups by MCP and SCAD based on 100 replications with α = 1, 2 in Example 2. α = 1 α = 2 mean median s.d. per mean median s.d. per MCP SCAD

31 Example 3 (No treatment heterogeneity). We generate data from a model with homogeneous treatment eects y i = z T i η + x i β+ε i, i = 1,..., n, where z i, x i, ε i and η are simulated in the same way as in Example 1. Set β = 2 and n = 200. We use our proposed penalized estimation approach to t the model assuming the possible existence of treatment heterogeneity. The sample mean of the estimated number of groups K is 1.49 and 1.48 based on 100 replications, respectively, for the MCP and SCAD methods. The sample median is 1 for both methods.

32 Table 4 : The empirical bias (Bias) of the estimates of β and η, and the average asymptotic standard deviation (ASD) and the empirical standard deviation (ESD) of MCP and SCAD, as well as the oracle estimator (ORAC) in Example 3. β η 1 η 2 η 3 η 4 η 5 Bias MCP ASE ESE Bias SCAD ASE ESE Bias ORAC ASE ESE

33 ACTG175 data We apply our method to the AIDS Clinical Trials Group Study 175 (ACTG175) (Tsiatis et al., 2007), ACTG175 was a randomized clinical trial to compare the 4 treatments: zidovudine with other three therapies including zidovudine and didanosine, zidovudine and zalcitabine, didanosine in adults infected with the human immunodeciency virus type I. We randomly select 300 patients from the study to consist of our dataset.

34 ACTG175 data The response variable is the log-transformed value of the CD4 counts at 20±5 weeks. We use binary variables for the treatments x i = (x i1, x i2, x i3 ) T. There are 12 baseline covariates in the model, 1 age (years), 2 weight (kg), 3 Karnofsky score, 4 CD4 counts at baseline, 5 CD8 counts at baseline, 6 hemophilia (0 =no, 1 =yes), 7 homosexual activity (0 =no, 1 =yes), 8 history of intravenous drug use (0 =no, 1 =yes), 9 race (0 =white, 1 =not white), 10 gender (0 =female, 1 =male), 11 antiretroviral history (0 =naive, 1 =experienced) and 12 symptomatic status (0 =asymptomatic, 1 =symptomatic).

35 ACTG175 data We t the heterogeneous model y i = z T i η + x T i β i +ε i, i = 1,..., 300, where z i = (1, z i2..., z i13 ) T with the rst component for intercept and other components being the 12 covariates described above. All the predictors are centered and standardized. We identied two subgroups.

36 ACTG175 data Figure 5 : Fusiongrams for β 1 = (β 11,..., β 1n ), β 2 = (β 21,..., β 2n ), and β 3 = (β 31,..., β 3n ). MCP MCP MCP β β β λ λ λ

37 Table 5 : The estimates (Est.), standard deviations (s.d.) and p-values (P-value) of α 1 and α 2 by the MCP and SCAD methods, and those values of β = α 1 by the OLS method. α 11 α 12 α 13 α 21 α 22 α 23 MCP Est s.d p-value SCAD Est s.d p-value OLS Est s.d p-value <

38 Concluding remarks Extension to other important models (e.g., logistic regression, Cox regression) is conceptually straightforward, but theoretical analysis and computation are more dicult. Extension to p n models is also possible, but requires further sparsity assumption to ensure model identiability, and theoretical analysis is more dicult. It is of interest to speed up the ADMM so that it can handle large n problems. It is possible to weaken the conditions for the theoretical results, but this will not change the basic story. The theoretical results are derived for xed λ values. It is much more dicult to derive the results for λ values selected based on a data-driven procedure.

39 Thank you!

A concave pairwise fusion approach to subgroup analysis

A concave pairwise fusion approach to subgroup analysis Journal of the American Statistical Association ISSN: 0162-1459 (Print) 1537-274X (Online) Journal homepage: http://www.tandfonline.com/loi/uasa20 A concave pairwise fusion approach to subgroup analysis

More information

The MNet Estimator. Patrick Breheny. Department of Biostatistics Department of Statistics University of Kentucky. August 2, 2010

The MNet Estimator. Patrick Breheny. Department of Biostatistics Department of Statistics University of Kentucky. August 2, 2010 Department of Biostatistics Department of Statistics University of Kentucky August 2, 2010 Joint work with Jian Huang, Shuangge Ma, and Cun-Hui Zhang Penalized regression methods Penalized methods have

More information

Homogeneity Pursuit. Jianqing Fan

Homogeneity Pursuit. Jianqing Fan Jianqing Fan Princeton University with Tracy Ke and Yichao Wu http://www.princeton.edu/ jqfan June 5, 2014 Get my own profile - Help Amazing Follow this author Grace Wahba 9 Followers Follow new articles

More information

Consistent high-dimensional Bayesian variable selection via penalized credible regions

Consistent high-dimensional Bayesian variable selection via penalized credible regions Consistent high-dimensional Bayesian variable selection via penalized credible regions Howard Bondell bondell@stat.ncsu.edu Joint work with Brian Reich Howard Bondell p. 1 Outline High-Dimensional Variable

More information

Bi-level feature selection with applications to genetic association

Bi-level feature selection with applications to genetic association Bi-level feature selection with applications to genetic association studies October 15, 2008 Motivation In many applications, biological features possess a grouping structure Categorical variables may

More information

Semi-Penalized Inference with Direct FDR Control

Semi-Penalized Inference with Direct FDR Control Jian Huang University of Iowa April 4, 2016 The problem Consider the linear regression model y = p x jβ j + ε, (1) j=1 where y IR n, x j IR n, ε IR n, and β j is the jth regression coefficient, Here p

More information

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson Bayesian variable selection via penalized credible regions Brian Reich, NC State Joint work with Howard Bondell and Ander Wilson Brian Reich, NCSU Penalized credible regions 1 Motivation big p, small n

More information

MSA220/MVE440 Statistical Learning for Big Data

MSA220/MVE440 Statistical Learning for Big Data MSA220/MVE440 Statistical Learning for Big Data Lecture 9-10 - High-dimensional regression Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology Recap from

More information

Personalized Treatment Selection Based on Randomized Clinical Trials. Tianxi Cai Department of Biostatistics Harvard School of Public Health

Personalized Treatment Selection Based on Randomized Clinical Trials. Tianxi Cai Department of Biostatistics Harvard School of Public Health Personalized Treatment Selection Based on Randomized Clinical Trials Tianxi Cai Department of Biostatistics Harvard School of Public Health Outline Motivation A systematic approach to separating subpopulations

More information

Stability and the elastic net

Stability and the elastic net Stability and the elastic net Patrick Breheny March 28 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/32 Introduction Elastic Net Our last several lectures have concentrated on methods for

More information

Smoothly Clipped Absolute Deviation (SCAD) for Correlated Variables

Smoothly Clipped Absolute Deviation (SCAD) for Correlated Variables Smoothly Clipped Absolute Deviation (SCAD) for Correlated Variables LIB-MA, FSSM Cadi Ayyad University (Morocco) COMPSTAT 2010 Paris, August 22-27, 2010 Motivations Fan and Li (2001), Zou and Li (2008)

More information

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider the problem of

More information

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1 Variable Selection in Restricted Linear Regression Models Y. Tuaç 1 and O. Arslan 1 Ankara University, Faculty of Science, Department of Statistics, 06100 Ankara/Turkey ytuac@ankara.edu.tr, oarslan@ankara.edu.tr

More information

Lecture 14: Variable Selection - Beyond LASSO

Lecture 14: Variable Selection - Beyond LASSO Fall, 2017 Extension of LASSO To achieve oracle properties, L q penalty with 0 < q < 1, SCAD penalty (Fan and Li 2001; Zhang et al. 2007). Adaptive LASSO (Zou 2006; Zhang and Lu 2007; Wang et al. 2007)

More information

High-dimensional Ordinary Least-squares Projection for Screening Variables

High-dimensional Ordinary Least-squares Projection for Screening Variables 1 / 38 High-dimensional Ordinary Least-squares Projection for Screening Variables Chenlei Leng Joint with Xiangyu Wang (Duke) Conference on Nonparametric Statistics for Big Data and Celebration to Honor

More information

Grouping Pursuit in regression. Xiaotong Shen

Grouping Pursuit in regression. Xiaotong Shen Grouping Pursuit in regression Xiaotong Shen School of Statistics University of Minnesota Email xshen@stat.umn.edu Joint with Hsin-Cheng Huang (Sinica, Taiwan) Workshop in honor of John Hartigan Innovation

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

Generalized Elastic Net Regression

Generalized Elastic Net Regression Abstract Generalized Elastic Net Regression Geoffroy MOURET Jean-Jules BRAULT Vahid PARTOVINIA This work presents a variation of the elastic net penalization method. We propose applying a combined l 1

More information

Lecture 5: Soft-Thresholding and Lasso

Lecture 5: Soft-Thresholding and Lasso High Dimensional Data and Statistical Learning Lecture 5: Soft-Thresholding and Lasso Weixing Song Department of Statistics Kansas State University Weixing Song STAT 905 October 23, 2014 1/54 Outline Penalized

More information

Machine Learning Linear Classification. Prof. Matteo Matteucci

Machine Learning Linear Classification. Prof. Matteo Matteucci Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)

More information

Simultaneous variable selection and class fusion for high-dimensional linear discriminant analysis

Simultaneous variable selection and class fusion for high-dimensional linear discriminant analysis Biostatistics (2010), 11, 4, pp. 599 608 doi:10.1093/biostatistics/kxq023 Advance Access publication on May 26, 2010 Simultaneous variable selection and class fusion for high-dimensional linear discriminant

More information

Sparse Linear Models (10/7/13)

Sparse Linear Models (10/7/13) STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine

More information

Regularization and Variable Selection via the Elastic Net

Regularization and Variable Selection via the Elastic Net p. 1/1 Regularization and Variable Selection via the Elastic Net Hui Zou and Trevor Hastie Journal of Royal Statistical Society, B, 2005 Presenter: Minhua Chen, Nov. 07, 2008 p. 2/1 Agenda Introduction

More information

Iterative Selection Using Orthogonal Regression Techniques

Iterative Selection Using Orthogonal Regression Techniques Iterative Selection Using Orthogonal Regression Techniques Bradley Turnbull 1, Subhashis Ghosal 1 and Hao Helen Zhang 2 1 Department of Statistics, North Carolina State University, Raleigh, NC, USA 2 Department

More information

Sparse Gaussian conditional random fields

Sparse Gaussian conditional random fields Sparse Gaussian conditional random fields Matt Wytock, J. ico Kolter School of Computer Science Carnegie Mellon University Pittsburgh, PA 53 {mwytock, zkolter}@cs.cmu.edu Abstract We propose sparse Gaussian

More information

Univariate shrinkage in the Cox model for high dimensional data

Univariate shrinkage in the Cox model for high dimensional data Univariate shrinkage in the Cox model for high dimensional data Robert Tibshirani January 6, 2009 Abstract We propose a method for prediction in Cox s proportional model, when the number of features (regressors)

More information

The picasso Package for Nonconvex Regularized M-estimation in High Dimensions in R

The picasso Package for Nonconvex Regularized M-estimation in High Dimensions in R The picasso Package for Nonconvex Regularized M-estimation in High Dimensions in R Xingguo Li Tuo Zhao Tong Zhang Han Liu Abstract We describe an R package named picasso, which implements a unified framework

More information

Analysis Methods for Supersaturated Design: Some Comparisons

Analysis Methods for Supersaturated Design: Some Comparisons Journal of Data Science 1(2003), 249-260 Analysis Methods for Supersaturated Design: Some Comparisons Runze Li 1 and Dennis K. J. Lin 2 The Pennsylvania State University Abstract: Supersaturated designs

More information

A Modern Look at Classical Multivariate Techniques

A Modern Look at Classical Multivariate Techniques A Modern Look at Classical Multivariate Techniques Yoonkyung Lee Department of Statistics The Ohio State University March 16-20, 2015 The 13th School of Probability and Statistics CIMAT, Guanajuato, Mexico

More information

A Significance Test for the Lasso

A Significance Test for the Lasso A Significance Test for the Lasso Lockhart R, Taylor J, Tibshirani R, and Tibshirani R Ashley Petersen May 14, 2013 1 Last time Problem: Many clinical covariates which are important to a certain medical

More information

Unsupervised Regressive Learning in High-dimensional Space

Unsupervised Regressive Learning in High-dimensional Space Unsupervised Regressive Learning in High-dimensional Space University of Kent ATRC Leicester 31st July, 2018 Outline Data Linkage Analysis High dimensionality and variable screening Variable screening

More information

Paper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001)

Paper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001) Paper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001) Presented by Yang Zhao March 5, 2010 1 / 36 Outlines 2 / 36 Motivation

More information

Ensemble estimation and variable selection with semiparametric regression models

Ensemble estimation and variable selection with semiparametric regression models Ensemble estimation and variable selection with semiparametric regression models Sunyoung Shin Department of Mathematical Sciences University of Texas at Dallas Joint work with Jason Fine, Yufeng Liu,

More information

Chapter 3. Linear Models for Regression

Chapter 3. Linear Models for Regression Chapter 3. Linear Models for Regression Wei Pan Division of Biostatistics, School of Public Health, University of Minnesota, Minneapolis, MN 55455 Email: weip@biostat.umn.edu PubH 7475/8475 c Wei Pan Linear

More information

VARIABLE SELECTION AND ESTIMATION WITH THE SEAMLESS-L 0 PENALTY

VARIABLE SELECTION AND ESTIMATION WITH THE SEAMLESS-L 0 PENALTY Statistica Sinica 23 (2013), 929-962 doi:http://dx.doi.org/10.5705/ss.2011.074 VARIABLE SELECTION AND ESTIMATION WITH THE SEAMLESS-L 0 PENALTY Lee Dicker, Baosheng Huang and Xihong Lin Rutgers University,

More information

Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion

Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Glenn Heller and Jing Qin Department of Epidemiology and Biostatistics Memorial

More information

Comparisons of penalized least squares. methods by simulations

Comparisons of penalized least squares. methods by simulations Comparisons of penalized least squares arxiv:1405.1796v1 [stat.co] 8 May 2014 methods by simulations Ke ZHANG, Fan YIN University of Science and Technology of China, Hefei 230026, China Shifeng XIONG Academy

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

arxiv: v1 [stat.me] 30 Dec 2017

arxiv: v1 [stat.me] 30 Dec 2017 arxiv:1801.00105v1 [stat.me] 30 Dec 2017 An ISIS screening approach involving threshold/partition for variable selection in linear regression 1. Introduction Yu-Hsiang Cheng e-mail: 96354501@nccu.edu.tw

More information

SOLVING NON-CONVEX LASSO TYPE PROBLEMS WITH DC PROGRAMMING. Gilles Gasso, Alain Rakotomamonjy and Stéphane Canu

SOLVING NON-CONVEX LASSO TYPE PROBLEMS WITH DC PROGRAMMING. Gilles Gasso, Alain Rakotomamonjy and Stéphane Canu SOLVING NON-CONVEX LASSO TYPE PROBLEMS WITH DC PROGRAMMING Gilles Gasso, Alain Rakotomamonjy and Stéphane Canu LITIS - EA 48 - INSA/Universite de Rouen Avenue de l Université - 768 Saint-Etienne du Rouvray

More information

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression Noah Simon Jerome Friedman Trevor Hastie November 5, 013 Abstract In this paper we purpose a blockwise descent

More information

The OSCAR for Generalized Linear Models

The OSCAR for Generalized Linear Models Sebastian Petry & Gerhard Tutz The OSCAR for Generalized Linear Models Technical Report Number 112, 2011 Department of Statistics University of Munich http://www.stat.uni-muenchen.de The OSCAR for Generalized

More information

A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models

A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models Jingyi Jessica Li Department of Statistics University of California, Los

More information

Lecture 14: Shrinkage

Lecture 14: Shrinkage Lecture 14: Shrinkage Reading: Section 6.2 STATS 202: Data mining and analysis October 27, 2017 1 / 19 Shrinkage methods The idea is to perform a linear regression, while regularizing or shrinking the

More information

Stepwise Searching for Feature Variables in High-Dimensional Linear Regression

Stepwise Searching for Feature Variables in High-Dimensional Linear Regression Stepwise Searching for Feature Variables in High-Dimensional Linear Regression Qiwei Yao Department of Statistics, London School of Economics q.yao@lse.ac.uk Joint work with: Hongzhi An, Chinese Academy

More information

Exam: high-dimensional data analysis January 20, 2014

Exam: high-dimensional data analysis January 20, 2014 Exam: high-dimensional data analysis January 20, 204 Instructions: - Write clearly. Scribbles will not be deciphered. - Answer each main question not the subquestions on a separate piece of paper. - Finish

More information

RECENT ADVANCES IN STATISTICAL MODELS: TOPICS IN MODEL SELECTION AND SEMI-PARAMETRIC INFERENCE

RECENT ADVANCES IN STATISTICAL MODELS: TOPICS IN MODEL SELECTION AND SEMI-PARAMETRIC INFERENCE RECENT ADVANCES IN STATISTICAL MODELS: TOPICS IN MODEL SELECTION AND SEMI-PARAMETRIC INFERENCE BY WENQIAN QIAO A dissertation submitted to the Graduate School New Brunswick Rutgers, The State University

More information

Machine Learning for OR & FE

Machine Learning for OR & FE Machine Learning for OR & FE Regression II: Regularization and Shrinkage Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

Group exponential penalties for bi-level variable selection

Group exponential penalties for bi-level variable selection for bi-level variable selection Department of Biostatistics Department of Statistics University of Kentucky July 31, 2011 Introduction In regression, variables can often be thought of as grouped: Indicator

More information

ISyE 691 Data mining and analytics

ISyE 691 Data mining and analytics ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)

More information

A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables

A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables Qi Tang (Joint work with Kam-Wah Tsui and Sijian Wang) Department of Statistics University of Wisconsin-Madison Feb. 8,

More information

arxiv: v2 [stat.me] 4 Jun 2016

arxiv: v2 [stat.me] 4 Jun 2016 Variable Selection for Additive Partial Linear Quantile Regression with Missing Covariates 1 Variable Selection for Additive Partial Linear Quantile Regression with Missing Covariates Ben Sherwood arxiv:1510.00094v2

More information

Nonconcave Penalized Likelihood with A Diverging Number of Parameters

Nonconcave Penalized Likelihood with A Diverging Number of Parameters Nonconcave Penalized Likelihood with A Diverging Number of Parameters Jianqing Fan and Heng Peng Presenter: Jiale Xu March 12, 2010 Jianqing Fan and Heng Peng Presenter: JialeNonconcave Xu () Penalized

More information

Semiparametric Mixed Effects Models with Flexible Random Effects Distribution

Semiparametric Mixed Effects Models with Flexible Random Effects Distribution Semiparametric Mixed Effects Models with Flexible Random Effects Distribution Marie Davidian North Carolina State University davidian@stat.ncsu.edu www.stat.ncsu.edu/ davidian Joint work with A. Tsiatis,

More information

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015 MS&E 226 In-Class Midterm Examination Solutions Small Data October 20, 2015 PROBLEM 1. Alice uses ordinary least squares to fit a linear regression model on a dataset containing outcome data Y and covariates

More information

Institute of Statistics Mimeo Series No Simultaneous regression shrinkage, variable selection and clustering of predictors with OSCAR

Institute of Statistics Mimeo Series No Simultaneous regression shrinkage, variable selection and clustering of predictors with OSCAR DEPARTMENT OF STATISTICS North Carolina State University 2501 Founders Drive, Campus Box 8203 Raleigh, NC 27695-8203 Institute of Statistics Mimeo Series No. 2583 Simultaneous regression shrinkage, variable

More information

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING DATE AND TIME: August 30, 2018, 14.00 19.00 RESPONSIBLE TEACHER: Niklas Wahlström NUMBER OF PROBLEMS: 5 AIDING MATERIAL: Calculator, mathematical

More information

Group descent algorithms for nonconvex penalized linear and logistic regression models with grouped predictors

Group descent algorithms for nonconvex penalized linear and logistic regression models with grouped predictors Group descent algorithms for nonconvex penalized linear and logistic regression models with grouped predictors Patrick Breheny Department of Biostatistics University of Iowa Jian Huang Department of Statistics

More information

The Iterated Lasso for High-Dimensional Logistic Regression

The Iterated Lasso for High-Dimensional Logistic Regression The Iterated Lasso for High-Dimensional Logistic Regression By JIAN HUANG Department of Statistics and Actuarial Science, 241 SH University of Iowa, Iowa City, Iowa 52242, U.S.A. SHUANGE MA Division of

More information

ORIE 4741: Learning with Big Messy Data. Regularization

ORIE 4741: Learning with Big Messy Data. Regularization ORIE 4741: Learning with Big Messy Data Regularization Professor Udell Operations Research and Information Engineering Cornell October 26, 2017 1 / 24 Regularized empirical risk minimization choose model

More information

Divide-and-combine Strategies in Statistical Modeling for Massive Data

Divide-and-combine Strategies in Statistical Modeling for Massive Data Divide-and-combine Strategies in Statistical Modeling for Massive Data Liqun Yu Washington University in St. Louis March 30, 2017 Liqun Yu (WUSTL) D&C Statistical Modeling for Massive Data March 30, 2017

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR

More information

Classification. Chapter Introduction. 6.2 The Bayes classifier

Classification. Chapter Introduction. 6.2 The Bayes classifier Chapter 6 Classification 6.1 Introduction Often encountered in applications is the situation where the response variable Y takes values in a finite set of labels. For example, the response Y could encode

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA

The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA Presented by Dongjun Chung March 12, 2010 Introduction Definition Oracle Properties Computations Relationship: Nonnegative Garrote Extensions:

More information

9/26/17. Ridge regression. What our model needs to do. Ridge Regression: L2 penalty. Ridge coefficients. Ridge coefficients

9/26/17. Ridge regression. What our model needs to do. Ridge Regression: L2 penalty. Ridge coefficients. Ridge coefficients What our model needs to do regression Usually, we are not just trying to explain observed data We want to uncover meaningful trends And predict future observations Our questions then are Is β" a good estimate

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Matrix Data: Prediction Instructor: Yizhou Sun yzsun@ccs.neu.edu September 14, 2014 Today s Schedule Course Project Introduction Linear Regression Model Decision Tree 2 Methods

More information

Regression Shrinkage and Selection via the Lasso

Regression Shrinkage and Selection via the Lasso Regression Shrinkage and Selection via the Lasso ROBERT TIBSHIRANI, 1996 Presenter: Guiyun Feng April 27 () 1 / 20 Motivation Estimation in Linear Models: y = β T x + ɛ. data (x i, y i ), i = 1, 2,...,

More information

Variable Selection for Highly Correlated Predictors

Variable Selection for Highly Correlated Predictors Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu arxiv:1709.04840v1 [stat.me] 14 Sep 2017 Abstract Penalty-based variable selection methods are powerful in selecting relevant covariates

More information

CMSC858P Supervised Learning Methods

CMSC858P Supervised Learning Methods CMSC858P Supervised Learning Methods Hector Corrada Bravo March, 2010 Introduction Today we discuss the classification setting in detail. Our setting is that we observe for each subject i a set of p predictors

More information

Penalized methods in genome-wide association studies

Penalized methods in genome-wide association studies University of Iowa Iowa Research Online Theses and Dissertations Summer 2011 Penalized methods in genome-wide association studies Jin Liu University of Iowa Copyright 2011 Jin Liu This dissertation is

More information

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS023) p.3938 An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Vitara Pungpapong

More information

Simultaneous regression shrinkage, variable selection, and supervised clustering of predictors with OSCAR

Simultaneous regression shrinkage, variable selection, and supervised clustering of predictors with OSCAR Simultaneous regression shrinkage, variable selection, and supervised clustering of predictors with OSCAR Howard D. Bondell and Brian J. Reich Department of Statistics, North Carolina State University,

More information

A UNIFIED APPROACH TO MODEL SELECTION AND SPARS. REGULARIZED LEAST SQUARES by Jinchi Lv and Yingying Fan The annals of Statistics (2009)

A UNIFIED APPROACH TO MODEL SELECTION AND SPARS. REGULARIZED LEAST SQUARES by Jinchi Lv and Yingying Fan The annals of Statistics (2009) A UNIFIED APPROACH TO MODEL SELECTION AND SPARSE RECOVERY USING REGULARIZED LEAST SQUARES by Jinchi Lv and Yingying Fan The annals of Statistics (2009) Mar. 19. 2010 Outline 1 2 Sideline information Notations

More information

Linear regression methods

Linear regression methods Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response

More information

TGDR: An Introduction

TGDR: An Introduction TGDR: An Introduction Julian Wolfson Student Seminar March 28, 2007 1 Variable Selection 2 Penalization, Solution Paths and TGDR 3 Applying TGDR 4 Extensions 5 Final Thoughts Some motivating examples We

More information

Or How to select variables Using Bayesian LASSO

Or How to select variables Using Bayesian LASSO Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO On Bayesian Variable Selection

More information

Variable Selection for Highly Correlated Predictors

Variable Selection for Highly Correlated Predictors Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu Department of Statistics, University of Illinois at Urbana-Champaign WHOA-PSI, Aug, 2017 St. Louis, Missouri 1 / 30 Background Variable

More information

Bayesian Grouped Horseshoe Regression with Application to Additive Models

Bayesian Grouped Horseshoe Regression with Application to Additive Models Bayesian Grouped Horseshoe Regression with Application to Additive Models Zemei Xu, Daniel F. Schmidt, Enes Makalic, Guoqi Qian, and John L. Hopper Centre for Epidemiology and Biostatistics, Melbourne

More information

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology

More information

CONTRIBUTIONS TO PENALIZED ESTIMATION. Sunyoung Shin

CONTRIBUTIONS TO PENALIZED ESTIMATION. Sunyoung Shin CONTRIBUTIONS TO PENALIZED ESTIMATION Sunyoung Shin A dissertation submitted to the faculty of the University of North Carolina at Chapel Hill in partial fulfillment of the requirements for the degree

More information

Supplement to Bayesian inference for high-dimensional linear regression under the mnet priors

Supplement to Bayesian inference for high-dimensional linear regression under the mnet priors The Canadian Journal of Statistics Vol. xx No. yy 0?? Pages?? La revue canadienne de statistique Supplement to Bayesian inference for high-dimensional linear regression under the mnet priors Aixin Tan

More information

Spatial Lasso with Application to GIS Model Selection. F. Jay Breidt Colorado State University

Spatial Lasso with Application to GIS Model Selection. F. Jay Breidt Colorado State University Spatial Lasso with Application to GIS Model Selection F. Jay Breidt Colorado State University with Hsin-Cheng Huang, Nan-Jung Hsu, and Dave Theobald September 25 The work reported here was developed under

More information

Variable selection in high-dimensional regression problems

Variable selection in high-dimensional regression problems Variable selection in high-dimensional regression problems Jan Polish Academy of Sciences and Warsaw University of Technology Based on joint research with P. Pokarowski, A. Prochenka, P. Teisseyre and

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

Lecture 6: Methods for high-dimensional problems

Lecture 6: Methods for high-dimensional problems Lecture 6: Methods for high-dimensional problems Hector Corrada Bravo and Rafael A. Irizarry March, 2010 In this Section we will discuss methods where data lies on high-dimensional spaces. In particular,

More information

A Significance Test for the Lasso

A Significance Test for the Lasso A Significance Test for the Lasso Lockhart R, Taylor J, Tibshirani R, and Tibshirani R Ashley Petersen June 6, 2013 1 Motivation Problem: Many clinical covariates which are important to a certain medical

More information

Improving Efficiency of Inferences in Randomized Clinical Trials Using Auxiliary Covariates

Improving Efficiency of Inferences in Randomized Clinical Trials Using Auxiliary Covariates Improving Efficiency of Inferences in Randomized Clinical Trials Using Auxiliary Covariates Anastasios (Butch) Tsiatis Department of Statistics North Carolina State University http://www.stat.ncsu.edu/

More information

Individualized Treatment Effects with Censored Data via Nonparametric Accelerated Failure Time Models

Individualized Treatment Effects with Censored Data via Nonparametric Accelerated Failure Time Models Individualized Treatment Effects with Censored Data via Nonparametric Accelerated Failure Time Models Nicholas C. Henderson Thomas A. Louis Gary Rosner Ravi Varadhan Johns Hopkins University July 31, 2018

More information

A Confidence Region Approach to Tuning for Variable Selection

A Confidence Region Approach to Tuning for Variable Selection A Confidence Region Approach to Tuning for Variable Selection Funda Gunes and Howard D. Bondell Department of Statistics North Carolina State University Abstract We develop an approach to tuning of penalized

More information

Robust Variable Selection Through MAVE

Robust Variable Selection Through MAVE Robust Variable Selection Through MAVE Weixin Yao and Qin Wang Abstract Dimension reduction and variable selection play important roles in high dimensional data analysis. Wang and Yin (2008) proposed sparse

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Sparse Recovery using L1 minimization - algorithms Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations

A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations Joint work with Karim Oualkacha (UQÀM), Yi Yang (McGill), Celia Greenwood

More information

sparse and low-rank tensor recovery Cubic-Sketching

sparse and low-rank tensor recovery Cubic-Sketching Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru

More information

ASYMPTOTIC PROPERTIES OF BRIDGE ESTIMATORS IN SPARSE HIGH-DIMENSIONAL REGRESSION MODELS

ASYMPTOTIC PROPERTIES OF BRIDGE ESTIMATORS IN SPARSE HIGH-DIMENSIONAL REGRESSION MODELS ASYMPTOTIC PROPERTIES OF BRIDGE ESTIMATORS IN SPARSE HIGH-DIMENSIONAL REGRESSION MODELS Jian Huang 1, Joel L. Horowitz 2, and Shuangge Ma 3 1 Department of Statistics and Actuarial Science, University

More information

Chris Fraley and Daniel Percival. August 22, 2008, revised May 14, 2010

Chris Fraley and Daniel Percival. August 22, 2008, revised May 14, 2010 Model-Averaged l 1 Regularization using Markov Chain Monte Carlo Model Composition Technical Report No. 541 Department of Statistics, University of Washington Chris Fraley and Daniel Percival August 22,

More information

Statistical Data Mining and Machine Learning Hilary Term 2016

Statistical Data Mining and Machine Learning Hilary Term 2016 Statistical Data Mining and Machine Learning Hilary Term 2016 Dino Sejdinovic Department of Statistics Oxford Slides and other materials available at: http://www.stats.ox.ac.uk/~sejdinov/sdmml Naïve Bayes

More information

High-dimensional regression

High-dimensional regression High-dimensional regression Advanced Methods for Data Analysis 36-402/36-608) Spring 2014 1 Back to linear regression 1.1 Shortcomings Suppose that we are given outcome measurements y 1,... y n R, and

More information

arxiv: v1 [stat.ap] 14 Apr 2011

arxiv: v1 [stat.ap] 14 Apr 2011 The Annals of Applied Statistics 2011, Vol. 5, No. 1, 232 253 DOI: 10.1214/10-AOAS388 c Institute of Mathematical Statistics, 2011 arxiv:1104.2748v1 [stat.ap] 14 Apr 2011 COORDINATE DESCENT ALGORITHMS

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Matrix Data: Prediction Instructor: Yizhou Sun yzsun@ccs.neu.edu September 21, 2015 Announcements TA Monisha s office hour has changed to Thursdays 10-12pm, 462WVH (the same

More information