A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations

Size: px
Start display at page:

Download "A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations"

Transcription

1 A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations Joint work with Karim Oualkacha (UQÀM), Yi Yang (McGill), Celia Greenwood (McGill) JSM, Vancouver 2018 sahirbhatnagar.com/ggmix 1 / 28

2 Motivation Genetic Analysis Workshop (GAW20, March 4-7, 2017, San Diego, US) 1 GOLDEN project: Genetics of Lipid Lowering Drugs and Diet Network Study 2 / 28

3 Motivation Our contribution in GAW20 3 / 28

4 Motivation Our contribution in GAW20 Contribution of the Causal modelling group 3 / 28

5 Motivation Our contribution in GAW20 consisted of investigation of causal relationship between DNA methylation (exposure) within some genes and HDL (outcome) 4 / 28

6 Motivation Our contribution in GAW20 consisted of investigation of causal relationship between DNA methylation (exposure) within some genes and HDL (outcome) DNA methylation in these genes has been shown association with HDL 4 / 28

7 Motivation Our contribution in GAW20 consisted of investigation of causal relationship between DNA methylation (exposure) within some genes and HDL (outcome) DNA methylation in these genes has been shown association with HDL We used Mendelian randomization to explore causal relationship We used SNPs around the analyzed genes as Instrumental Variables (IVs) to interrogate the causal relationship 4 / 28

8 Motivation Our contribution in GAW20 consisted of investigation of causal relationship between DNA methylation (exposure) within some genes and HDL (outcome) DNA methylation in these genes has been shown association with HDL We used Mendelian randomization to explore causal relationship We used SNPs around the analyzed genes as Instrumental Variables (IVs) to interrogate the causal relationship 4 / 28

9 Challenges in GAW20 Data Sets GAW20 SNPs data was high-dimensional There was a need for data regularization in order to select SNPs strongly associated with the exposure Penalized LS regression can be used (Lasso, SCAD, MCP or Elastic net) 5 / 28

10 Challenges in GAW20 Data Sets GAW20 SNPs data was high-dimensional There was a need for data regularization in order to select SNPs strongly associated with the exposure Penalized LS regression can be used (Lasso, SCAD, MCP or Elastic net) But, data consists of families! In the GAW20, all regularized methods either did not control for the family structure 5 / 28

11 Challenges in GAW20 Data Sets GAW20 SNPs data was high-dimensional There was a need for data regularization in order to select SNPs strongly associated with the exposure Penalized LS regression can be used (Lasso, SCAD, MCP or Elastic net) But, data consists of families! In the GAW20, all regularized methods either did not control for the family structure or used two-steps adjustment for the family structure (including our group) 5 / 28

12 Challenges in GAW20 Data Sets Two-steps adjustment: Step 1 : uses LMM to adjust for subjects relationship 1 Oualkacha et al. Gene. Epi. (2013) 6 / 28

13 Challenges in GAW20 Data Sets Two-steps adjustment: Step 1 : uses LMM to adjust for subjects relationship Step 2 : uses residuals from Step 1 in variable-selection LS-regression methods to select SNPs 1 Oualkacha et al. Gene. Epi. (2013) 6 / 28

14 Challenges in GAW20 Data Sets Two-steps adjustment: Step 1 : uses LMM to adjust for subjects relationship Step 2 : uses residuals from Step 1 in variable-selection LS-regression methods to select SNPs Two-steps procedure is a valid approach In association testing, (GRAMMAR) it is known to suffer from huge power loss 1 1 Oualkacha et al. Gene. Epi. (2013) 6 / 28

15 Proposal Aim: We believe that performing variable selection and controlling for familial and/or hidden relationships simultaneously in high-dimensional settings, are likely to be of great interest to the genetic scientists community 7 / 28

16 Proposal Aim: We believe that performing variable selection and controlling for familial and/or hidden relationships simultaneously in high-dimensional settings, are likely to be of great interest to the genetic scientists community Proposal: We propose, ggmix, a two-in-one procedure which controls for structured populations and performs variable selection in Linear Mixed Models 7 / 28

17 Data and Model Phenotype: Y = (y 1,..., y n ) R n SNPs: X = (X 1 ;..., X n ) T R n p, where p n Twice the Kinship matrix or Realized Relationship matrix: Φ R n n Regression Coefficients: β = (β 1,..., β p ) T R p Polygenic random effect: P = (P 1,..., P n ) R n Error: ε = (ε 1,..., ε n ) R n 8 / 28

18 Data and Model Phenotype: Y = (y 1,..., y n ) R n SNPs: X = (X 1 ;..., X n ) T R n p, where p n Twice the Kinship matrix or Realized Relationship matrix: Φ R n n Regression Coefficients: β = (β 1,..., β p ) T R p Polygenic random effect: P = (P 1,..., P n ) R n Error: ε = (ε 1,..., ε n ) R n We consider the following LMM with a single random effect: Y = Xβ + P + ε P N (0, ησ 2 Φ) ε N (0, (1 η)σ 2 I) σ 2 is the phenotype total variance η [0, 1] is the phenotype heritability (narrow sens) Y (β, η, σ 2 ) N (Xβ, ησ 2 Φ + (1 η)σ 2 I) 8 / 28

19 Likelihood The negative log-likelihood is given by l(θ) n 2 log(σ2 ) log (det(v)) + 1 2σ 2 (Y Xβ)T V 1 (Y Xβ) V = ηφ + (1 η)i 9 / 28

20 Likelihood The negative log-likelihood is given by l(θ) n 2 log(σ2 ) log (det(v)) + 1 2σ 2 (Y Xβ)T V 1 (Y Xβ) V = ηφ + (1 η)i Assume the spectral decomposition of Φ Φ = UDU U is an n n orthogonal matrix and D is an n n diagonal matrix 9 / 28

21 Likelihood The negative log-likelihood is given by l(θ) n 2 log(σ2 ) log (det(v)) + 1 2σ 2 (Y Xβ)T V 1 (Y Xβ) V = ηφ + (1 η)i Assume the spectral decomposition of Φ Φ = UDU U is an n n orthogonal matrix and D is an n n diagonal matrix One can write V = U(ηD + (1 η)i)u = UWU 9 / 28

22 Likelihood The negative log-likelihood is given by l(θ) n 2 log(σ2 ) log (det(v)) + 1 2σ 2 (Y Xβ)T V 1 (Y Xβ) V = ηφ + (1 η)i Assume the spectral decomposition of Φ Φ = UDU U is an n n orthogonal matrix and D is an n n diagonal matrix One can write V = U(ηD + (1 η)i)u = UWU with W = diag (w i ) n i=1, w i = ηd ii + (1 η) 9 / 28

23 Likelihood Projection of Y (and columns of X) into Span(U) leads to a simplified correlation structure for the transformed data: Ỹ = U Y Ỹ (β, η, σ 2 ) N ( Xβ, σ 2 W), with X = U X 10 / 28

24 Likelihood Projection of Y (and columns of X) into Span(U) leads to a simplified correlation structure for the transformed data: Ỹ = U Y Ỹ (β, η, σ 2 ) N ( Xβ, σ 2 W), with X = U X The negative log-likelihood can then be expressed as l(θ) n 2 log(σ2 ) n i=1 log (w i ) + 1 2σ 2 (Ỹ Xβ ) T W 1 ( Ỹ Xβ ) 10 / 28

25 Likelihood Projection of Y (and columns of X) into Span(U) leads to a simplified correlation structure for the transformed data: Ỹ = U Y Ỹ (β, η, σ 2 ) N ( Xβ, σ 2 W), with X = U X The negative log-likelihood can then be expressed as l(θ) n 2 log(σ2 ) n i=1 log (w i ) + 1 2σ 2 (Ỹ Xβ ) T W 1 ( Ỹ Xβ ) For fixed σ 2 and η, solving for β is a weighted least squares problem 10 / 28

26 Penalized Maximum Likelihood Estimator Define the objective function: Q λ (Θ) = l(θ) + λ j p j (β j ) p j ( ) is a penalty term on β 1,..., β p An estimate of the model parameters Θ λ is obtained by Θ λ = arg min Q λ (Θ) Θ 11 / 28

27 Block Relaxation (De Leeuw, 1994) To solve for the optimization problem we use a block relaxation technique Set k 0, initial values for the parameter vector Θ (0) and ɛ; for λ {λ max,..., λ min } do repeat For j = 1,..., p, β (k+1) j ( arg min Q λ β (k) j, η(k), σ 2 (k)) β j ( η (k+1) arg min Q λ β (k+1), η, σ 2 (k)) η (β (k+1), η (k+1), σ 2) σ 2 (k+1) arg min Q λ σ 2 k k + 1 until convergence criterion is satisfied: Θ (k+1) Θ (k) 2 < ɛ; end Algorithm 1: Block Relaxation Algorithm 12 / 28

28 Coordinate Gradient Descent Method We take advantage of smoothness of l(θ) We approximate Q λ (Θ) by a strictly convex quadratic function (using gradient) We use CGD to calculate a descent direction To achieve the descent property for the objective function, we employ further line search 1 Tseng P& Yun S. Math. Program., Ser. B, (2009) 13 / 28

29 Coordinate Gradient Descent Method We take advantage of smoothness of l(θ) We approximate Q λ (Θ) by a strictly convex quadratic function (using gradient) We use CGD to calculate a descent direction To achieve the descent property for the objective function, we employ further line search Theorem [Convergence] 1 : If {Θ (k), k = 0, 1, 2,...} is a sequence of iterates generated by the iteration map of Algorithm 1, then each cluster point (i.e. limit point) of {Θ (k), k = 0, 1, 2,...} is a stationary point of Q λ (Θ) 1 Tseng P& Yun S. Math. Program., Ser. B, (2009) 13 / 28

30 Choice of the tuning parameter We use the BIC: BIC λ = 2l( β, σ 2, η) + c df λ df λ is the number of non-zero elements in β λ plus two 1 Several authors 2 have used this criterion for variable selection in mixed models with c = log n Other authors 3 have proposed c = log(log(n)) log(n) 1 Zou et al. The Annals of Statistics, (2007) 2 Bondell et al. Biometrics (2010) 3 Wang et al. JRSS(Ser. B), (2009) 14 / 28

31 Simulation study We simulate genotypes from the BN-PSD Admixture Model a : percentage of causal SNPs X (test) : n 5000 matrix of SNPs randomly sampled across the genome X (causal) : n (a 5000) matrix of SNPs that are truly associated with the simulated phenotype, X (causal) X (test) β j : effect size for the j th SNP, simulated from a Uniform(0.3, 0.7) for j = 1,..., (a 5000) Y (β, η, σ 2 ) N (X (causal) β, ησ 2 Φ + (1 η)σ 2 I) / 28

32 RRM/Kinship matrix construction X (other) : n 10, 000 matrix of simulated SNPs X (kinship) : matrix of SNPs used to construct the RRM/Kinship matrix Scenario 1: X (kinship) = X (other) No overlap Scenario 2: X (kinship) = [X (other), X (causal) ] 100% overlap In each scenario we considered a = 0, 0.01, η = 0.1, 0.5 and σ 2 = 1 16 / 28

33 Empirical Kinship Matrix Empirical Kinship Matrix with Block Structure Block Structure Individuals Kinship 2nd principal component st principal component (a) Kinship Matrix (b) 1st and 2nd PC 17 / 28

34 Correct Sparsity for Null Model Correct Sparsity Results for the Null Model Based on 200 simulations 10% Heritability 50% Heritability No causal SNPs in Kinship twostep lasso ggmix twostep lasso ggmix Method twostep lasso ggmix η = 10% 18 / 28

35 Correct Sparsity for Model with 1% Causal SNPs Correct Sparsity results for the Model with 1% Causal SNPs Based on 200 simulations 10% Heritability 50% Heritability twostep lasso ggmix twostep lasso ggmix All causal SNPs in Kinship No causal SNPs in Kinship Method twostep lasso ggmix η = 10% 19 / 28

36 True Positive vs. False Positive Rate True Positive Rate vs. False Positive Rate (Mean +/- 1 SD) for the Model with 1% Causal SNPs Based on 200 simulations True positive rate 10% Heritability 50% Heritability lasso twostep ggmix lasso twostep ggmix ggmix ggmix 0.6 lasso twostep lasso twostep All causal SNPs in Kinship No causal SNPs in Kinship False positive rate Method a twostep a lasso a ggmix 20 / 28

37 Heritability Estimates for Null Model Estimated Heritability for the Null Model Based on 200 simulations 10% Heritability 50% Heritability η^ No causal SNPs in Kinship twostep ggmix twostep ggmix horizontal dashed line is the true value 21 / 28

38 Heritability Estimates for Model with 1% Causal SNPs Estimated Heritability for the Model with 1% Causal SNPs Based on 200 simulations η^ % Heritability 50% Heritability All causal SNPs in Kinship No causal SNPs in Kinship twostep ggmix twostep ggmix horizontal dashed line is the true value 22 / 28

39 Error Variance for Null Model Estimated Error Variance for the Null Model Based on 200 simulations 10% Heritability 50% Heritability Error Variance 2 2 No causal SNPs in Kinship 1 1 twostep lasso ggmix twostep lasso ggmix horizontal dashed line is the true value 23 / 28

40 Error Variance for Model with 1% Causal SNPs Estimated Error Variance for the Model with 1% Causal SNPs Based on 200 simulations Error Variance % Heritability 50% Heritability All causal SNPs in Kinship No causal SNPs in Kinship twostep lasso ggmix twostep lasso ggmix horizontal dashed line is the true value 24 / 28

41 Model Error vs. Number Active for Model with 1% Causal SNPs Model Error vs. Number of Active Variable (Mean +/- 1 SD) for Model with 1% Causal SNPs Based on 200 simulations 10% Heritability 50% Heritability Model Error ggmix ggmix lasso lasso twostep twostep ggmix ggmix lasso lasso twostep twostep Number of active variables Method a twostep a lasso a ggmix All causal SNPs in Kinship No causal SNPs in Kinship 25 / 28

42 Discussion/Future work In some situations, prior information of the predictors (e.g. SNPs) groups structure is available Theoretical development of group-lasso in LMM is already done 26 / 28

43 Discussion/Future work In some situations, prior information of the predictors (e.g. SNPs) groups structure is available Theoretical development of group-lasso in LMM is already done In situations where the RRM matrix is of low rank (i.e. n >> # of SNPs used to construct RRM). ex: UK Biobank Computational time of fitting ggmix can be reduced using SVD decomposition of X (kinship) in order to construct Φ and in order to transforme the data Theoretical development of low-rank trick is already done 26 / 28

44 Discussion/Future work Capturing the subjects relationship using random effect requires VCs estimation Random effect modelling leads to a non-convex optimization problem Fixed effects models are good alternatives to random effects models for analysis of Longitudinal/Panel data 1 Capturing familial structure using a penalized FE model could be an interesting avenue to explore 1 Roger Koenker, JMA, (2004) 27 / 28

45 References 1. Christoph Lippert, Jennifer Listgarten, Ying Liu, Carl M Kadie, Robert I Davidson, and David Heckerman. Fast linear mixed models for genome-wide association studies. Nature methods, 8(10): , Matti Pirinen, Peter Donnelly, Chris CA Spencer, et al. Efficient computation with a linear mixed model on large-scale data sets with applications to genetic studies. The Annals of Applied Statistics, 7(1): , Paul Tseng and Sangwoon Yun. A coordinate gradient descent method for nonsmooth separable minimization. Mathematical Programming, 117(1): , Jerome Friedman, Trevor Hastie, and Rob Tibshirani. Regularization paths for gen- eralized linear models via coordinate descent. Journal of statistical software, 33(1):1, sahirbhatnagar.com/ggmix 28 / 28

FaST linear mixed models for genome-wide association studies

FaST linear mixed models for genome-wide association studies Nature Methods FaS linear mixed models for genome-wide association studies Christoph Lippert, Jennifer Listgarten, Ying Liu, Carl M Kadie, Robert I Davidson & David Heckerman Supplementary Figure Supplementary

More information

25 : Graphical induced structured input/output models

25 : Graphical induced structured input/output models 10-708: Probabilistic Graphical Models 10-708, Spring 2013 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Meghana Kshirsagar (mkshirsa), Yiwen Chen (yiwenche) 1 Graph

More information

25 : Graphical induced structured input/output models

25 : Graphical induced structured input/output models 10-708: Probabilistic Graphical Models 10-708, Spring 2016 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Raied Aljadaany, Shi Zong, Chenchen Zhu Disclaimer: A large

More information

Machine Learning for OR & FE

Machine Learning for OR & FE Machine Learning for OR & FE Regression II: Regularization and Shrinkage Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

Regularization and Variable Selection via the Elastic Net

Regularization and Variable Selection via the Elastic Net p. 1/1 Regularization and Variable Selection via the Elastic Net Hui Zou and Trevor Hastie Journal of Royal Statistical Society, B, 2005 Presenter: Minhua Chen, Nov. 07, 2008 p. 2/1 Agenda Introduction

More information

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression Noah Simon Jerome Friedman Trevor Hastie November 5, 013 Abstract In this paper we purpose a blockwise descent

More information

Group exponential penalties for bi-level variable selection

Group exponential penalties for bi-level variable selection for bi-level variable selection Department of Biostatistics Department of Statistics University of Kentucky July 31, 2011 Introduction In regression, variables can often be thought of as grouped: Indicator

More information

Pathwise coordinate optimization

Pathwise coordinate optimization Stanford University 1 Pathwise coordinate optimization Jerome Friedman, Trevor Hastie, Holger Hoefling, Robert Tibshirani Stanford University Acknowledgements: Thanks to Stephen Boyd, Michael Saunders,

More information

SUPPLEMENTARY TEXT: EFFICIENT COMPUTATION WITH A LINEAR MIXED MODEL ON LARGE-SCALE DATA SETS WITH APPLICATIONS TO GENETIC STUDIES

SUPPLEMENTARY TEXT: EFFICIENT COMPUTATION WITH A LINEAR MIXED MODEL ON LARGE-SCALE DATA SETS WITH APPLICATIONS TO GENETIC STUDIES SUPPLEMENTARY TEXT: EFFICIENT COMPUTATION WITH A LINEAR MIXED MODEL ON LARGE-SCALE DATA SETS WITH APPLICATIONS TO GENETIC STUDIES By Matti Pirinen, Peter Donnelly and Chris C.A. Spencer University of Oxford

More information

Stability and the elastic net

Stability and the elastic net Stability and the elastic net Patrick Breheny March 28 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/32 Introduction Elastic Net Our last several lectures have concentrated on methods for

More information

Chapter 3. Linear Models for Regression

Chapter 3. Linear Models for Regression Chapter 3. Linear Models for Regression Wei Pan Division of Biostatistics, School of Public Health, University of Minnesota, Minneapolis, MN 55455 Email: weip@biostat.umn.edu PubH 7475/8475 c Wei Pan Linear

More information

MSA220/MVE440 Statistical Learning for Big Data

MSA220/MVE440 Statistical Learning for Big Data MSA220/MVE440 Statistical Learning for Big Data Lecture 9-10 - High-dimensional regression Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology Recap from

More information

Regularization: Ridge Regression and the LASSO

Regularization: Ridge Regression and the LASSO Agenda Wednesday, November 29, 2006 Agenda Agenda 1 The Bias-Variance Tradeoff 2 Ridge Regression Solution to the l 2 problem Data Augmentation Approach Bayesian Interpretation The SVD and Ridge Regression

More information

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage Lingrui Gan, Naveen N. Narisetty, Feng Liang Department of Statistics University of Illinois at Urbana-Champaign Problem Statement

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider the problem of

More information

Variable Selection for Highly Correlated Predictors

Variable Selection for Highly Correlated Predictors Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu Department of Statistics, University of Illinois at Urbana-Champaign WHOA-PSI, Aug, 2017 St. Louis, Missouri 1 / 30 Background Variable

More information

Prediction & Feature Selection in GLM

Prediction & Feature Selection in GLM Tarigan Statistical Consulting & Coaching statistical-coaching.ch Doctoral Program in Computer Science of the Universities of Fribourg, Geneva, Lausanne, Neuchâtel, Bern and the EPFL Hands-on Data Analysis

More information

EFFICIENT COMPUTATION WITH A LINEAR MIXED MODEL ON LARGE-SCALE DATA SETS WITH APPLICATIONS TO GENETIC STUDIES

EFFICIENT COMPUTATION WITH A LINEAR MIXED MODEL ON LARGE-SCALE DATA SETS WITH APPLICATIONS TO GENETIC STUDIES Submitted to the Annals of Applied Statistics EFFICIENT COMPUTATION WITH A LINEAR MIXED MODEL ON LARGE-SCALE DATA SETS WITH APPLICATIONS TO GENETIC STUDIES By Matti Pirinen, Peter Donnelly and Chris C.A.

More information

A direct formulation for sparse PCA using semidefinite programming

A direct formulation for sparse PCA using semidefinite programming A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon

More information

27: Case study with popular GM III. 1 Introduction: Gene association mapping for complex diseases 1

27: Case study with popular GM III. 1 Introduction: Gene association mapping for complex diseases 1 10-708: Probabilistic Graphical Models, Spring 2015 27: Case study with popular GM III Lecturer: Eric P. Xing Scribes: Hyun Ah Song & Elizabeth Silver 1 Introduction: Gene association mapping for complex

More information

A Survey of L 1. Regression. Céline Cunen, 20/10/2014. Vidaurre, Bielza and Larranaga (2013)

A Survey of L 1. Regression. Céline Cunen, 20/10/2014. Vidaurre, Bielza and Larranaga (2013) A Survey of L 1 Regression Vidaurre, Bielza and Larranaga (2013) Céline Cunen, 20/10/2014 Outline of article 1.Introduction 2.The Lasso for Linear Regression a) Notation and Main Concepts b) Statistical

More information

Generalized Elastic Net Regression

Generalized Elastic Net Regression Abstract Generalized Elastic Net Regression Geoffroy MOURET Jean-Jules BRAULT Vahid PARTOVINIA This work presents a variation of the elastic net penalization method. We propose applying a combined l 1

More information

A Significance Test for the Lasso

A Significance Test for the Lasso A Significance Test for the Lasso Lockhart R, Taylor J, Tibshirani R, and Tibshirani R Ashley Petersen May 14, 2013 1 Last time Problem: Many clinical covariates which are important to a certain medical

More information

Theorems. Least squares regression

Theorems. Least squares regression Theorems In this assignment we are trying to classify AML and ALL samples by use of penalized logistic regression. Before we indulge on the adventure of classification we should first explain the most

More information

ISyE 691 Data mining and analytics

ISyE 691 Data mining and analytics ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)

More information

Exploratory quantile regression with many covariates: An application to adverse birth outcomes

Exploratory quantile regression with many covariates: An application to adverse birth outcomes Exploratory quantile regression with many covariates: An application to adverse birth outcomes June 3, 2011 eappendix 30 Percent of Total 20 10 0 0 1000 2000 3000 4000 5000 Birth weights efigure 1: Histogram

More information

Lecture 14: Variable Selection - Beyond LASSO

Lecture 14: Variable Selection - Beyond LASSO Fall, 2017 Extension of LASSO To achieve oracle properties, L q penalty with 0 < q < 1, SCAD penalty (Fan and Li 2001; Zhang et al. 2007). Adaptive LASSO (Zou 2006; Zhang and Lu 2007; Wang et al. 2007)

More information

MLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT

MLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT MLCC 2018 Variable Selection and Sparsity Lorenzo Rosasco UNIGE-MIT-IIT Outline Variable Selection Subset Selection Greedy Methods: (Orthogonal) Matching Pursuit Convex Relaxation: LASSO & Elastic Net

More information

Bi-level feature selection with applications to genetic association

Bi-level feature selection with applications to genetic association Bi-level feature selection with applications to genetic association studies October 15, 2008 Motivation In many applications, biological features possess a grouping structure Categorical variables may

More information

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization Proximal Newton Method Zico Kolter (notes by Ryan Tibshirani) Convex Optimization 10-725 Consider the problem Last time: quasi-newton methods min x f(x) with f convex, twice differentiable, dom(f) = R

More information

A direct formulation for sparse PCA using semidefinite programming

A direct formulation for sparse PCA using semidefinite programming A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley A. d Aspremont, INFORMS, Denver,

More information

A note on the group lasso and a sparse group lasso

A note on the group lasso and a sparse group lasso A note on the group lasso and a sparse group lasso arxiv:1001.0736v1 [math.st] 5 Jan 2010 Jerome Friedman Trevor Hastie and Robert Tibshirani January 5, 2010 Abstract We consider the group lasso penalty

More information

Package Grace. R topics documented: April 9, Type Package

Package Grace. R topics documented: April 9, Type Package Type Package Package Grace April 9, 2017 Title Graph-Constrained Estimation and Hypothesis Tests Version 0.5.3 Date 2017-4-8 Author Sen Zhao Maintainer Sen Zhao Description Use

More information

A SURVEY OF SOME SPARSE METHODS FOR HIGH DIMENSIONAL DATA

A SURVEY OF SOME SPARSE METHODS FOR HIGH DIMENSIONAL DATA A SURVEY OF SOME SPARSE METHODS FOR HIGH DIMENSIONAL DATA Gilbert Saporta CEDRIC, Conservatoire National des Arts et Métiers, Paris gilbert.saporta@cnam.fr With inputs from Anne Bernard, Ph.D student Industrial

More information

ESL Chap3. Some extensions of lasso

ESL Chap3. Some extensions of lasso ESL Chap3 Some extensions of lasso 1 Outline Consistency of lasso for model selection Adaptive lasso Elastic net Group lasso 2 Consistency of lasso for model selection A number of authors have studied

More information

Penalized methods in genome-wide association studies

Penalized methods in genome-wide association studies University of Iowa Iowa Research Online Theses and Dissertations Summer 2011 Penalized methods in genome-wide association studies Jin Liu University of Iowa Copyright 2011 Jin Liu This dissertation is

More information

On Mixture Regression Shrinkage and Selection via the MR-LASSO

On Mixture Regression Shrinkage and Selection via the MR-LASSO On Mixture Regression Shrinage and Selection via the MR-LASSO Ronghua Luo, Hansheng Wang, and Chih-Ling Tsai Guanghua School of Management, Peing University & Graduate School of Management, University

More information

CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS

CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS * Some contents are adapted from Dr. Hung Huang and Dr. Chengkai Li at UT Arlington Mingon Kang, Ph.D. Computer Science, Kennesaw State University Problems

More information

Least Angle Regression, Forward Stagewise and the Lasso

Least Angle Regression, Forward Stagewise and the Lasso January 2005 Rob Tibshirani, Stanford 1 Least Angle Regression, Forward Stagewise and the Lasso Brad Efron, Trevor Hastie, Iain Johnstone and Robert Tibshirani Stanford University Annals of Statistics,

More information

MS-C1620 Statistical inference

MS-C1620 Statistical inference MS-C1620 Statistical inference 10 Linear regression III Joni Virta Department of Mathematics and Systems Analysis School of Science Aalto University Academic year 2018 2019 Period III - IV 1 / 32 Contents

More information

Adaptive Lasso for correlated predictors

Adaptive Lasso for correlated predictors Adaptive Lasso for correlated predictors Keith Knight Department of Statistics University of Toronto e-mail: keith@utstat.toronto.edu This research was supported by NSERC of Canada. OUTLINE 1. Introduction

More information

Semi-Penalized Inference with Direct FDR Control

Semi-Penalized Inference with Direct FDR Control Jian Huang University of Iowa April 4, 2016 The problem Consider the linear regression model y = p x jβ j + ε, (1) j=1 where y IR n, x j IR n, ε IR n, and β j is the jth regression coefficient, Here p

More information

Recap from previous lecture

Recap from previous lecture Recap from previous lecture Learning is using past experience to improve future performance. Different types of learning: supervised unsupervised reinforcement active online... For a machine, experience

More information

Linear regression methods

Linear regression methods Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response

More information

Likelihood Ratio Tests. that Certain Variance Components Are Zero. Ciprian M. Crainiceanu. Department of Statistical Science

Likelihood Ratio Tests. that Certain Variance Components Are Zero. Ciprian M. Crainiceanu. Department of Statistical Science 1 Likelihood Ratio Tests that Certain Variance Components Are Zero Ciprian M. Crainiceanu Department of Statistical Science www.people.cornell.edu/pages/cmc59 Work done jointly with David Ruppert, School

More information

Consistent Group Identification and Variable Selection in Regression with Correlated Predictors

Consistent Group Identification and Variable Selection in Regression with Correlated Predictors Consistent Group Identification and Variable Selection in Regression with Correlated Predictors Dhruv B. Sharma, Howard D. Bondell and Hao Helen Zhang Abstract Statistical procedures for variable selection

More information

Efficient Bayesian mixed model analysis increases association power in large cohorts

Efficient Bayesian mixed model analysis increases association power in large cohorts Linear regression Existing mixed model methods New method: BOLT-LMM Time O(MM) O(MN 2 ) O MN 1.5 Corrects for confounding? Power Efficient Bayesian mixed model analysis increases association power in large

More information

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population

More information

Predicting causal effects in large-scale systems from observational data

Predicting causal effects in large-scale systems from observational data nature methods Predicting causal effects in large-scale systems from observational data Marloes H Maathuis 1, Diego Colombo 1, Markus Kalisch 1 & Peter Bühlmann 1,2 Supplementary figures and text: Supplementary

More information

SOLVING NON-CONVEX LASSO TYPE PROBLEMS WITH DC PROGRAMMING. Gilles Gasso, Alain Rakotomamonjy and Stéphane Canu

SOLVING NON-CONVEX LASSO TYPE PROBLEMS WITH DC PROGRAMMING. Gilles Gasso, Alain Rakotomamonjy and Stéphane Canu SOLVING NON-CONVEX LASSO TYPE PROBLEMS WITH DC PROGRAMMING Gilles Gasso, Alain Rakotomamonjy and Stéphane Canu LITIS - EA 48 - INSA/Universite de Rouen Avenue de l Université - 768 Saint-Etienne du Rouvray

More information

The lasso, persistence, and cross-validation

The lasso, persistence, and cross-validation The lasso, persistence, and cross-validation Daniel J. McDonald Department of Statistics Indiana University http://www.stat.cmu.edu/ danielmc Joint work with: Darren Homrighausen Colorado State University

More information

Some Sparse Methods for High Dimensional Data. Gilbert Saporta CEDRIC, Conservatoire National des Arts et Métiers, Paris

Some Sparse Methods for High Dimensional Data. Gilbert Saporta CEDRIC, Conservatoire National des Arts et Métiers, Paris Some Sparse Methods for High Dimensional Data Gilbert Saporta CEDRIC, Conservatoire National des Arts et Métiers, Paris gilbert.saporta@cnam.fr A joint work with Anne Bernard (QFAB Bioinformatics) Stéphanie

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Sparse Recovery using L1 minimization - algorithms Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

Association studies and regression

Association studies and regression Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration

More information

arxiv: v3 [stat.ml] 14 Apr 2016

arxiv: v3 [stat.ml] 14 Apr 2016 arxiv:1307.0048v3 [stat.ml] 14 Apr 2016 Simple one-pass algorithm for penalized linear regression with cross-validation on MapReduce Kun Yang April 15, 2016 Abstract In this paper, we propose a one-pass

More information

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson Bayesian variable selection via penalized credible regions Brian Reich, NC State Joint work with Howard Bondell and Ander Wilson Brian Reich, NCSU Penalized credible regions 1 Motivation big p, small n

More information

Bayesian Grouped Horseshoe Regression with Application to Additive Models

Bayesian Grouped Horseshoe Regression with Application to Additive Models Bayesian Grouped Horseshoe Regression with Application to Additive Models Zemei Xu, Daniel F. Schmidt, Enes Makalic, Guoqi Qian, and John L. Hopper Centre for Epidemiology and Biostatistics, Melbourne

More information

Linear Regression. Volker Tresp 2018

Linear Regression. Volker Tresp 2018 Linear Regression Volker Tresp 2018 1 Learning Machine: The Linear Model / ADALINE As with the Perceptron we start with an activation functions that is a linearly weighted sum of the inputs h = M j=0 w

More information

Sparse PCA with applications in finance

Sparse PCA with applications in finance Sparse PCA with applications in finance A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon 1 Introduction

More information

The lasso. Patrick Breheny. February 15. The lasso Convex optimization Soft thresholding

The lasso. Patrick Breheny. February 15. The lasso Convex optimization Soft thresholding Patrick Breheny February 15 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/24 Introduction Last week, we introduced penalized regression and discussed ridge regression, in which the penalty

More information

Linear Regression Linear Regression with Shrinkage

Linear Regression Linear Regression with Shrinkage Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle

More information

Linear Regression Models. Based on Chapter 3 of Hastie, Tibshirani and Friedman

Linear Regression Models. Based on Chapter 3 of Hastie, Tibshirani and Friedman Linear Regression Models Based on Chapter 3 of Hastie, ibshirani and Friedman Linear Regression Models Here the X s might be: p f ( X = " + " 0 j= 1 X j Raw predictor variables (continuous or coded-categorical

More information

Statistics for high-dimensional data: Group Lasso and additive models

Statistics for high-dimensional data: Group Lasso and additive models Statistics for high-dimensional data: Group Lasso and additive models Peter Bühlmann and Sara van de Geer Seminar für Statistik, ETH Zürich May 2012 The Group Lasso (Yuan & Lin, 2006) high-dimensional

More information

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS023) p.3938 An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Vitara Pungpapong

More information

An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss

An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss arxiv:1811.04545v1 [stat.co] 12 Nov 2018 Cheng Wang School of Mathematical Sciences, Shanghai Jiao

More information

Nonconvex penalties: Signal-to-noise ratio and algorithms

Nonconvex penalties: Signal-to-noise ratio and algorithms Nonconvex penalties: Signal-to-noise ratio and algorithms Patrick Breheny March 21 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/22 Introduction In today s lecture, we will return to nonconvex

More information

Group descent algorithms for nonconvex penalized linear and logistic regression models with grouped predictors

Group descent algorithms for nonconvex penalized linear and logistic regression models with grouped predictors Group descent algorithms for nonconvex penalized linear and logistic regression models with grouped predictors Patrick Breheny Department of Biostatistics University of Iowa Jian Huang Department of Statistics

More information

SCMA292 Mathematical Modeling : Machine Learning. Krikamol Muandet. Department of Mathematics Faculty of Science, Mahidol University.

SCMA292 Mathematical Modeling : Machine Learning. Krikamol Muandet. Department of Mathematics Faculty of Science, Mahidol University. SCMA292 Mathematical Modeling : Machine Learning Krikamol Muandet Department of Mathematics Faculty of Science, Mahidol University February 9, 2016 Outline Quick Recap of Least Square Ridge Regression

More information

Smoothly Clipped Absolute Deviation (SCAD) for Correlated Variables

Smoothly Clipped Absolute Deviation (SCAD) for Correlated Variables Smoothly Clipped Absolute Deviation (SCAD) for Correlated Variables LIB-MA, FSSM Cadi Ayyad University (Morocco) COMPSTAT 2010 Paris, August 22-27, 2010 Motivations Fan and Li (2001), Zou and Li (2008)

More information

Robust Variable Selection Through MAVE

Robust Variable Selection Through MAVE Robust Variable Selection Through MAVE Weixin Yao and Qin Wang Abstract Dimension reduction and variable selection play important roles in high dimensional data analysis. Wang and Yin (2008) proposed sparse

More information

Multiple regression. CM226: Machine Learning for Bioinformatics. Fall Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar

Multiple regression. CM226: Machine Learning for Bioinformatics. Fall Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Multiple regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Multiple regression 1 / 36 Previous two lectures Linear and logistic

More information

Linear Regression Linear Regression with Shrinkage

Linear Regression Linear Regression with Shrinkage Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle

More information

Big Data Analytics: Optimization and Randomization

Big Data Analytics: Optimization and Randomization Big Data Analytics: Optimization and Randomization Tianbao Yang Tutorial@ACML 2015 Hong Kong Department of Computer Science, The University of Iowa, IA, USA Nov. 20, 2015 Yang Tutorial for ACML 15 Nov.

More information

Lecture 2 Part 1 Optimization

Lecture 2 Part 1 Optimization Lecture 2 Part 1 Optimization (January 16, 2015) Mu Zhu University of Waterloo Need for Optimization E(y x), P(y x) want to go after them first, model some examples last week then, estimate didn t discuss

More information

Estimating subgroup specific treatment effects via concave fusion

Estimating subgroup specific treatment effects via concave fusion Estimating subgroup specific treatment effects via concave fusion Jian Huang University of Iowa April 6, 2016 Outline 1 Motivation and the problem 2 The proposed model and approach Concave pairwise fusion

More information

Regularization Paths. Theme

Regularization Paths. Theme June 00 Trevor Hastie, Stanford Statistics June 00 Trevor Hastie, Stanford Statistics Theme Regularization Paths Trevor Hastie Stanford University drawing on collaborations with Brad Efron, Mee-Young Park,

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

Lecture 6: Methods for high-dimensional problems

Lecture 6: Methods for high-dimensional problems Lecture 6: Methods for high-dimensional problems Hector Corrada Bravo and Rafael A. Irizarry March, 2010 In this Section we will discuss methods where data lies on high-dimensional spaces. In particular,

More information

Final Overview. Introduction to ML. Marek Petrik 4/25/2017

Final Overview. Introduction to ML. Marek Petrik 4/25/2017 Final Overview Introduction to ML Marek Petrik 4/25/2017 This Course: Introduction to Machine Learning Build a foundation for practice and research in ML Basic machine learning concepts: max likelihood,

More information

Lecture 17: October 27

Lecture 17: October 27 0-725/36-725: Convex Optimiation Fall 205 Lecturer: Ryan Tibshirani Lecture 7: October 27 Scribes: Brandon Amos, Gines Hidalgo Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These

More information

Permutation-invariant regularization of large covariance matrices. Liza Levina

Permutation-invariant regularization of large covariance matrices. Liza Levina Liza Levina Permutation-invariant covariance regularization 1/42 Permutation-invariant regularization of large covariance matrices Liza Levina Department of Statistics University of Michigan Joint work

More information

Linear Model Selection and Regularization

Linear Model Selection and Regularization Linear Model Selection and Regularization Recall the linear model Y = β 0 + β 1 X 1 + + β p X p + ɛ. In the lectures that follow, we consider some approaches for extending the linear model framework. In

More information

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1 Variable Selection in Restricted Linear Regression Models Y. Tuaç 1 and O. Arslan 1 Ankara University, Faculty of Science, Department of Statistics, 06100 Ankara/Turkey ytuac@ankara.edu.tr, oarslan@ankara.edu.tr

More information

Outlier detection and variable selection via difference based regression model and penalized regression

Outlier detection and variable selection via difference based regression model and penalized regression Journal of the Korean Data & Information Science Society 2018, 29(3), 815 825 http://dx.doi.org/10.7465/jkdi.2018.29.3.815 한국데이터정보과학회지 Outlier detection and variable selection via difference based regression

More information

Package tsvd. March 24, 2015

Package tsvd. March 24, 2015 Package tsvd March 24, 2015 Type Package Title Thresholding-based SVD for multivariate reduced rank regression Version 1.3 Date 2015-03-24 Author Xin Ma, Luo Xiao, and Wing H. Wong* Maintainer Xin Ma

More information

Regularization Path Algorithms for Detecting Gene Interactions

Regularization Path Algorithms for Detecting Gene Interactions Regularization Path Algorithms for Detecting Gene Interactions Mee Young Park Trevor Hastie July 16, 2006 Abstract In this study, we consider several regularization path algorithms with grouped variable

More information

Fused elastic net EPoC

Fused elastic net EPoC Fused elastic net EPoC Tobias Abenius November 28, 2013 Data Set and Method Use mrna and copy number aberration (CNA) to build a network using an extension of our EPoC method (Jörnsten et al., 2011; Abenius

More information

Sparsity Regularization

Sparsity Regularization Sparsity Regularization Bangti Jin Course Inverse Problems & Imaging 1 / 41 Outline 1 Motivation: sparsity? 2 Mathematical preliminaries 3 l 1 solvers 2 / 41 problem setup finite-dimensional formulation

More information

Package tsvd. March 11, 2015

Package tsvd. March 11, 2015 Package tsvd March 11, 2015 Type Package Title Thresholding-based SVD for multivariate reduced rank regression Version 1.1 Date 2015-03-09 Author Xin Ma, Luo Xiao, and Wing H. Wong* Maintainer Xin Ma

More information

Variable Selection for Highly Correlated Predictors

Variable Selection for Highly Correlated Predictors Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu arxiv:1709.04840v1 [stat.me] 14 Sep 2017 Abstract Penalty-based variable selection methods are powerful in selecting relevant covariates

More information

High-dimensional regression with unknown variance

High-dimensional regression with unknown variance High-dimensional regression with unknown variance Christophe Giraud Ecole Polytechnique march 2012 Setting Gaussian regression with unknown variance: Y i = f i + ε i with ε i i.i.d. N (0, σ 2 ) f = (f

More information

Correlate. A method for the integrative analysis of two genomic data sets

Correlate. A method for the integrative analysis of two genomic data sets Correlate A method for the integrative analysis of two genomic data sets Sam Gross, Balasubramanian Narasimhan, Robert Tibshirani, and Daniela Witten February 19, 2010 Introduction Sparse Canonical Correlation

More information

Sparse Linear Models (10/7/13)

Sparse Linear Models (10/7/13) STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine

More information

Least Absolute Shrinkage is Equivalent to Quadratic Penalization

Least Absolute Shrinkage is Equivalent to Quadratic Penalization Least Absolute Shrinkage is Equivalent to Quadratic Penalization Yves Grandvalet Heudiasyc, UMR CNRS 6599, Université de Technologie de Compiègne, BP 20.529, 60205 Compiègne Cedex, France Yves.Grandvalet@hds.utc.fr

More information

Sparse Graph Learning via Markov Random Fields

Sparse Graph Learning via Markov Random Fields Sparse Graph Learning via Markov Random Fields Xin Sui, Shao Tang Sep 23, 2016 Xin Sui, Shao Tang Sparse Graph Learning via Markov Random Fields Sep 23, 2016 1 / 36 Outline 1 Introduction to graph learning

More information

Homogeneity Pursuit. Jianqing Fan

Homogeneity Pursuit. Jianqing Fan Jianqing Fan Princeton University with Tracy Ke and Yichao Wu http://www.princeton.edu/ jqfan June 5, 2014 Get my own profile - Help Amazing Follow this author Grace Wahba 9 Followers Follow new articles

More information

Linear model selection and regularization

Linear model selection and regularization Linear model selection and regularization Problems with linear regression with least square 1. Prediction Accuracy: linear regression has low bias but suffer from high variance, especially when n p. It

More information

Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty

Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty Journal of Data Science 9(2011), 549-564 Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty Masaru Kanba and Kanta Naito Shimane University Abstract: This paper discusses the

More information

Generalized Linear Models and Its Asymptotic Properties

Generalized Linear Models and Its Asymptotic Properties for High Dimensional Generalized Linear Models and Its Asymptotic Properties April 21, 2012 for High Dimensional Generalized L Abstract Literature Review In this talk, we present a new prior setting for

More information