Graphlet Screening (GS)

Size: px
Start display at page:

Download "Graphlet Screening (GS)"

Transcription

1 Graphlet Screening (GS) Jiashun Jin Carnegie Mellon University April 11, 2014 Jiashun Jin Graphlet Screening (GS) 1 / 36

2 Collaborators Alphabetically: Zheng (Tracy) Ke Cun-Hui Zhang Qi Zhang Princeton Rutgers University of Wisconsin Jiashun Jin Graphlet Screening (GS) 2 / 36

3 Rare/Weak signals in LSI Rare. Signals sparsely scatter across different observation units; no priori where there are Weak. Signals are individually weak (new) Box and Meyer (1986) Jiashun Jin Graphlet Screening (GS) 3 / 36

4 A linear model Y = X β + z, X = X n,p, z N(0, I n ) p n 1 β is sparse: many coordinates are 0 Gram matrix G = X X has unit diagonals is sparse (few large coordinates in each row) Jiashun Jin Graphlet Screening (GS) 4 / 36

5 Linkage Disequilibrium (LD) Jiashun Jin Graphlet Screening (GS) 5 / 36

6 Idea Jiashun Jin Graphlet Screening (GS) 5 / 36

7 Univariate Screening Y = X β + z, X = X n,p = [x 1, x 2,..., x p ] For a threshold t > 0, apply Hard-Thresholding: { ˆβ j HT (Y, xj ), (Y, x = j ) t, 0, otherwise Jiashun Jin Graphlet Screening (GS) 6 / 36

8 Signal cancellation Even if β j is large, E[(x j, Y )] could be small Denote the support of β by S = S(β) = {1 j p : β j 0} p (Y, x j ) = (x l, x j )β l + (z, x j ) l=1 = β j + l S(β)\{j} (x j, x l )β l + N(0, 1), signal cancellation may happen as x j {x l : l S(β) \ {j}} Jiashun Jin Graphlet Screening (GS) 7 / 36

9 Exhaustive Multivariate Screening (EMS) Fix m 0 1 and a small π 0 > 0 For m = 1, 2,..., m 0 and any subset J = {j 1, j 2,..., j m }, j 1 < j 2 <... < j m Project Y onto {x j1, x j2,..., x jm }: Y P J Y Retain the nodes in J if and only if P ( χ 2 m(0) P J Y 2) π 0 Fit the model with all retained nodes Jiashun Jin Graphlet Screening (GS) 8 / 36

10 Rationale Hope: maybe for some small-size set J, {x j : j J } {x l : l S(β) \ J }; Y = β j x j + β j x j + z j J j S(β)\J Possible signal cancellation if we look at any single projected coefficients (Y, x j ) No signal cancelation if look at the projected coefficients {(Y, x j ) : j J } together Jiashun Jin Graphlet Screening (GS) 9 / 36

11 Problems Computationally infeasible: m 0 m=1 ( ) p m Inefficiency: include too many candidates for screening; signals need to be stronger than necessary to survive the screening Our proposal: Graphlet Screening (GS) Jiashun Jin Graphlet Screening (GS) 10 / 36

12 Graph of Strong Dependence (GOSD) Define GOSD as the graph G = (V, E): V = {1, 2,..., p}: each variable is a node Nodes i and j have an edge iff (x i, x j ) δ, (δ = 1 log(p), say) G = X X sparse = G sparse Jiashun Jin Graphlet Screening (GS) 11 / 36

13 Graphlet Screening (GS) A(m 0 ) = {All connected subgraphs of G; size m 0 } GS: same as EMS, except for EMS exhaustively screens all J = {j 1, j 2,..., j m }, j 1 < j 2 <... < j m GS screens J if and only if J A(m 0 ) Lemma. If d be the maximum degree of G, then A(m 0 ) Cp(ed) m 0 ; e = Jiashun Jin Graphlet Screening (GS) 12 / 36

14 Mutual orthogonality G: there is an edge between i and j (x i, x j ) 1/ log(p) S = S(β) = {1 i p : β i 0} Restricting nodes to S forms a subgraph G S, and G S = G S,1... G S,M : G S,l : components By how G is defined, approximately, {x j : j G S,1 } {x j : j G S,2 }... {x j : j G S,M } Jiashun Jin Graphlet Screening (GS) 13 / 36

15 Maximum component size m0 (β, G, δ) Surprise. m0 (β, G, δ) is small in many cases, where m 0(β, G, δ) = max 1 l M G S,l Lemma. If I {β j 0} iid Bernoulli(ɛ), then P ( m 0(β, G, δ) > m ) p(edɛ) m+1, m > 1, where d = d(g) is maximum degree of G Jiashun Jin Graphlet Screening (GS) 14 / 36

16 Signal archipelago I Signals split into many small-size disconnected islands I Columns indexed by different islands: mutual orthogonal I No signal cancellation when we screen each island individually Jiashun Jin Graphlet Screening (GS) 15 / 36

17 Why GS works A(m 0 ) = {All connected subgraphs of G; size m 0 }, G S = G S,1 G S,2... G S,M ; maximum size m0 (β, G, δ) GS screens a set J J A(m 0 ) If m 0 m0 (β, G, δ) G S,l A(m 0 ) : on our screen list! Approximately, no signal cancellation for {x j : G S,l } {x j : G S \ G S,l }, approx. Y = β j x j + j G S,l j G S \G S,l β j x j + z Jiashun Jin Graphlet Screening (GS) 16 / 36

18 Summary Method US ( EMS GS Computation p p ) m 0 Cp(ed) m 0 Efficiency yes no yes Robust to Signal Cancellation no yes yes Jiashun Jin Graphlet Screening (GS) 17 / 36

19 Methods Jiashun Jin Graphlet Screening (GS) 17 / 36

20 Application I: rank features For feature j, find all J j, calculate P-values; use π j = minimum of all such P-values for ranking Example. Rows of X = X n,p are iid N(0, Ω) Ω block-wise diagonal, with building blocks ( 1 ρ ) ρ 1 β: partitioned to blocks of 2, (0, 0), prob. (1 ɛ), (0, τ), prob. ɛ/4, (β 2k 1, β 2k ) = (τ, 0), prob. ɛ/4, (τ, τ), prob. ɛ/2 Jiashun Jin Graphlet Screening (GS) 18 / 36

21 ROC of US vs ROC of GS (m 0 = 2) p = 1000, n = 500, ɛ = ρ = 0.8,τ = 4 US GS ρ = 0.8,τ = 1.5 US GS Jiashun Jin Graphlet Screening (GS) 19 / 36

22 SNP data (chromosome 21) GS: m 0 = 2; p = 3937, n = Left: (β 2k 1, β 2k ) = (τ, τ), prob Right: (β 3k 2, β 3k 1, β 3k ) = (τ, τ, τ), prob pair signals, τ = triplet signals, τ = US GS US GS Jiashun Jin Graphlet Screening (GS) 20 / 36

23 Application II. Variable Selection A two-stage screen and clean procedure: Apply GS, with small modifications Clean Jiashun Jin Graphlet Screening (GS) 21 / 36

24 Screen step A(m 0 ) = {All connected subgraphs of G with size m 0 } Arrange by size, ties breaking lexicographically: A(m 0 ) = {J 1, J 2,..., J T } Initializing with S 0 = For t = 1, 2,..., T, letting S t 1 be the set of retained indices in stage t 1, update S t 1 by { St 1 J S t = t, if P Jt Y 2 P Jt S t 1 Y 2 2q log(p), S t 1, otherwise Ŝ = S T : all retained nodes Jiashun Jin Graphlet Screening (GS) 22 / 36

25 Clean step Fixing tuning parameters (u gs, v gs ), j / Ŝ: set ˆβ gs j = 0 j Ŝ: decompose GŜ = GŜ,1 GŜ,2... GŜ, ˆM, and estimate {β j : j GŜ,l } by minimizing P G Ŝ,l (Y j GŜ,l β j x j ) 2 + (u gs ) 2 β 0, subject to either β j = 0 or β j v gs Jiashun Jin Graphlet Screening (GS) 23 / 36

26 Tuning parameters of GS Feature ranking: need (δ, m 0 ); choices of which can be guided by G and computation capacity Variable selection: also need (q, u gs, v gs ) q: flexible and insensitive u gs is relatively easy to estimate v gs is relatively hard to estimate Jiashun Jin Graphlet Screening (GS) 24 / 36

27 Minimax Theory Jiashun Jin Graphlet Screening (GS) 24 / 36

28 Sparse model Y = X β + z, z N(0, I n ) β = b µ, b i iid Bernoulli(ɛ), µ M p (τ, a) b µ R p : (b µ) j = b j µ j M p(τ, a) = {µ R p : τ µ j a τ} As p, link (ɛ, τ) to p by fixed parameters: ɛ = ɛ p = p ϑ, τ = τ p = 2r log(p) Jiashun Jin Graphlet Screening (GS) 25 / 36

29 Sparse model, II. Random design Y = X β+z, X = X 1... X n, X i iid N(0, 1 n Ω) Ω: unknown correlation matrix n = n p = p θ and (1 ϑ) < θ < 1, so that pɛ p n p p Jiashun Jin Graphlet Screening (GS) 26 / 36

30 Minimax Hamming distance Measuring errors with Hamming distance: [ p H p ( ˆβ, ɛ p, µ; Ω) = E j=1 Minimax Hamming distance: Hamm p(ϑ, θ, r, a, Ω) = inf ˆβ 1 { sgn( ˆβ j ) sgn(β j ) }] sup H p ( ˆβ, ɛ p, µ; Ω) µ M p(τ p,a) Jiashun Jin Graphlet Screening (GS) 27 / 36

31 Exponent ρ j = ρ j (ϑ, r, Ω) { Define ω = ω(s 0, S 1 ; Ω) = inf δ δ Ωδ } where { δ u (0) u (1) u (k) : i = 0, i / S k 1 u (k), k = 0, 1 i a, i S k Define ρ(s 0, S 1 ; ϑ, r, a, Ω) = S 0 + S 1 ϑ + ωr ( S 1 S 0 ) 2 ϑ 2 4ωr Minimax rate critically depends on the exponents: ρ j = ρ j (ϑ, r; Ω) = min ρ(s 0, S 1, ϑ, r, a, Ω) (S 0,S 1 ):j S 0 S 1 not dependent on (θ, a) (mild regularity cond.) computable; has explicit form for some Ω Jiashun Jin Graphlet Screening (GS) 28 / 36

32 Asymptotic minimaxity of GS Assume p j=1 Ω(i, j) γ C, γ (0, 1), 1 i p Screen-step: q is a properly small number Clean-step: set u gs = 2ϑ log p, and v gs = τ p Theorem GS achieves optimal rate of convergence: sup µ M p(τ p,a) H p ( ˆβ gs, ɛ p, µ, Ω) L p [ ( p j=1 p ρ j ] ) + p 1 (m 0 +1)ϑ L p [ Hamm p(ϑ, θ, r, a, Ω) + p 1 (m 0+1)ϑ where L p is a generic multi-log(p) term Jiashun Jin Graphlet Screening (GS) 29 / 36 ]

33 L 0 -penalization method Donoho and Starck (1989) Y X β 2 + λ β 0 : λ > 0: tuning parameter Idea: Where there is no noise Infinite solutions to Y = X β But only one is very sparse Hope: the sparsest solution is the truth (Occam s Razor) Noiseless Strong noise Jiashun Jin Graphlet Screening (GS) 30 / 36

34 L 0 /L 1 -penalization methods L 1 -solution L 0 -solution truth Method L 0 /L 1 -penalization CASE Regime Rare/Strong Rare/Weak Loss I {sgn( ˆβ) sgn(β)} Hamm(sgn( ˆβ), sgn(β)) Optimality Not Yes Motivation Imaging/Engineering Genetics/Genomics Design Controllable/Nice Uncontrollable/Bad Key idea One-stage global method Multi-stage local method Hamm: Hamming distance Donoho and Stark (1989), Tibshiraini (1996), and many others Jiashun Jin Graphlet Screening (GS) 31 / 36

35 Phase Diagram Jiashun Jin Graphlet Screening (GS) 31 / 36

36 A Corollary Suppose the conditions of the theorem hold. If additionally, Ω(i, j) 4 2 5, i j, then { Hamm p(ϑ, θ, r, a, Ω) 1 + o(1), r < ϑ, = pɛ p L p p (ϑ r)2 4r, 1 < r ϑ < Right hand side: rate when Ω = I p ; Jiashun Jin Graphlet Screening (GS) 32 / 36

37 Phase Diagram Exact Recovery Exact Recovery 3 3 r r Almost Full Recovery 1.5 Almost Full Recovery No Recovery ϑ 0.5 No Recovery Left: Ω = I p ; line r = ϑ; curve r = (1 + 1 ϑ) 2. Right: Ω as in the corollary; green line: r/ϑ = ϑ Jiashun Jin Graphlet Screening (GS) 33 / 36

38 Simulation comparison LASSO Graphlet Screening LASSO Graphlet Screening LASSO Graphlet Screening by-2 blockwise 0.3 Penta-diagonal 0.3 Random Correlation τ p τ p τ p p = 5000, n = 4000, pɛ p = 250; τ p = 6, 7,..., 12. Left to right: G is block-wise, penta-diagonal, randomly generated ( sprandsym in matlab). Jiashun Jin Graphlet Screening (GS) 34 / 36

39 Take-home messages GS is a computationally feasible approach that overcomes the challenge of signal cancellation Key insight: GS splits into different small-size signal islands G S,l, and {x j : j G S,l } are mutual orthogonal (approx). minimax rate depends on X locally so we have to act locally Optimal in variable selection, while penalization methods are not A flexible idea and that is useful in many different situations Jiashun Jin Graphlet Screening (GS) 35 / 36

40 Main References Genovese C, Jin J, Wasserman L, and Yao Z (2012) A comparison of the lasso and marginal regression. J. Mach. Learn. Res., Ji P and Jin J (2012) UPS delivers optimal phase diagram in high-dimensional variable selection Ann. Statist. 40(1), Jin J, Zhang C-H and Zhang Q (2012) Optimality of Graphlet Screening in High Dimensional Variable Selection. arxiv: Ke T, Jin J and Fan J (2012) Covariance assisted screening and estimation. arxiv: Jiashun Jin Graphlet Screening (GS) 36 / 36

Covariate-Assisted Variable Ranking

Covariate-Assisted Variable Ranking Covariate-Assisted Variable Ranking Tracy Ke Department of Statistics Harvard University WHOA-PSI@St. Louis, Sep. 8, 2018 1/18 Sparse linear regression Y = X β + z, X R n,p, z N(0, σ 2 I n ) Signals (nonzero

More information

OPTIMAL PROCEDURES IN HIGH-DIMENSIONAL VARIABLE SELECTION

OPTIMAL PROCEDURES IN HIGH-DIMENSIONAL VARIABLE SELECTION OPTIMAL PROCEDURES IN HIGH-DIMENSIONAL VARIABLE SELECTION by Qi Zhang Bachelor of Science in Math and Applied Math, China Agricultural University, 2008 Submitted to the Graduate Faculty of the Department

More information

arxiv: v2 [math.st] 25 Aug 2012

arxiv: v2 [math.st] 25 Aug 2012 Submitted to the Annals of Statistics arxiv: math.st/1205.4645 COVARIANCE ASSISTED SCREENING AND ESTIMATION arxiv:1205.4645v2 [math.st] 25 Aug 2012 By Tracy Ke,, Jiashun Jin and Jianqing Fan Princeton

More information

COVARIATE ASSISTED SCREENING AND ESTIMATION

COVARIATE ASSISTED SCREENING AND ESTIMATION The Annals of Statistics 2014, Vol. 42, No. 6, 2202 2242 DOI: 10.1214/14-AOS1243 Institute of Mathematical Statistics, 2014 COVARIATE ASSISTED SCREENING AND ESTIMATION BY ZHENG TRACY KE 1,2,3,JIASHUN JIN

More information

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology

More information

arxiv: v1 [math.st] 20 Nov 2009

arxiv: v1 [math.st] 20 Nov 2009 REVISITING MARGINAL REGRESSION arxiv:0911.4080v1 [math.st] 20 Nov 2009 By Christopher R. Genovese Jiashun Jin and Larry Wasserman Carnegie Mellon University May 31, 2018 The lasso has become an important

More information

Clustering by Important Features PCA (IF-PCA)

Clustering by Important Features PCA (IF-PCA) Clustering by Important Features PCA (IF-PCA) Rare/Weak Signals and Phase Diagrams Jiashun Jin, CMU Zheng Tracy Ke (Univ. of Chicago) Wanjie Wang (Univ. of Pennsylvania) August 5, 2015 Jiashun Jin, CMU

More information

COVARIATE ASSISTED VARIABLE RANKING. By Zheng Tracy Ke and Fan Yang University of Chicago

COVARIATE ASSISTED VARIABLE RANKING. By Zheng Tracy Ke and Fan Yang University of Chicago Submitted to the Annals of Statistics arxiv: arxiv:0000.0000 COVARIATE ASSISTED VARIABLE RANKING By Zheng Tracy Ke and Fan Yang University of Chicago Consider a linear model y = Xβ + z, z N(0, σ 2 I n).

More information

Clustering by Important Features PCA (IF-PCA)

Clustering by Important Features PCA (IF-PCA) Clustering by Important Features PCA (IF-PCA) Rare/Weak Signals and Phase Diagrams Jiashun Jin, CMU David Donoho (Stanford) Zheng Tracy Ke (Univ. of Chicago) Wanjie Wang (Univ. of Pennsylvania) August

More information

Gaussian Graphical Models and Graphical Lasso

Gaussian Graphical Models and Graphical Lasso ELE 538B: Sparsity, Structure and Inference Gaussian Graphical Models and Graphical Lasso Yuxin Chen Princeton University, Spring 2017 Multivariate Gaussians Consider a random vector x N (0, Σ) with pdf

More information

Consistent high-dimensional Bayesian variable selection via penalized credible regions

Consistent high-dimensional Bayesian variable selection via penalized credible regions Consistent high-dimensional Bayesian variable selection via penalized credible regions Howard Bondell bondell@stat.ncsu.edu Joint work with Brian Reich Howard Bondell p. 1 Outline High-Dimensional Variable

More information

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28 Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:

More information

Marginal Regression For Multitask Learning

Marginal Regression For Multitask Learning Mladen Kolar Machine Learning Department Carnegie Mellon University mladenk@cs.cmu.edu Han Liu Biostatistics Johns Hopkins University hanliu@jhsph.edu Abstract Variable selection is an important and practical

More information

Sample Size Requirement For Some Low-Dimensional Estimation Problems

Sample Size Requirement For Some Low-Dimensional Estimation Problems Sample Size Requirement For Some Low-Dimensional Estimation Problems Cun-Hui Zhang, Rutgers University September 10, 2013 SAMSI Thanks for the invitation! Acknowledgements/References Sun, T. and Zhang,

More information

A Comparison of the Lasso and Marginal Regression

A Comparison of the Lasso and Marginal Regression Journal of Machine Learning Research 13 (2012) 2107-2143 Submitted 5/11; Revised 1/12; Published 6/12 A Comparison of the Lasso and Marginal Regression Christopher R. Genovese Jiashun Jin Larry Wasserman

More information

Optimal Linear Estimation under Unknown Nonlinear Transform

Optimal Linear Estimation under Unknown Nonlinear Transform Optimal Linear Estimation under Unknown Nonlinear Transform Xinyang Yi The University of Texas at Austin yixy@utexas.edu Constantine Caramanis The University of Texas at Austin constantine@utexas.edu Zhaoran

More information

Robust Principal Component Analysis

Robust Principal Component Analysis ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M

More information

High-dimensional covariance estimation based on Gaussian graphical models

High-dimensional covariance estimation based on Gaussian graphical models High-dimensional covariance estimation based on Gaussian graphical models Shuheng Zhou Department of Statistics, The University of Michigan, Ann Arbor IMA workshop on High Dimensional Phenomena Sept. 26,

More information

Nonconcave Penalized Likelihood with A Diverging Number of Parameters

Nonconcave Penalized Likelihood with A Diverging Number of Parameters Nonconcave Penalized Likelihood with A Diverging Number of Parameters Jianqing Fan and Heng Peng Presenter: Jiale Xu March 12, 2010 Jianqing Fan and Heng Peng Presenter: JialeNonconcave Xu () Penalized

More information

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider the problem of

More information

sparse and low-rank tensor recovery Cubic-Sketching

sparse and low-rank tensor recovery Cubic-Sketching Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru

More information

19.1 Problem setup: Sparse linear regression

19.1 Problem setup: Sparse linear regression ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 19: Minimax rates for sparse linear regression Lecturer: Yihong Wu Scribe: Subhadeep Paul, April 13/14, 2016 In

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

High-dimensional Covariance Estimation Based On Gaussian Graphical Models

High-dimensional Covariance Estimation Based On Gaussian Graphical Models High-dimensional Covariance Estimation Based On Gaussian Graphical Models Shuheng Zhou, Philipp Rutimann, Min Xu and Peter Buhlmann February 3, 2012 Problem definition Want to estimate the covariance matrix

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57

More information

Compressed Sensing and Sparse Recovery

Compressed Sensing and Sparse Recovery ELE 538B: Sparsity, Structure and Inference Compressed Sensing and Sparse Recovery Yuxin Chen Princeton University, Spring 217 Outline Restricted isometry property (RIP) A RIPless theory Compressed sensing

More information

Homogeneity Pursuit. Jianqing Fan

Homogeneity Pursuit. Jianqing Fan Jianqing Fan Princeton University with Tracy Ke and Yichao Wu http://www.princeton.edu/ jqfan June 5, 2014 Get my own profile - Help Amazing Follow this author Grace Wahba 9 Followers Follow new articles

More information

Semi-Penalized Inference with Direct FDR Control

Semi-Penalized Inference with Direct FDR Control Jian Huang University of Iowa April 4, 2016 The problem Consider the linear regression model y = p x jβ j + ε, (1) j=1 where y IR n, x j IR n, ε IR n, and β j is the jth regression coefficient, Here p

More information

DISCUSSION OF INFLUENTIAL FEATURE PCA FOR HIGH DIMENSIONAL CLUSTERING. By T. Tony Cai and Linjun Zhang University of Pennsylvania

DISCUSSION OF INFLUENTIAL FEATURE PCA FOR HIGH DIMENSIONAL CLUSTERING. By T. Tony Cai and Linjun Zhang University of Pennsylvania Submitted to the Annals of Statistics DISCUSSION OF INFLUENTIAL FEATURE PCA FOR HIGH DIMENSIONAL CLUSTERING By T. Tony Cai and Linjun Zhang University of Pennsylvania We would like to congratulate the

More information

Composite Loss Functions and Multivariate Regression; Sparse PCA

Composite Loss Functions and Multivariate Regression; Sparse PCA Composite Loss Functions and Multivariate Regression; Sparse PCA G. Obozinski, B. Taskar, and M. I. Jordan (2009). Joint covariate selection and joint subspace selection for multiple classification problems.

More information

Optimal prediction for sparse linear models? Lower bounds for coordinate-separable M-estimators

Optimal prediction for sparse linear models? Lower bounds for coordinate-separable M-estimators Electronic Journal of Statistics ISSN: 935-7524 arxiv: arxiv:503.0388 Optimal prediction for sparse linear models? Lower bounds for coordinate-separable M-estimators Yuchen Zhang, Martin J. Wainwright

More information

The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression

The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression Cun-hui Zhang and Jian Huang Presenter: Quefeng Li Feb. 26, 2010 un-hui Zhang and Jian Huang Presenter: Quefeng The Sparsity

More information

Regularization and Variable Selection via the Elastic Net

Regularization and Variable Selection via the Elastic Net p. 1/1 Regularization and Variable Selection via the Elastic Net Hui Zou and Trevor Hastie Journal of Royal Statistical Society, B, 2005 Presenter: Minhua Chen, Nov. 07, 2008 p. 2/1 Agenda Introduction

More information

The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA

The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA Presented by Dongjun Chung March 12, 2010 Introduction Definition Oracle Properties Computations Relationship: Nonnegative Garrote Extensions:

More information

19.1 Maximum Likelihood estimator and risk upper bound

19.1 Maximum Likelihood estimator and risk upper bound ECE598: Information-theoretic methods in high-dimensional statistics Spring 016 Lecture 19: Denoising sparse vectors - Ris upper bound Lecturer: Yihong Wu Scribe: Ravi Kiran Raman, Apr 1, 016 This lecture

More information

Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate

Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate Lucas Janson, Stanford Department of Statistics WADAPT Workshop, NIPS, December 2016 Collaborators: Emmanuel

More information

Causal Inference: Discussion

Causal Inference: Discussion Causal Inference: Discussion Mladen Kolar The University of Chicago Booth School of Business Sept 23, 2016 Types of machine learning problems Based on the information available: Supervised learning Reinforcement

More information

Paper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001)

Paper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001) Paper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001) Presented by Yang Zhao March 5, 2010 1 / 36 Outlines 2 / 36 Motivation

More information

Stat 710: Mathematical Statistics Lecture 31

Stat 710: Mathematical Statistics Lecture 31 Stat 710: Mathematical Statistics Lecture 31 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 31 April 13, 2009 1 / 13 Lecture 31:

More information

STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song

STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song Presenter: Jiwei Zhao Department of Statistics University of Wisconsin Madison April

More information

Analysis of Greedy Algorithms

Analysis of Greedy Algorithms Analysis of Greedy Algorithms Jiahui Shen Florida State University Oct.26th Outline Introduction Regularity condition Analysis on orthogonal matching pursuit Analysis on forward-backward greedy algorithm

More information

Penalized Squared Error and Likelihood: Risk Bounds and Fast Algorithms

Penalized Squared Error and Likelihood: Risk Bounds and Fast Algorithms university-logo Penalized Squared Error and Likelihood: Risk Bounds and Fast Algorithms Andrew Barron Cong Huang Xi Luo Department of Statistics Yale University 2008 Workshop on Sparsity in High Dimensional

More information

Divide-and-combine Strategies in Statistical Modeling for Massive Data

Divide-and-combine Strategies in Statistical Modeling for Massive Data Divide-and-combine Strategies in Statistical Modeling for Massive Data Liqun Yu Washington University in St. Louis March 30, 2017 Liqun Yu (WUSTL) D&C Statistical Modeling for Massive Data March 30, 2017

More information

Message passing and approximate message passing

Message passing and approximate message passing Message passing and approximate message passing Arian Maleki Columbia University 1 / 47 What is the problem? Given pdf µ(x 1, x 2,..., x n ) we are interested in arg maxx1,x 2,...,x n µ(x 1, x 2,..., x

More information

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson Bayesian variable selection via penalized credible regions Brian Reich, NC State Joint work with Howard Bondell and Ander Wilson Brian Reich, NCSU Penalized credible regions 1 Motivation big p, small n

More information

Finite Time Analysis of Vector Autoregressive Models under Linear Restrictions

Finite Time Analysis of Vector Autoregressive Models under Linear Restrictions Finite Time Analysis of Vector Autoregressive Models under Linear Restrictions Yao Zheng collaboration with Guang Cheng Department of Statistics, Purdue University Talk at Math, IUPUI October 23, 2018

More information

Reconstruction from Anisotropic Random Measurements

Reconstruction from Anisotropic Random Measurements Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013

More information

The MNet Estimator. Patrick Breheny. Department of Biostatistics Department of Statistics University of Kentucky. August 2, 2010

The MNet Estimator. Patrick Breheny. Department of Biostatistics Department of Statistics University of Kentucky. August 2, 2010 Department of Biostatistics Department of Statistics University of Kentucky August 2, 2010 Joint work with Jian Huang, Shuangge Ma, and Cun-Hui Zhang Penalized regression methods Penalized methods have

More information

Stochastic Proximal Gradient Algorithm

Stochastic Proximal Gradient Algorithm Stochastic Institut Mines-Télécom / Telecom ParisTech / Laboratoire Traitement et Communication de l Information Joint work with: Y. Atchade, Ann Arbor, USA, G. Fort LTCI/Télécom Paristech and the kind

More information

The deterministic Lasso

The deterministic Lasso The deterministic Lasso Sara van de Geer Seminar für Statistik, ETH Zürich Abstract We study high-dimensional generalized linear models and empirical risk minimization using the Lasso An oracle inequality

More information

Phase Transition Phenomenon in Sparse Approximation

Phase Transition Phenomenon in Sparse Approximation Phase Transition Phenomenon in Sparse Approximation University of Utah/Edinburgh L1 Approximation: May 17 st 2008 Convex polytopes Counting faces Sparse Representations via l 1 Regularization Underdetermined

More information

Peter Hoff Minimax estimation November 12, Motivation and definition. 2 Least favorable prior 3. 3 Least favorable prior sequence 11

Peter Hoff Minimax estimation November 12, Motivation and definition. 2 Least favorable prior 3. 3 Least favorable prior sequence 11 Contents 1 Motivation and definition 1 2 Least favorable prior 3 3 Least favorable prior sequence 11 4 Nonparametric problems 15 5 Minimax and admissibility 18 6 Superefficiency and sparsity 19 Most of

More information

Learning discrete graphical models via generalized inverse covariance matrices

Learning discrete graphical models via generalized inverse covariance matrices Learning discrete graphical models via generalized inverse covariance matrices Duzhe Wang, Yiming Lv, Yongjoon Kim, Young Lee Department of Statistics University of Wisconsin-Madison {dwang282, lv23, ykim676,

More information

Solving Corrupted Quadratic Equations, Provably

Solving Corrupted Quadratic Equations, Provably Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin

More information

TECHNICAL REPORT NO. 1091r. A Note on the Lasso and Related Procedures in Model Selection

TECHNICAL REPORT NO. 1091r. A Note on the Lasso and Related Procedures in Model Selection DEPARTMENT OF STATISTICS University of Wisconsin 1210 West Dayton St. Madison, WI 53706 TECHNICAL REPORT NO. 1091r April 2004, Revised December 2004 A Note on the Lasso and Related Procedures in Model

More information

Lecture 5: January 30

Lecture 5: January 30 CS71 Randomness & Computation Spring 018 Instructor: Alistair Sinclair Lecture 5: January 30 Disclaimer: These notes have not been subjected to the usual scrutiny accorded to formal publications. They

More information

Stepwise Searching for Feature Variables in High-Dimensional Linear Regression

Stepwise Searching for Feature Variables in High-Dimensional Linear Regression Stepwise Searching for Feature Variables in High-Dimensional Linear Regression Qiwei Yao Department of Statistics, London School of Economics q.yao@lse.ac.uk Joint work with: Hongzhi An, Chinese Academy

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

Statistical Ranking Problem

Statistical Ranking Problem Statistical Ranking Problem Tong Zhang Statistics Department, Rutgers University Ranking Problems Rank a set of items and display to users in corresponding order. Two issues: performance on top and dealing

More information

1 Regression with High Dimensional Data

1 Regression with High Dimensional Data 6.883 Learning with Combinatorial Structure ote for Lecture 11 Instructor: Prof. Stefanie Jegelka Scribe: Xuhong Zhang 1 Regression with High Dimensional Data Consider the following regression problem:

More information

Summary and discussion of: Controlling the False Discovery Rate via Knockoffs

Summary and discussion of: Controlling the False Discovery Rate via Knockoffs Summary and discussion of: Controlling the False Discovery Rate via Knockoffs Statistics Journal Club, 36-825 Sangwon Justin Hyun and William Willie Neiswanger 1 Paper Summary 1.1 Quick intuitive summary

More information

high-dimensional inference robust to the lack of model sparsity

high-dimensional inference robust to the lack of model sparsity high-dimensional inference robust to the lack of model sparsity Jelena Bradic (joint with a PhD student Yinchu Zhu) www.jelenabradic.net Assistant Professor Department of Mathematics University of California,

More information

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage Lingrui Gan, Naveen N. Narisetty, Feng Liang Department of Statistics University of Illinois at Urbana-Champaign Problem Statement

More information

Peter Hoff Minimax estimation October 31, Motivation and definition. 2 Least favorable prior 3. 3 Least favorable prior sequence 11

Peter Hoff Minimax estimation October 31, Motivation and definition. 2 Least favorable prior 3. 3 Least favorable prior sequence 11 Contents 1 Motivation and definition 1 2 Least favorable prior 3 3 Least favorable prior sequence 11 4 Nonparametric problems 15 5 Minimax and admissibility 18 6 Superefficiency and sparsity 19 Most of

More information

Lecture 2 Part 1 Optimization

Lecture 2 Part 1 Optimization Lecture 2 Part 1 Optimization (January 16, 2015) Mu Zhu University of Waterloo Need for Optimization E(y x), P(y x) want to go after them first, model some examples last week then, estimate didn t discuss

More information

Elaine T. Hale, Wotao Yin, Yin Zhang

Elaine T. Hale, Wotao Yin, Yin Zhang , Wotao Yin, Yin Zhang Department of Computational and Applied Mathematics Rice University McMaster University, ICCOPT II-MOPTA 2007 August 13, 2007 1 with Noise 2 3 4 1 with Noise 2 3 4 1 with Noise 2

More information

High-dimensional Ordinary Least-squares Projection for Screening Variables

High-dimensional Ordinary Least-squares Projection for Screening Variables 1 / 38 High-dimensional Ordinary Least-squares Projection for Screening Variables Chenlei Leng Joint with Xiangyu Wang (Duke) Conference on Nonparametric Statistics for Big Data and Celebration to Honor

More information

Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise

Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published

More information

Recent developments on sparse representation

Recent developments on sparse representation Recent developments on sparse representation Zeng Tieyong Department of Mathematics, Hong Kong Baptist University Email: zeng@hkbu.edu.hk Hong Kong Baptist University Dec. 8, 2008 First Previous Next Last

More information

Cross-Validation with Confidence

Cross-Validation with Confidence Cross-Validation with Confidence Jing Lei Department of Statistics, Carnegie Mellon University WHOA-PSI Workshop, St Louis, 2017 Quotes from Day 1 and Day 2 Good model or pure model? Occam s razor We really

More information

ORACLE INEQUALITIES AND OPTIMAL INFERENCE UNDER GROUP SPARSITY. By Karim Lounici, Massimiliano Pontil, Sara van de Geer and Alexandre B.

ORACLE INEQUALITIES AND OPTIMAL INFERENCE UNDER GROUP SPARSITY. By Karim Lounici, Massimiliano Pontil, Sara van de Geer and Alexandre B. Submitted to the Annals of Statistics arxiv: 007.77v ORACLE INEQUALITIES AND OPTIMAL INFERENCE UNDER GROUP SPARSITY By Karim Lounici, Massimiliano Pontil, Sara van de Geer and Alexandre B. Tsybakov We

More information

LARGE NUMBERS OF EXPLANATORY VARIABLES. H.S. Battey. WHAO-PSI, St Louis, 9 September 2018

LARGE NUMBERS OF EXPLANATORY VARIABLES. H.S. Battey. WHAO-PSI, St Louis, 9 September 2018 LARGE NUMBERS OF EXPLANATORY VARIABLES HS Battey Department of Mathematics, Imperial College London WHAO-PSI, St Louis, 9 September 2018 Regression, broadly defined Response variable Y i, eg, blood pressure,

More information

New ways of dimension reduction? Cutting data sets into small pieces

New ways of dimension reduction? Cutting data sets into small pieces New ways of dimension reduction? Cutting data sets into small pieces Roman Vershynin University of Michigan, Department of Mathematics Statistical Machine Learning Ann Arbor, June 5, 2012 Joint work with

More information

Introduction to graphical models: Lecture III

Introduction to graphical models: Lecture III Introduction to graphical models: Lecture III Martin Wainwright UC Berkeley Departments of Statistics, and EECS Martin Wainwright (UC Berkeley) Some introductory lectures January 2013 1 / 25 Introduction

More information

Extended Bayesian Information Criteria for Gaussian Graphical Models

Extended Bayesian Information Criteria for Gaussian Graphical Models Extended Bayesian Information Criteria for Gaussian Graphical Models Rina Foygel University of Chicago rina@uchicago.edu Mathias Drton University of Chicago drton@uchicago.edu Abstract Gaussian graphical

More information

21.2 Example 1 : Non-parametric regression in Mean Integrated Square Error Density Estimation (L 2 2 risk)

21.2 Example 1 : Non-parametric regression in Mean Integrated Square Error Density Estimation (L 2 2 risk) 10-704: Information Processing and Learning Spring 2015 Lecture 21: Examples of Lower Bounds and Assouad s Method Lecturer: Akshay Krishnamurthy Scribes: Soumya Batra Note: LaTeX template courtesy of UC

More information

10708 Graphical Models: Homework 2

10708 Graphical Models: Homework 2 10708 Graphical Models: Homework 2 Due Monday, March 18, beginning of class Feburary 27, 2013 Instructions: There are five questions (one for extra credit) on this assignment. There is a problem involves

More information

Sparse PCA in High Dimensions

Sparse PCA in High Dimensions Sparse PCA in High Dimensions Jing Lei, Department of Statistics, Carnegie Mellon Workshop on Big Data and Differential Privacy Simons Institute, Dec, 2013 (Based on joint work with V. Q. Vu, J. Cho, and

More information

(Part 1) High-dimensional statistics May / 41

(Part 1) High-dimensional statistics May / 41 Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2

More information

Probabilistic Graphical Models

Probabilistic Graphical Models School of Computer Science Probabilistic Graphical Models Gaussian graphical models and Ising models: modeling networks Eric Xing Lecture 0, February 7, 04 Reading: See class website Eric Xing @ CMU, 005-04

More information

Signal Recovery from Permuted Observations

Signal Recovery from Permuted Observations EE381V Course Project Signal Recovery from Permuted Observations 1 Problem Shanshan Wu (sw33323) May 8th, 2015 We start with the following problem: let s R n be an unknown n-dimensional real-valued signal,

More information

Motivation Sparse Signal Recovery is an interesting area with many potential applications. Methods developed for solving sparse signal recovery proble

Motivation Sparse Signal Recovery is an interesting area with many potential applications. Methods developed for solving sparse signal recovery proble Bayesian Methods for Sparse Signal Recovery Bhaskar D Rao 1 University of California, San Diego 1 Thanks to David Wipf, Zhilin Zhang and Ritwik Giri Motivation Sparse Signal Recovery is an interesting

More information

Bayesian Methods for Sparse Signal Recovery

Bayesian Methods for Sparse Signal Recovery Bayesian Methods for Sparse Signal Recovery Bhaskar D Rao 1 University of California, San Diego 1 Thanks to David Wipf, Jason Palmer, Zhilin Zhang and Ritwik Giri Motivation Motivation Sparse Signal Recovery

More information

Cross-Validation with Confidence

Cross-Validation with Confidence Cross-Validation with Confidence Jing Lei Department of Statistics, Carnegie Mellon University UMN Statistics Seminar, Mar 30, 2017 Overview Parameter est. Model selection Point est. MLE, M-est.,... Cross-validation

More information

Post-selection Inference for Forward Stepwise and Least Angle Regression

Post-selection Inference for Forward Stepwise and Least Angle Regression Post-selection Inference for Forward Stepwise and Least Angle Regression Ryan & Rob Tibshirani Carnegie Mellon University & Stanford University Joint work with Jonathon Taylor, Richard Lockhart September

More information

arxiv: v3 [stat.me] 8 Jun 2018

arxiv: v3 [stat.me] 8 Jun 2018 Between hard and soft thresholding: optimal iterative thresholding algorithms Haoyang Liu and Rina Foygel Barber arxiv:804.0884v3 [stat.me] 8 Jun 08 June, 08 Abstract Iterative thresholding algorithms

More information

Predicting Workplace Incidents with Temporal Graph-guided Fused Lasso

Predicting Workplace Incidents with Temporal Graph-guided Fused Lasso Predicting Workplace Incidents with Temporal Graph-guided Fused Lasso Keerthiram Murugesan 1 and Jaime Carbonell 1 1 Language Technologies Institute Carnegie Mellon University Pittsburgh, USA CMU-LTI-15-??

More information

The adaptive and the thresholded Lasso for potentially misspecified models (and a lower bound for the Lasso)

The adaptive and the thresholded Lasso for potentially misspecified models (and a lower bound for the Lasso) Electronic Journal of Statistics Vol. 0 (2010) ISSN: 1935-7524 The adaptive the thresholded Lasso for potentially misspecified models ( a lower bound for the Lasso) Sara van de Geer Peter Bühlmann Seminar

More information

Statistical Measures of Uncertainty in Inverse Problems

Statistical Measures of Uncertainty in Inverse Problems Statistical Measures of Uncertainty in Inverse Problems Workshop on Uncertainty in Inverse Problems Institute for Mathematics and Its Applications Minneapolis, MN 19-26 April 2002 P.B. Stark Department

More information

An Introduction to Sparse Approximation

An Introduction to Sparse Approximation An Introduction to Sparse Approximation Anna C. Gilbert Department of Mathematics University of Michigan Basic image/signal/data compression: transform coding Approximate signals sparsely Compress images,

More information

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,

More information

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit New Coherence and RIP Analysis for Wea 1 Orthogonal Matching Pursuit Mingrui Yang, Member, IEEE, and Fran de Hoog arxiv:1405.3354v1 [cs.it] 14 May 2014 Abstract In this paper we define a new coherence

More information

Coordinate descent. Geoff Gordon & Ryan Tibshirani Optimization /

Coordinate descent. Geoff Gordon & Ryan Tibshirani Optimization / Coordinate descent Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Adding to the toolbox, with stats and ML in mind We ve seen several general and useful minimization tools First-order methods

More information

Bi-level feature selection with applications to genetic association

Bi-level feature selection with applications to genetic association Bi-level feature selection with applications to genetic association studies October 15, 2008 Motivation In many applications, biological features possess a grouping structure Categorical variables may

More information

SPARSE signal representations have gained popularity in recent

SPARSE signal representations have gained popularity in recent 6958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 10, OCTOBER 2011 Blind Compressed Sensing Sivan Gleichman and Yonina C. Eldar, Senior Member, IEEE Abstract The fundamental principle underlying

More information

Does l p -minimization outperform l 1 -minimization?

Does l p -minimization outperform l 1 -minimization? Does l p -minimization outperform l -minimization? Le Zheng, Arian Maleki, Haolei Weng, Xiaodong Wang 3, Teng Long Abstract arxiv:50.03704v [cs.it] 0 Jun 06 In many application areas ranging from bioinformatics

More information

Interpreting Regression Results

Interpreting Regression Results Interpreting Regression Results Carlo Favero Favero () Interpreting Regression Results 1 / 42 Interpreting Regression Results Interpreting regression results is not a simple exercise. We propose to split

More information

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming High Dimensional Inverse Covariate Matrix Estimation via Linear Programming Ming Yuan October 24, 2011 Gaussian Graphical Model X = (X 1,..., X p ) indep. N(µ, Σ) Inverse covariance matrix Σ 1 = Ω = (ω

More information

STA414/2104 Statistical Methods for Machine Learning II

STA414/2104 Statistical Methods for Machine Learning II STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements

More information

Variable Selection for Highly Correlated Predictors

Variable Selection for Highly Correlated Predictors Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu arxiv:1709.04840v1 [stat.me] 14 Sep 2017 Abstract Penalty-based variable selection methods are powerful in selecting relevant covariates

More information