MS&E 318 (CME 338) Large-Scale Numerical Optimization. A Lasso Solver
|
|
- Helen Poole
- 5 years ago
- Views:
Transcription
1 Stanford University, Dept of Management Science and Engineering MS&E 318 (CME 338) Large-Scale Numerical Optimization Instructor: Michael Saunders Spring 2011 Final Project Due Friday June 10 A Lasso Solver We follow the terminology of van den Berg and Friedlander [7] in defining three parameterized problems arising in the estimation of linear models A b (where A is m n). The problems are equivalent in the sense that if one of them has its tuning parameter specified, the optimal solution for the other two problems will be the same if those problems are solved with certain parameter values (though it is nontrivial to find the matching parameter values). The first problem was defined by Tibshirani (1996) [6] to be the Lasso problem (least absolute shrinkage and selection operator): LS τ minimize A b 2 2 subject to 1 τ, where smaller values of the tuning parameter τ t to shrink more components of to be eactly zero. Models with m > n were probably envisaged at the time, although the Lasso problem is well defined for any m and n if 0 τ <. For signal analysis problems in which A was likely to be a fast operator, Chen, Donoho, and Saunders (1998) [2, 3] defined the parameter-free Basis Pursuit (BP) problem min 1 subject to A = b (BP) and also the Basis Pursuit Denoising problem (BPDN), which is a conve quadratic optimization problem with tuning parameter λ: QP λ minimize A b λ 1 subject to no constraints. Models with m n were envisaged for BP and BPDN, although problem QP λ is well defined for any m and n if 0 < λ. Finally, van den Berg and Friedlander [7, 5] regard the following problem as a more practical formulation of BPDN: BP σ minimize 1 subject to A b 2 2 σ2, because σ is more likely to be known than λ, and σ = 0 gives the BP problem naturally (whereas QP λ needs λ 0, where it is not well defined unless A has full column rank). Problem BP σ is well defined for any m and n for all σ 0.
2 2 MS&E 318 (CME 338) Large-Scale Numerical Optimization At the time Lasso was proposed, good solvers were not known for problem LS τ. The BPDN solvers in [2, 3] could handle problem QP λ with a given value of λ, allowing A to be a large fast operator (typically an over-complete dictionary composed of wavelets, chirplets, warplets, curvelets, etc). Efron, Hastie, Johnstone, and Tibshirani (2004) [4] later developed the LARS algorithm, a version of which can solve problem QP λ for all 0 < λ. In principle, LARS can be used to solve LS τ and BP σ with any specific τ or σ. Later, van den Berg and Friedlander (2008) [7, 5] developed algorithm SPGL1 to solve LS τ with a given τ (allowing A to be a fast operator). They include a rootfinding algorithm to solve problem BP σ with a specified σ 0 (by solving a sequence of Lasso problems LS τ ). They further generalized SPGL1 to handle comple data, group sparsity, and multiple measurement vectors (MMV); see [5]. Although SPGL1 is a highly versatile solver, it should be of interest to develop a conventional active-set solver for the Lasso problem, using established algorithms for the classical problem of nonnegative least squares (NNLS). Our project is a step in this direction. Smooth Formulations Note that the Lasso problem can be solved as the linearly constrained least-squares problem lasso τ 1 minimize v, w 2 Av Aw b 2 2 subject to e T v + e T w τ, v, w 0, where e is a vector of 1s and = v w. nonnegative Lasso problem For simplicity, we first consider the nnlasso τ 1 minimize 2 A b 2 2 subject to e T τ, 0. Keep in mind that an algorithm for solving nnlasso τ can easily be used to solve lasso τ. The Lasso problem is not very interesting unless the constraint 1 τ is binding. Thus our strategy for nnlasso τ is to assume that e T = τ. We can then apply the reduced-gradient method (an active-set method). There will be a single basic variable, and the remaining variables will be superbasic or nonbasic. If the basic variable is chosen fortuitously it will never be knocked out of the basis (it will stay strictly positive). Also, if τ is small enough there will not be many superbasic variables. Most variables will remain nonbasic at zero. Project Resources We will use Matlab and/or Tomlab to solve nnlasso τ directly using algorithms in the optim toolbo or elsewhere. The files mentioned below can be downloaded from
3 Spring 2011 Final Project: a Lasso Solver 3 Project Tasks 1. File lassoproblem.m generates some typical data (A, b). Use it to generate an eample nnlasso τ problem (type=0) and a sensible value for τ. In general we trust that the norms of each column of A will be roughly equal. Since A is eplicit and dense, we can compute these reasonably efficiently. Find a vectorized way to do this (no for loops!). Report the largest and smallest norm. 2. File spgl1problem.m generates slightly different data (A, b). Do the same for the problem suggested in the help info. 3. File nnlasso.m knows about some of the solvers available in Matlab and Tomlab. It also knows how to start Tomlab on the campus Linu cluster. Give a brief description of the solvers lsqlin and lssol. Are they suitable for problem nnlasso τ? 4. Set the Matlab and Tomlab paths inside nnlasso.m and solve the problems in Q1 and Q2 using matlab/lsqlin, tomlab/lsqlin and tomlab/lssol. Report your results and any differences that you notice. 5. In order to develop our new solver for nnlasso τ, we need a good choice for the basic variable in the reduced-gradient algorithm. Let s suppose that eactly one variable is positive. By studying r A b for each of n possible 1-variable solutions, deduce a reasonable method for choosing a good basic variable. (This method is used in nnlasso.m. We need to justify it.) 6. Suppose the chosen variable is B and the remaining variables are. We can use the constraint e T = τ to eliminate the basic variable. If we ignore the implicit constraint B 0, derive a problem that should solve. (It is an NNLS problem. We need to describe it algebraically.) 7. [,inform] = nnlasso(a,b,tau, matlab/lsqnonneg ); or [,inform] = nnlasso(a,b,tau, tomlab/lsqnonneg ); solves the nnlasso τ problem. Run either or both of these and report your findings. 8. nnlasso warns us if B has gone negative. Suggest a strategy for choosing a different basic variable from the solution obtained so far. (Bear in mind that a good NNLS solver would be able to warm start from the active set that you construct.) 9. Michael Friedlander s solver BCLS [1] is designed for bound-constrained leastsquares problems. Study the BCLS website and describe the features that would make BCLS a good solver for implementing nnlasso τ. 10. Given our solver for nnlasso τ, use it to construct a solver for the original problem lasso τ. Test it using the problem generators with type=1.
4 4 MS&E 318 (CME 338) Large-Scale Numerical Optimization function [A,b,1norm] = lassoproblem(m,n,k,type) [A,b,1norm] = lassoproblem(m,n,k,type); generates data for two types of Lasso problem: type=0: min 1/2 A - b _2^2 st sum() <= tau, >= 0; type=1: min 1/2 A - b _2^2 st norm(,1) <= tau, where A is size m n and there would be k nonzeros in if tau = infinity. We assume m>0, n>0, 1<=k<=n. It doesn t matter if m<=n or m>= n. Eample: matlab get into Matlab cd <projectlasso> your own directory m = 10; n = 50; k = 5; eample dimensions [A,b,1norm] = lassoproblem(m,n,k,0); 0 => nonneg Lasso or [A,b,1norm] = lassoproblem(m,n,k,1); 1 => normal Lasso tau = 0.5*1norm; smaller than 1norm 19 May 2011: First version for class project. Michael Saunders, MS&E318/CME338 instructor, Stanford University. rand ( state,0); Initialize the good old way randn( state,0); (easy to remember) p = randperm(n); p = p(1:k); Position of nonzeros in = zeros(n,1); Generate sparse solution k = randn(k,1); switch type case 0 (p) = abs(k); >= 0 case 1 (p) = k; positive or negative otherwise error( type must be 0 or 1 ) A = randn(m,n); Gaussian m-by-n ensemble b = A*; The rhs vector 1norm = norm(k,1);
5 Spring 2011 Final Project: a Lasso Solver 5 function [,inform] = nnlasso(a,b,tau,solver) [,inform] = nnlasso(a,b,tau,solver); solves the nonnegative Lasso problem min 1/2 A - b _2^2 s.t. sum() <= tau, >= 0. where tau > 0 and "solver" is a string that specifies the solver to be used. Eamples: [,inform] = nnlasso(a,b,tau, matlab/lsqlin ); or [,inform] = nnlasso(a,b,tau, tomlab/lsqlin ); or [,inform] = nnlasso(a,b,tau, tomlab/lssol ); or [,inform] = nnlasso(a,b,tau, matlab/lsqnonneg ); or [,inform] = nnlasso(a,b,tau, tomlab/lsqnonneg ); The first 3 eamples solve the nonnegative Lasso problem directly. The last 2 eamples convert it to a nonnegative least-squares problem. Before running nnlasso, set matlabpath and tomlabpath inside nnlasso.m. 22 May 2011: First version for class project. Michael Saunders, MS&E318/CME338 instructor, Stanford University. Select 2 of the net 4 lines: matlabpath = /afs/ir/software/matlab-2010b ; on Linu cluster tomlabpath = /afs/ir/software/matlab-2010b/tomlab ; matlabpath = /Applications/MATLAB_R2009b.app ; on Michael s imac tomlabpath = /Applications/tomlab ; Save wd, A, b, and size of problem wd = cd; AA = A; bb = b; [m,n] = size(a); lab = solver(1:6); matlab or tomlab alg = solver(8:); lsqlin, lssol, or lsqnonneg Start TOMLAB if necessary if lab== tomlab & ~eist( tomlab ) cd(tomlabpath); startup Set path to solver switch solver case matlab/lsqlin ; cd([matlabpath /toolbo/optim/optim ]); case tomlab/lsqlin ; cd([tomlabpath /optim ]); case tomlab/lssol ; cd([tomlabpath /me ]); case matlab/lsqnonneg ; cd([matlabpath /toolbo/matlab/matfun ]); case tomlab/lsqnonneg ; cd([tomlabpath /optim ]); otherwise error( Unknown solver )
6 6 MS&E 318 (CME 338) Large-Scale Numerical Optimization Setup inputs for the specified algorithm, then solve the problem switch alg case lsqlin Solve nonnegative Lasso directly C = A; d = b; A = ones(1,n); b = tau; Aeq = []; beq = []; bl = zeros(n,1); bu = inf(n,1); 0 = zeros(n,1); [,resnorm,residual,eitflag,output,lambda] = lsqlin(c,d,a,b,aeq,beq,bl,bu,0); inform = eitflag; output case lssol Solve nonnegative Lasso directly H = A; A = ones(1,n); bl = [zeros(n,1); 0]; bu = [ inf(n,1); tau]; c = []; zeros(n,1); = zeros(n,1); optpar = []; [,inform,istate,clamda,iter,fobj,r,k] = lssol(a,bl,bu,c,,optpar,h,b); case lsqnonneg Reduce the problem to nonnegative least squares by choosing a basic variable (jb) and eliminating it via (jb) = tau - sum((j),j~=jb) (assuming the linear constraint is binding) [rb,jb] = ma(a *b); ab = A(:,jB); C = A; C(:,jB) = []; C = C - ab*ones(1,n-1); d = b - ab*tau; 0 = zeros(n-1,1); [S,resnorm,residual,eitflag] = lsqnonneg(c,d); inform = eitflag; B = tau - sum(s); Evaluate the basic variable js = 1:n; Epand B,S into full js(jb) = []; = zeros(n,1); (jb) = B; (js) = S; fprintf( \n The basic variable is ) fprintf( \n 7i 15.7e, jb,b) if B <= -1e-12 fprintf( \n WARNING: it has the wrong sign ) inform = -1; otherwise error( Unknown algorithm )
7 Spring 2011 Final Project: a Lasso Solver 7 Restore wd and output a few things about the solution cd(wd); 1norm = norm(,1); r2norm = norm(bb - AA*); fprintf( \n ) fprintf( \n Solution obtained by s, solver) fprintf( \n _1 = 13.7e, 1norm) fprintf( \n r _2 = 13.7e, r2norm) fprintf( \n\n Nonzero ) K = find( > 1e-12) ; for j = K fprintf( \n 7i 15.7e, j,(j)) fprintf( \n ) References [1] BCLS solver for bound-constrained least-squares. mpf/bcls. [2] S. S. Chen, D. L. Donoho, and M. A. Saunders. Atomic decomposition by basis pursuit. SIAM J. Sci. Comput., 20(1):33 61, [3] S. S. Chen, D. L. Donoho, and M. A. Saunders. Atomic decomposition by basis pursuit. SIAM Review, 43(1): , SIGEST article. [4] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani. Least angle regression. Ann. Statist., 32(2): , [5] SPGL1 solver for large-scale sparse reconstruction. [6] R. Tibshirani. Regression shrinkage and selection via the lasso. J. Roy. Statist. Soc. Ser. B, 58(1): , [7] E. van den Berg and M. P. Friedlander. Probing the Pareto frontier for basis pursuit solutions. SIAM J. Sci. Comput., 31(2): , 2008.
Basis Pursuit Denoising and the Dantzig Selector
BPDN and DS p. 1/16 Basis Pursuit Denoising and the Dantzig Selector West Coast Optimization Meeting University of Washington Seattle, WA, April 28 29, 2007 Michael Friedlander and Michael Saunders Dept
More informationMaking Flippy Floppy
Making Flippy Floppy James V. Burke UW Mathematics jvburke@uw.edu Aleksandr Y. Aravkin IBM, T.J.Watson Research sasha.aravkin@gmail.com Michael P. Friedlander UBC Computer Science mpf@cs.ubc.ca Current
More informationMaking Flippy Floppy
Making Flippy Floppy James V. Burke UW Mathematics jvburke@uw.edu Aleksandr Y. Aravkin IBM, T.J.Watson Research sasha.aravkin@gmail.com Michael P. Friedlander UBC Computer Science mpf@cs.ubc.ca Vietnam
More informationComputing approximate PageRank vectors by Basis Pursuit Denoising
Computing approximate PageRank vectors by Basis Pursuit Denoising Michael Saunders Systems Optimization Laboratory, Stanford University Joint work with Holly Jin, LinkedIn Corp SIAM Annual Meeting San
More informationSparse Optimization Lecture: Basic Sparse Optimization Models
Sparse Optimization Lecture: Basic Sparse Optimization Models Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know basic l 1, l 2,1, and nuclear-norm
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization 1 Origins Instructor: Michael Saunders Spring 2015 Notes 9: Augmented Lagrangian Methods
More information1-Bit Compressive Sensing
1-Bit Compressive Sensing Petros T. Boufounos, Richard G. Baraniuk Rice University, Electrical and Computer Engineering 61 Main St. MS 38, Houston, TX 775 Abstract Compressive sensing is a new signal acquisition
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization. Generalized Least Squares
Stanford University, Dept of Management Science and Engineering MS&E 318 (CME 338) Large-Scale Numerical Optimization Instructor: Michael Saunders Spring 2010 Project Due Wednesday June 9 Generalized Least
More informationPathwise coordinate optimization
Stanford University 1 Pathwise coordinate optimization Jerome Friedman, Trevor Hastie, Holger Hoefling, Robert Tibshirani Stanford University Acknowledgements: Thanks to Stephen Boyd, Michael Saunders,
More informationFWI with Compressive Updates Aleksandr Aravkin, Felix Herrmann, Tristan van Leeuwen, Xiang Li, James Burke
Consortium 2010 FWI with Compressive Updates Aleksandr Aravkin, Felix Herrmann, Tristan van Leeuwen, Xiang Li, James Burke SLIM University of British Columbia Full Waveform Inversion The Full Waveform
More informationSparsity in Underdetermined Systems
Sparsity in Underdetermined Systems Department of Statistics Stanford University August 19, 2005 Classical Linear Regression Problem X n y p n 1 > Given predictors and response, y Xβ ε = + ε N( 0, σ 2
More informationECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference
ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Sparse Recovery using L1 minimization - algorithms Yuejie Chi Department of Electrical and Computer Engineering Spring
More informationOptimal Value Function Methods in Numerical Optimization Level Set Methods
Optimal Value Function Methods in Numerical Optimization Level Set Methods James V Burke Mathematics, University of Washington, (jvburke@uw.edu) Joint work with Aravkin (UW), Drusvyatskiy (UW), Friedlander
More informationMachine Learning for Signal Processing Sparse and Overcomplete Representations. Bhiksha Raj (slides from Sourish Chaudhuri) Oct 22, 2013
Machine Learning for Signal Processing Sparse and Overcomplete Representations Bhiksha Raj (slides from Sourish Chaudhuri) Oct 22, 2013 1 Key Topics in this Lecture Basics Component-based representations
More informationLasso: Algorithms and Extensions
ELE 538B: Sparsity, Structure and Inference Lasso: Algorithms and Extensions Yuxin Chen Princeton University, Spring 2017 Outline Proximal operators Proximal gradient methods for lasso and its extensions
More informationSparse regression. Optimization-Based Data Analysis. Carlos Fernandez-Granda
Sparse regression Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 3/28/2016 Regression Least-squares regression Example: Global warming Logistic
More informationSaharon Rosset 1 and Ji Zhu 2
Aust. N. Z. J. Stat. 46(3), 2004, 505 510 CORRECTED PROOF OF THE RESULT OF A PREDICTION ERROR PROPERTY OF THE LASSO ESTIMATOR AND ITS GENERALIZATION BY HUANG (2003) Saharon Rosset 1 and Ji Zhu 2 IBM T.J.
More informationGeometric Modeling Summer Semester 2010 Mathematical Tools (1)
Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric
More informationAn Homotopy Algorithm for the Lasso with Online Observations
An Homotopy Algorithm for the Lasso with Online Observations Pierre J. Garrigues Department of EECS Redwood Center for Theoretical Neuroscience University of California Berkeley, CA 94720 garrigue@eecs.berkeley.edu
More informationSparsity-promoting migration with multiples
Sparsity-promoting migration with multiples Tim Lin, Ning Tu and Felix Herrmann SLIM Seismic Laboratory for Imaging and Modeling the University of British Columbia Courtesy of Verschuur, 29 SLIM Motivation..
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME MS&E 38 (CME 338 Large-Scale Numerical Optimization Course description Instructor: Michael Saunders Spring 28 Notes : Review The course teaches
More informationA robust hybrid of lasso and ridge regression
A robust hybrid of lasso and ridge regression Art B. Owen Stanford University October 2006 Abstract Ridge regression and the lasso are regularized versions of least squares regression using L 2 and L 1
More informationConvex Optimization Problems. Prof. Daniel P. Palomar
Conve Optimization Problems Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics Fall 2018-19, HKUST,
More informationMachine Learning for Signal Processing Sparse and Overcomplete Representations
Machine Learning for Signal Processing Sparse and Overcomplete Representations Abelino Jimenez (slides from Bhiksha Raj and Sourish Chaudhuri) Oct 1, 217 1 So far Weights Data Basis Data Independent ICA
More informationConsider the following example of a linear system:
LINEAR SYSTEMS Consider the following example of a linear system: Its unique solution is x + 2x 2 + 3x 3 = 5 x + x 3 = 3 3x + x 2 + 3x 3 = 3 x =, x 2 = 0, x 3 = 2 In general we want to solve n equations
More informationSparse & Redundant Signal Representation, and its Role in Image Processing
Sparse & Redundant Signal Representation, and its Role in Michael Elad The CS Department The Technion Israel Institute of technology Haifa 3000, Israel Wave 006 Wavelet and Applications Ecole Polytechnique
More informationTractable Upper Bounds on the Restricted Isometry Constant
Tractable Upper Bounds on the Restricted Isometry Constant Alex d Aspremont, Francis Bach, Laurent El Ghaoui Princeton University, École Normale Supérieure, U.C. Berkeley. Support from NSF, DHS and Google.
More informationOn Algorithms for Solving Least Squares Problems under an L 1 Penalty or an L 1 Constraint
On Algorithms for Solving Least Squares Problems under an L 1 Penalty or an L 1 Constraint B.A. Turlach School of Mathematics and Statistics (M19) The University of Western Australia 35 Stirling Highway,
More informationAccelerated large-scale inversion with message passing Felix J. Herrmann, the University of British Columbia, Canada
Accelerated large-scale inversion with message passing Felix J. Herrmann, the University of British Columbia, Canada SUMMARY To meet current-day challenges, exploration seismology increasingly relies on
More informationMultiple Change Point Detection by Sparse Parameter Estimation
Multiple Change Point Detection by Sparse Parameter Estimation Department of Econometrics Fac. of Economics and Management University of Defence Brno, Czech Republic Dept. of Appl. Math. and Comp. Sci.
More informationSparsity Regularization
Sparsity Regularization Bangti Jin Course Inverse Problems & Imaging 1 / 41 Outline 1 Motivation: sparsity? 2 Mathematical preliminaries 3 l 1 solvers 2 / 41 problem setup finite-dimensional formulation
More informationOrthogonal Matching Pursuit for Sparse Signal Recovery With Noise
Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization Instructor: Michael Saunders Spring 2017 Notes 10: MINOS Part 1: the Reduced-Gradient
More informationDepartment of Computer Science, University of British Columbia Technical Report TR , January 2008
Department of Computer Science, University of British Columbia Technical Report TR-2008-01, January 2008 PROBING THE PARETO FRONTIER FOR BASIS PURSUIT SOLUTIONS EWOUT VAN DEN BERG AND MICHAEL P. FRIEDLANDER
More informationHomotopy methods based on l 0 norm for the compressed sensing problem
Homotopy methods based on l 0 norm for the compressed sensing problem Wenxing Zhu, Zhengshan Dong Center for Discrete Mathematics and Theoretical Computer Science, Fuzhou University, Fuzhou 350108, China
More information10-725/36-725: Convex Optimization Spring Lecture 21: April 6
10-725/36-725: Conve Optimization Spring 2015 Lecturer: Ryan Tibshirani Lecture 21: April 6 Scribes: Chiqun Zhang, Hanqi Cheng, Waleed Ammar Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:
More informationA complex version of the LASSO algorithm and its application to beamforming
The 7 th International Telecommunications Symposium ITS 21) A complex version of the LASSO algorithm and its application to beamforming José F. de Andrade Jr. and Marcello L. R. de Campos Program of Electrical
More informationSeismic data interpolation and denoising using SVD-free low-rank matrix factorization
Seismic data interpolation and denoising using SVD-free low-rank matrix factorization R. Kumar, A.Y. Aravkin,, H. Mansour,, B. Recht and F.J. Herrmann Dept. of Earth and Ocean sciences, University of British
More informationc 2008 Society for Industrial and Applied Mathematics
SIAM J. SCI. COMPUT. Vol. 31, No. 2, pp. 890 912 c 2008 Society for Industrial and Applied Mathematics PROBING THE PARETO FRONTIER FOR BASIS PURSUIT SOLUTIONS EWOUT VAN DEN BERG AND MICHAEL P. FRIEDLANDER
More informationITERATED SRINKAGE ALGORITHM FOR BASIS PURSUIT MINIMIZATION
ITERATED SRINKAGE ALGORITHM FOR BASIS PURSUIT MINIMIZATION Michael Elad The Computer Science Department The Technion Israel Institute o technology Haia 3000, Israel * SIAM Conerence on Imaging Science
More informationThe Risk of James Stein and Lasso Shrinkage
Econometric Reviews ISSN: 0747-4938 (Print) 1532-4168 (Online) Journal homepage: http://tandfonline.com/loi/lecr20 The Risk of James Stein and Lasso Shrinkage Bruce E. Hansen To cite this article: Bruce
More informationParNes: A rapidly convergent algorithm for accurate recovery of sparse and approximately sparse signals
Preprint manuscript No. (will be inserted by the editor) ParNes: A rapidly convergent algorithm for accurate recovery of sparse and approximately sparse signals Ming Gu ek-heng im Cinna Julie Wu Received:
More informationTECHNICAL REPORT NO. 1091r. A Note on the Lasso and Related Procedures in Model Selection
DEPARTMENT OF STATISTICS University of Wisconsin 1210 West Dayton St. Madison, WI 53706 TECHNICAL REPORT NO. 1091r April 2004, Revised December 2004 A Note on the Lasso and Related Procedures in Model
More informationOptimization Algorithms for Compressed Sensing
Optimization Algorithms for Compressed Sensing Stephen Wright University of Wisconsin-Madison SIAM Gator Student Conference, Gainesville, March 2009 Stephen Wright (UW-Madison) Optimization and Compressed
More informationGetting Started with Communications Engineering
1 Linear algebra is the algebra of linear equations: the term linear being used in the same sense as in linear functions, such as: which is the equation of a straight line. y ax c (0.1) Of course, if we
More informationPrediction & Feature Selection in GLM
Tarigan Statistical Consulting & Coaching statistical-coaching.ch Doctoral Program in Computer Science of the Universities of Fribourg, Geneva, Lausanne, Neuchâtel, Bern and the EPFL Hands-on Data Analysis
More informationApplications with l 1 -Norm Objective Terms
Applications with l 1 -Norm Objective Terms Stephen Wright University of Wisconsin-Madison Huatulco, January 2007 Stephen Wright (UW-Madison) Applications with l 1 -Norm Objective Terms Huatulco, January
More informationData Sparse Matrix Computation - Lecture 20
Data Sparse Matrix Computation - Lecture 20 Yao Cheng, Dongping Qi, Tianyi Shi November 9, 207 Contents Introduction 2 Theorems on Sparsity 2. Example: A = [Φ Ψ]......................... 2.2 General Matrix
More informationExploiting Sparsity for Wireless Communications
Exploiting Sparsity for Wireless Communications Georgios B. Giannakis Dept. of ECE, Univ. of Minnesota http://spincom.ece.umn.edu Acknowledgements: D. Angelosante, J.-A. Bazerque, H. Zhu; and NSF grants
More informationBackpropagation Rules for Sparse Coding (Task-Driven Dictionary Learning)
Backpropagation Rules for Sparse Coding (Task-Driven Dictionary Learning) Julien Mairal UC Berkeley Edinburgh, ICML, June 2012 Julien Mairal, UC Berkeley Backpropagation Rules for Sparse Coding 1/57 Other
More informationMessage Passing Algorithms for Compressed Sensing: II. Analysis and Validation
Message Passing Algorithms for Compressed Sensing: II. Analysis and Validation David L. Donoho Department of Statistics Arian Maleki Department of Electrical Engineering Andrea Montanari Department of
More informationLinear Model Selection and Regularization
Linear Model Selection and Regularization Recall the linear model Y = β 0 + β 1 X 1 + + β p X p + ɛ. In the lectures that follow, we consider some approaches for extending the linear model framework. In
More informationStatistical Geometry Processing Winter Semester 2011/2012
Statistical Geometry Processing Winter Semester 2011/2012 Linear Algebra, Function Spaces & Inverse Problems Vector and Function Spaces 3 Vectors vectors are arrows in space classically: 2 or 3 dim. Euclidian
More informationThe lasso, persistence, and cross-validation
The lasso, persistence, and cross-validation Daniel J. McDonald Department of Statistics Indiana University http://www.stat.cmu.edu/ danielmc Joint work with: Darren Homrighausen Colorado State University
More informationThe problem is: given an n 1 vector y and an n p matrix X, find the b(λ) that minimizes
LARS/lasso 1 Statistics 312/612, fall 21 Homework # 9 Due: Monday 29 November The questions on this sheet all refer to (a slight modification of) the lasso modification of the LARS algorithm, based on
More informationCompressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles
Or: the equation Ax = b, revisited University of California, Los Angeles Mahler Lecture Series Acquiring signals Many types of real-world signals (e.g. sound, images, video) can be viewed as an n-dimensional
More informationFast Hard Thresholding with Nesterov s Gradient Method
Fast Hard Thresholding with Nesterov s Gradient Method Volkan Cevher Idiap Research Institute Ecole Polytechnique Federale de ausanne volkan.cevher@epfl.ch Sina Jafarpour Department of Computer Science
More information6. Regularized linear regression
Foundations of Machine Learning École Centrale Paris Fall 2015 6. Regularized linear regression Chloé-Agathe Azencot Centre for Computational Biology, Mines ParisTech chloe agathe.azencott@mines paristech.fr
More informationSparse Recovery of Streaming Signals Using. M. Salman Asif and Justin Romberg. Abstract
Sparse Recovery of Streaming Signals Using l 1 -Homotopy 1 M. Salman Asif and Justin Romberg arxiv:136.3331v1 [cs.it] 14 Jun 213 Abstract Most of the existing methods for sparse signal recovery assume
More informationFast Regularization Paths via Coordinate Descent
KDD August 2008 Trevor Hastie, Stanford Statistics 1 Fast Regularization Paths via Coordinate Descent Trevor Hastie Stanford University joint work with Jerry Friedman and Rob Tibshirani. KDD August 2008
More informationLinear Equations in Linear Algebra
1 Linear Equations in Linear Algebra 1.1 SYSTEMS OF LINEAR EQUATIONS LINEAR EQUATION,, 1 n A linear equation in the variables equation that can be written in the form a a a b 1 1 2 2 n n a a is an where
More informationRecent developments on sparse representation
Recent developments on sparse representation Zeng Tieyong Department of Mathematics, Hong Kong Baptist University Email: zeng@hkbu.edu.hk Hong Kong Baptist University Dec. 8, 2008 First Previous Next Last
More informationThe lasso. Patrick Breheny. February 15. The lasso Convex optimization Soft thresholding
Patrick Breheny February 15 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/24 Introduction Last week, we introduced penalized regression and discussed ridge regression, in which the penalty
More informationCompressive Phase Retrieval From Squared Output Measurements Via Semidefinite Programming
Compressive Phase Retrieval From Squared Output Measurements Via Semidefinite Programming Henrik Ohlsson, Allen Y. Yang Roy Dong S. Shankar Sastry Department of Electrical Engineering and Computer Sciences,
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu
More informationarxiv:cs/ v1 [cs.ms] 13 Aug 2004
tsnnls: A solver for large sparse least squares problems with non-negative variables Jason Cantarella Department of Mathematics, University of Georgia, Athens, GA 30602 Michael Piatek arxiv:cs/0408029v
More informationMachine Learning for OR & FE
Machine Learning for OR & FE Regression II: Regularization and Shrinkage Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com
More informationA Comparison between LARS and LASSO for Initialising the Time-Series Forecasting Auto-Regressive Equations
Available online at www.sciencedirect.com Procedia Technology 7 ( 2013 ) 282 288 2013 Iberoamerican Conference on Electronics Engineering and Computer Science A Comparison between LARS and LASSO for Initialising
More informationwhere κ is a gauge function. The majority of the solvers available for the problems under consideration work with the penalized version,
SPARSE OPTIMIZATION WITH LEAST-SQUARES CONSTRAINTS ON-LINE APPENDIX EWOUT VAN DEN BERG AND MICHAEL P. FRIEDLANDER Appendix A. Numerical experiments. This on-line appendix accompanies the paper Sparse optimization
More informationLASSO. Step. My modified algorithm getlasso() produces the same sequence of coefficients as the R function lars() when run on the diabetes data set:
lars/lasso 1 Homework 9 stepped you through the lasso modification of the LARS algorithm, based on the papers by Efron, Hastie, Johnstone, and Tibshirani (24) (= the LARS paper) and by Rosset and Zhu (27).
More informationORIE 4741: Learning with Big Messy Data. Regularization
ORIE 4741: Learning with Big Messy Data Regularization Professor Udell Operations Research and Information Engineering Cornell October 26, 2017 1 / 24 Regularized empirical risk minimization choose model
More informationImage Compression Using Simulated Annealing
Image Compression Using Simulated Annealing Aritra Dutta,GeonwooKim,MeiqinLi, Carlos Ortiz Marrero,MohitSinha,ColeStiegler k August 5 14, 2015 Mathematical Modeling in Industry XIX Institute of Mathematics
More informationACCELERATED LINEARIZED BREGMAN METHOD. June 21, Introduction. In this paper, we are interested in the following optimization problem.
ACCELERATED LINEARIZED BREGMAN METHOD BO HUANG, SHIQIAN MA, AND DONALD GOLDFARB June 21, 2011 Abstract. In this paper, we propose and analyze an accelerated linearized Bregman (A) method for solving the
More informationOptimization methods
Optimization methods Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda /8/016 Introduction Aim: Overview of optimization methods that Tend to
More informationIEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 58, NO. 6, JUNE
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 58, NO. 6, JUNE 2010 2935 Variance-Component Based Sparse Signal Reconstruction and Model Selection Kun Qiu, Student Member, IEEE, and Aleksandar Dogandzic,
More informationLinear Systems of Equations. ChEn 2450
Linear Systems of Equations ChEn 450 LinearSystems-directkey - August 5, 04 Example Circuit analysis (also used in heat transfer) + v _ R R4 I I I3 R R5 R3 Kirchoff s Laws give the following equations
More informationLeast Sparsity of p-norm based Optimization Problems with p > 1
Least Sparsity of p-norm based Optimization Problems with p > Jinglai Shen and Seyedahmad Mousavi Original version: July, 07; Revision: February, 08 Abstract Motivated by l p -optimization arising from
More informationInverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France.
Inverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France remi.gribonval@inria.fr Overview of the course Introduction sparsity & data compression inverse problems
More informationOptimization for Compressed Sensing
Optimization for Compressed Sensing Robert J. Vanderbei 2014 March 21 Dept. of Industrial & Systems Engineering University of Florida http://www.princeton.edu/ rvdb Lasso Regression The problem is to solve
More informationSparse Solutions of Systems of Equations and Sparse Modelling of Signals and Images
Sparse Solutions of Systems of Equations and Sparse Modelling of Signals and Images Alfredo Nava-Tudela ant@umd.edu John J. Benedetto Department of Mathematics jjb@umd.edu Abstract In this project we are
More informationA Survey of L 1. Regression. Céline Cunen, 20/10/2014. Vidaurre, Bielza and Larranaga (2013)
A Survey of L 1 Regression Vidaurre, Bielza and Larranaga (2013) Céline Cunen, 20/10/2014 Outline of article 1.Introduction 2.The Lasso for Linear Regression a) Notation and Main Concepts b) Statistical
More informationBias-free Sparse Regression with Guaranteed Consistency
Bias-free Sparse Regression with Guaranteed Consistency Wotao Yin (UCLA Math) joint with: Stanley Osher, Ming Yan (UCLA) Feng Ruan, Jiechao Xiong, Yuan Yao (Peking U) UC Riverside, STATS Department March
More informationNumerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization
Numerical Linear Algebra Primer Ryan Tibshirani Convex Optimization 10-725 Consider Last time: proximal Newton method min x g(x) + h(x) where g, h convex, g twice differentiable, and h simple. Proximal
More informationGeneralized Orthogonal Matching Pursuit- A Review and Some
Generalized Orthogonal Matching Pursuit- A Review and Some New Results Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur, INDIA Table of Contents
More informationRobustly Stable Signal Recovery in Compressed Sensing with Structured Matrix Perturbation
Robustly Stable Signal Recovery in Compressed Sensing with Structured Matri Perturbation Zai Yang, Cishen Zhang, and Lihua Xie, Fellow, IEEE arxiv:.7v [cs.it] 4 Mar Abstract The sparse signal recovery
More informationOPTIMAL SURE PARAMETERS FOR SIGMOIDAL WAVELET SHRINKAGE
17th European Signal Processing Conference (EUSIPCO 009) Glasgow, Scotland, August 4-8, 009 OPTIMAL SURE PARAMETERS FOR SIGMOIDAL WAVELET SHRINKAGE Abdourrahmane M. Atto 1, Dominique Pastor, Gregoire Mercier
More informationGeneralized Elastic Net Regression
Abstract Generalized Elastic Net Regression Geoffroy MOURET Jean-Jules BRAULT Vahid PARTOVINIA This work presents a variation of the elastic net penalization method. We propose applying a combined l 1
More informationNoise Removal? The Evolution Of Pr(x) Denoising By Energy Minimization. ( x) An Introduction to Sparse Representation and the K-SVD Algorithm
Sparse Representation and the K-SV Algorithm he CS epartment he echnion Israel Institute of technology Haifa 3, Israel University of Erlangen - Nürnberg April 8 Noise Removal? Our story begins with image
More informationHigh-dimensional regression
High-dimensional regression Advanced Methods for Data Analysis 36-402/36-608) Spring 2014 1 Back to linear regression 1.1 Shortcomings Suppose that we are given outcome measurements y 1,... y n R, and
More informationCompressed Sensing in Cancer Biology? (A Work in Progress)
Compressed Sensing in Cancer Biology? (A Work in Progress) M. Vidyasagar FRS Cecil & Ida Green Chair The University of Texas at Dallas M.Vidyasagar@utdallas.edu www.utdallas.edu/ m.vidyasagar University
More information0.1. Linear transformations
Suggestions for midterm review #3 The repetitoria are usually not complete; I am merely bringing up the points that many people didn t now on the recitations Linear transformations The following mostly
More informationNecessary and sufficient conditions of solution uniqueness in l 1 minimization
1 Necessary and sufficient conditions of solution uniqueness in l 1 minimization Hui Zhang, Wotao Yin, and Lizhi Cheng arxiv:1209.0652v2 [cs.it] 18 Sep 2012 Abstract This paper shows that the solutions
More informationMATH 3511 Lecture 1. Solving Linear Systems 1
MATH 3511 Lecture 1 Solving Linear Systems 1 Dmitriy Leykekhman Spring 2012 Goals Review of basic linear algebra Solution of simple linear systems Gaussian elimination D Leykekhman - MATH 3511 Introduction
More informationOWL to the rescue of LASSO
OWL to the rescue of LASSO IISc IBM day 2018 Joint Work R. Sankaran and Francis Bach AISTATS 17 Chiranjib Bhattacharyya Professor, Department of Computer Science and Automation Indian Institute of Science,
More informationCoordinate descent. Geoff Gordon & Ryan Tibshirani Optimization /
Coordinate descent Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Adding to the toolbox, with stats and ML in mind We ve seen several general and useful minimization tools First-order methods
More informationThe lasso: some novel algorithms and applications
1 The lasso: some novel algorithms and applications Newton Institute, June 25, 2008 Robert Tibshirani Stanford University Collaborations with Trevor Hastie, Jerome Friedman, Holger Hoefling, Gen Nowak,
More informationCME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6
CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 GENE H GOLUB Issues with Floating-point Arithmetic We conclude our discussion of floating-point arithmetic by highlighting two issues that frequently
More informationA significance test for the lasso
1 First part: Joint work with Richard Lockhart (SFU), Jonathan Taylor (Stanford), and Ryan Tibshirani (Carnegie-Mellon Univ.) Second part: Joint work with Max Grazier G Sell, Stefan Wager and Alexandra
More informationNear Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing
Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas M.Vidyasagar@utdallas.edu www.utdallas.edu/ m.vidyasagar
More informationColor Scheme. swright/pcmi/ M. Figueiredo and S. Wright () Inference and Optimization PCMI, July / 14
Color Scheme www.cs.wisc.edu/ swright/pcmi/ M. Figueiredo and S. Wright () Inference and Optimization PCMI, July 2016 1 / 14 Statistical Inference via Optimization Many problems in statistical inference
More information