SUPPLEMENTARY MATERIAL TO REGULARIZATION AND THE SMALL-BALL METHOD I: SPARSE RECOVERY
|
|
- Brianne Wright
- 5 years ago
- Views:
Transcription
1 arxiv: arxiv: SUPPLEMETARY MATERIAL TO REGULARIZATIO AD THE SMALL-BALL METHOD I: SPARSE RECOVERY By Guillaume Lecué, and Shahar Mendelson, CREST, CRS, Université Paris Saclay. Technion I.I.T. and Mathematical Sciences Institute, AU. An inspection of Theorem 3. reveals no mention of an isotropicity assumption. There is no choice of a Euclidean structure, and in fact, the statement itself is not even finite dimensional. All that isotropicity has been used for was to bound the complexity function r and the sparsity function in the three applications the LASSO in Theorem.4, SLOPE in Theorem.6 and the trace norm regularization in Theorem.7. We may apply Theorem 3. to situations that do not involve an isotropic vector and here we give an example of how this may be done. To simplify our presentation we will only consider l and SLOPE regularization, which may both be written as Ψt = d β j t j, j= where β β d > 0 and t t d 0 is the nondecreasing rearrangement of t j. As mentioned previously, the LASSO case is recovered for β = = β d = and the SLOPE norm is obtained for β j = C loged/j for some constant C. We also denote by B Ψ resp. S Ψ the unit ball resp. sphere associated with the Ψ-norm. Let Σ R d d be the covariance matrix of X and set D = {x R d : Σ / x } to be the corresponding ellipsoid. aturally, if X is not isotropic than Σ is not the identity matrix. Supported by Labex ECODEC AR - -LABEX-0047 and by the Chaire Economie et Gestion des ouvelles Données, under the auspices of Institut Louis Bachelier, Havas-Media and Paris-Dauphine. Supported by the Israel Science Foundation.
2 G. LECUÉ AD S. MEDELSO In order to apply Theorem 3., we need to bound from above the expectation of the remum of the Gaussian process indexed by ρb Ψ rd: 0. l ρb Ψ rd = E Σ / G, w where G is a standard Gaussian vector in R d. We also need to solve the sparsity equation that is, find ρ > 0 for which ρ 4ρ /5 where, for every ρ > 0, ρ = inf h, g h ρs Ψ rd g Γ t ρ and Γ t ρ is the collection of all subgradients of Ψ of vectors in t + ρ/0b Ψ. We will show that the same results that have been obtained for the LASSO and SLOPE in Theorem.4 and Theorem.6 actually hold under the following assumption. Assumption 0.. Let j {,..., d} and denote by Σ / j Σ /. Let s {,..., d} and set B s = s j= β j/ j.. There exists σ > 0 such that for all j {,..., d}, Σ / j. For all x 0B s S Ψ D, Σ / x J s x J. the j-th row of σ. SLOPE. We first control the Gaussian mean width in 0. when Ψ is the SLOPE norm. Lemma 0.. Set β j = C loged/j and let Σ be a d d symmetric nonnegative matrix for which max j σ. If D = {x R d : Σ / x Σ / } then E E Proof. ote that j Σ / G, w min { ρ C 3 6σ + r 8 π, r } d Σ / G, w re G, Σ / w / = re G r E G r d. w D
3 REGULARIZATIO AD THE SMALL-BALL METHOD 3 ext, let H : R d R be defined by Hu = Σ / u, w : w ρb Ψ rd and recall that G is the standard Gaussian vector in R d. It is straightforward to verify that HG Σ / G, w ρ w ρb Ψ C max ξ j j d loged/j where we set ξ j d j= = Σ/ G and ξ j d j= is the non-increasing rearrangement of ξ j d j=. Observe that for u, v Rd, Hu Hv Σ / u v, w r u v, w = r u v, w B d implying that H is a Lipschitz function with constant r; thus, it follows from p. in Chapter of [] that π 0. EHG MedHG + r, where MedHG is the median of HG. Hence, to obtain the claimed bound on E w ρbψ Σ / G, w it suffices to establish a suitable upper estimate on the median of max j d ξ j loged/j. With that in mind, let ξ,..., ξ be mean-zero Gaussian variables and assume that for every j =,..., d, 0.3 E expξ j /L e for some L > 0. ote that in our case, ξ,..., ξ satisfying 0.3 for L = 3σ/8. By Jensen s inequality, E exp hence, j j k= 0.4 P r ξ k L j j k= j ξ k L j E exp k= log ed/j ξ k L j d k= ξ E exp k L ed j ; exp log ed/j = j ed. Let q 0 be the integer that satisfies q d < q+. It follows from 0.4 that with probability at least q l ed = q > ed, l=0
4 4 G. LECUÉ AD S. MEDELSO for every l = 0,, q, ξ l l l ξ j L loged/ l. max j d j= Moreover, for l j < l+, we have ξ j ξ and loged/ l loged/j; l also for q j d, we have ξ j ξ and loged/ l 3 loged/j. q Therefore, ξ j 0.5 P r 6L > loged/j, proving the requested bound on MedHG. Observe that up to constant σ, we actually recover the same result as in Equation 5.; therefore, one may choose the same complexity function r as in the proof of Theorem.6. Let us turn to a lower bound on the sparsity function. Lemma 0.. There exists an absolute constant 0 < c < 80 for which the following holds. Let s {,..., d} and set B s = β j/ j. Assume that for every x 80B s S ψ D one has Σ / x / J s x J. Let ρ > 0 and assume further that there is a s-sparse vector in t + ρ/0b Ψ. If 80B s ρ/rρ then ρ = inf h ρs Ψ rρd g Γ t ρ 4ρ h, g 5. Proof. Let h ρs Ψ rρd and denote by h j the non-increasing rearrangement of h j. It follows from the proof of Lemma 4. that 7ρ h, g g Γ t ρ 0 β j h j. Let h,s be the s-sparse vector with coordinates given by h j for j s and 0 otherwise. We have h rρ ρ S Ψ D S Ψ D ρ 80B s
5 REGULARIZATIO AD THE SMALL-BALL METHOD 5 implying that Σ / h h,s. Furthermore, since h j h / j for every j s, we have β j h h j,s Σ B s B / s h B s rρ. Hence, if ρ 80rρB s then ρ 4ρ/5. We thus recover the same condition as in Lemma 4.3, implying that Theorem.6 actually holds under the weaker Assumption 0.: let X be an L-subgaussian random vector whose covariance matrix satisfies Assumption 0.. The SLOPE estimator with regularization parameter λ ξ Lq / satisfies, with probability at least δ exp c 0 L 8, Ψˆt t s ed c 3 ξ Lq log and Σ / ˆt t s c 3 ξ s ed L q log s when c 4 s loged/s and when there is a s-sparse vector close enough to t. The LASSO. Here, for every j d, β j = ; B Ψ = B d ; and B s = s j= / j s. Lemma 0.3. Let Σ be a d d symmetric nonnegative matrix for which Σ / max j j σ. If D = {x R d : Σ / x } then every ρ > 0 and r > 0, E Σ / G, w { min r d, ρσ } loged. w ρb d rd Proof. ote that E Σ / G, w re G, Σ / w / = re G r E G r d. w ρb d rd w D ext, set ξ j d j= = Σ/ G and let ξ j d j= be the non-increasing rearrangement of ξ j. Therefore, ξ,..., ξ d are mean-zero Gaussian variables and satisfy E expξj /L e for L = 3σ/8. It is evident that ξ E ξ d L log E exp ξ L log j E exp L log ed, and therefore, E Σ / G, w E w ρb d rd w ρb d j= Σ / G, w ρe ξ / 3ρσL log ed. 8
6 6 G. LECUÉ AD S. MEDELSO Lemma 0.3 leads to a slightly different result than in the isotropic case Lemma 5.3, and as a consequence, r has to be slightly modified. A straightforward computation shows that rmρ ξ Lq d loged L,q,δ min, ρσ ξ L q and r Qρ L { ρ σ 0 if L d log cld otherwise. and still rρ = max{r M ρ, r Q ρ}. Finally, let us prove the sparsity condition. Lemma 0.4. There exists an absolute constant 0 < c < 80 for which the following holds. Let s {,..., d} and set B s = / j. Assume that for every x 0B s Sl d D one has Σ / x / J s x J. Let ρ > 0 and assume further that there is a s-sparse vector in t + ρ/0b d. If 0B s ρ/rρ then ρ = inf h ρsl d rρd g Γ t ρ 4ρ h, g 5. Proof. Let h ρsl d rρd and denote by h j the non-increasing rearrangement of h j. It follows from the proof of Lemma 4. that g Γ t ρ h, g ρ h j. Let h,s be the s-sparse vector with coordinates given by h j for j s and 0 otherwise. Observe that h rρ ρ Sld D Sl d ρ D, 40B s and therefore Σ / h h,s. It follows that h j s h h Σ B s B / s h B s rρ, and in particular, if ρ 0rρB s then ρ 4ρ/5.
7 REGULARIZATIO AD THE SMALL-BALL METHOD 7 Using the estimate on r and Lemma 0.4, it is evident that when L,q,δ sσ loged, one has ρ 4ρ /5 for loged ρ L,q,δ ξ Lq s and if there is a s-sparse vector in t + ρ /0B d. Finally, one may choose the regularization parameter by setting λ r ρ ρ loged L,q,δ ξ Lq s. It follows that if X is an L-subgaussian random vector that satisfies Assumption 0. then with probability larger than δ exp c 0 L 8, ˆt t loged ρ = c δ ξ Lq s and Σ / ˆt t s loged rρ = c L, δ ξ Lq. References. [] Michel Ledoux. The concentration of measure phenomenon, volume 89 of Mathematical Surveys and Monographs. American Mathematical Society, Providence, RI, 00. ESAE, 3, avenue Pierre Larousse, 945 MALAKOFF. France. Department of Mathematics, Technion, I.I.T., Haifa, Israel
REGULARIZATION AND THE SMALL-BALL METHOD I: SPARSE RECOVERY
Submitted to the Annals of Applied Statistics arxiv: arxiv:1601.05584 REGULARIZATIO AD THE SMALL-BALL METHOD I: SPARSE RECOVERY By Guillaume Lecué, and Shahar Mendelson, CREST, CRS, Université Paris Saclay.
More informationarxiv: v2 [math.st] 7 Feb 2017
Estimation bounds and sharp oracle inequalities of regularized procedures with Lipschitz loss functions Pierre Alquier,3,4, Vincent Cottet,3,4, Guillaume Lecué 2,3,4 CREST, ESAE, Université Paris Saclay
More informationMinimax rate of convergence and the performance of ERM in phase recovery
Minimax rate of convergence and the performance of ERM in phase recovery Guillaume Lecué,3 Shahar Mendelson,4,5 ovember 0, 03 Abstract We study the performance of Empirical Risk Minimization in noisy phase
More informationLearning subgaussian classes : Upper and minimax bounds
Learning subgaussian classes : Upper and minimax bounds Guillaume Lecué 1,3 Shahar Mendelson,4,5 May 0, 013 Abstract We obtain sharp oracle inequalities for the empirical risk minimization procedure in
More informationOn the singular values of random matrices
On the singular values of random matrices Shahar Mendelson Grigoris Paouris Abstract We present an approach that allows one to bound the largest and smallest singular values of an N n random matrix with
More informationReconstruction from Anisotropic Random Measurements
Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013
More informationLearning from MOM s principles : Le Cam s approach
Learning from MOM s principles : Le Cam s approach Guillaume Lecué a, Matthieu Lerasle b a CREST, CNRS, Université aris Saclay b Laboratoire de Mathmatiques d Orsay, Univ aris-sud, CNRS, Université aris
More informationSparse recovery under weak moment assumptions
Sparse recovery under weak moment assumptions Guillaume Lecué 1,3 Shahar Mendelson 2,4,5 February 23, 2015 Abstract We prove that iid random vectors that satisfy a rather weak moment assumption can be
More informationUniform uncertainty principle for Bernoulli and subgaussian ensembles
arxiv:math.st/0608665 v1 27 Aug 2006 Uniform uncertainty principle for Bernoulli and subgaussian ensembles Shahar MENDELSON 1 Alain PAJOR 2 Nicole TOMCZAK-JAEGERMANN 3 1 Introduction August 21, 2006 In
More informationLARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011
LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS S. G. Bobkov and F. L. Nazarov September 25, 20 Abstract We study large deviations of linear functionals on an isotropic
More informationarxiv: v2 [math.st] 29 Nov 2017
arxiv:1701.04112v2 [math.st] 29 Nov 2017 Regularization, sparse recovery, and median-of-means tournaments Gábor Lugosi Shahar Mendelson November 30, 2017 Abstract We introduce a regularized risk minimization
More informationMajorizing measures and proportional subsets of bounded orthonormal systems
Majorizing measures and proportional subsets of bounded orthonormal systems Olivier GUÉDON Shahar MENDELSON1 Alain PAJOR Nicole TOMCZAK-JAEGERMANN Abstract In this article we prove that for any orthonormal
More informationl 1 -Regularized Linear Regression: Persistence and Oracle Inequalities
l -Regularized Linear Regression: Persistence and Oracle Inequalities Peter Bartlett EECS and Statistics UC Berkeley slides at http://www.stat.berkeley.edu/ bartlett Joint work with Shahar Mendelson and
More informationarxiv: v1 [math.st] 21 Sep 2016
Bounds on the prediction error of penalized least squares estimators with convex penalty Pierre C. Bellec and Alexandre B. Tsybakov arxiv:609.06675v math.st] Sep 06 Abstract This paper considers the penalized
More informationConvex Optimization & Parsimony of L p-balls representation
Convex Optimization & Parsimony of L p -balls representation LAAS-CNRS and Institute of Mathematics, Toulouse, France IMA, January 2016 Motivation Unit balls associated with nonnegative homogeneous polynomials
More informationIEOR 265 Lecture 3 Sparse Linear Regression
IOR 65 Lecture 3 Sparse Linear Regression 1 M Bound Recall from last lecture that the reason we are interested in complexity measures of sets is because of the following result, which is known as the M
More informationGeometry of log-concave Ensembles of random matrices
Geometry of log-concave Ensembles of random matrices Nicole Tomczak-Jaegermann Joint work with Radosław Adamczak, Rafał Latała, Alexander Litvak, Alain Pajor Cortona, June 2011 Nicole Tomczak-Jaegermann
More informationSparse and Low Rank Recovery via Null Space Properties
Sparse and Low Rank Recovery via Null Space Properties Holger Rauhut Lehrstuhl C für Mathematik (Analysis), RWTH Aachen Convexity, probability and discrete structures, a geometric viewpoint Marne-la-Vallée,
More informationarxiv: v2 [math.st] 12 Feb 2008
arxiv:080.460v2 [math.st] 2 Feb 2008 Electronic Journal of Statistics Vol. 2 2008 90 02 ISSN: 935-7524 DOI: 0.24/08-EJS77 Sup-norm convergence rate and sign concentration property of Lasso and Dantzig
More informationSupremum of simple stochastic processes
Subspace embeddings Daniel Hsu COMS 4772 1 Supremum of simple stochastic processes 2 Recap: JL lemma JL lemma. For any ε (0, 1/2), point set S R d of cardinality 16 ln n S = n, and k N such that k, there
More informationMulti-stage convex relaxation approach for low-rank structured PSD matrix recovery
Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun
More informationGaussian Graphical Models and Graphical Lasso
ELE 538B: Sparsity, Structure and Inference Gaussian Graphical Models and Graphical Lasso Yuxin Chen Princeton University, Spring 2017 Multivariate Gaussians Consider a random vector x N (0, Σ) with pdf
More informationThe Stability of Low-Rank Matrix Reconstruction: a Constrained Singular Value Perspective
Forty-Eighth Annual Allerton Conference Allerton House UIUC Illinois USA September 9 - October 1 010 The Stability of Low-Rank Matrix Reconstruction: a Constrained Singular Value Perspective Gongguo Tang
More information19.1 Problem setup: Sparse linear regression
ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 19: Minimax rates for sparse linear regression Lecturer: Yihong Wu Scribe: Subhadeep Paul, April 13/14, 2016 In
More informationRisk minimization by median-of-means tournaments
Risk minimization by median-of-means tournaments Gábor Lugosi Shahar Mendelson August 2, 206 Abstract We consider the classical statistical learning/regression problem, when the value of a real random
More informationSparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28
Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:
More informationInverse Brascamp-Lieb inequalities along the Heat equation
Inverse Brascamp-Lieb inequalities along the Heat equation Franck Barthe and Dario Cordero-Erausquin October 8, 003 Abstract Adapting Borell s proof of Ehrhard s inequality for general sets, we provide
More informationApproximately Gaussian marginals and the hyperplane conjecture
Approximately Gaussian marginals and the hyperplane conjecture Tel-Aviv University Conference on Asymptotic Geometric Analysis, Euler Institute, St. Petersburg, July 2010 Lecture based on a joint work
More informationSubspaces and orthogonal decompositions generated by bounded orthogonal systems
Subspaces and orthogonal decompositions generated by bounded orthogonal systems Olivier GUÉDON Shahar MENDELSON Alain PAJOR Nicole TOMCZAK-JAEGERMANN August 3, 006 Abstract We investigate properties of
More informationOn the minimum of several random variables
On the minimum of several random variables Yehoram Gordon Alexander Litvak Carsten Schütt Elisabeth Werner Abstract For a given sequence of real numbers a,..., a n we denote the k-th smallest one by k-
More information1 Regression with High Dimensional Data
6.883 Learning with Combinatorial Structure ote for Lecture 11 Instructor: Prof. Stefanie Jegelka Scribe: Xuhong Zhang 1 Regression with High Dimensional Data Consider the following regression problem:
More informationA SIMPLE TOOL FOR BOUNDING THE DEVIATION OF RANDOM MATRICES ON GEOMETRIC SETS
A SIMPLE TOOL FOR BOUNDING THE DEVIATION OF RANDOM MATRICES ON GEOMETRIC SETS CHRISTOPHER LIAW, ABBAS MEHRABIAN, YANIV PLAN, AND ROMAN VERSHYNIN Abstract. Let A be an isotropic, sub-gaussian m n matrix.
More informationRisk Bounds for Lévy Processes in the PAC-Learning Framework
Risk Bounds for Lévy Processes in the PAC-Learning Framework Chao Zhang School of Computer Engineering anyang Technological University Dacheng Tao School of Computer Engineering anyang Technological University
More informationThe Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression
The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression Cun-hui Zhang and Jian Huang Presenter: Quefeng Li Feb. 26, 2010 un-hui Zhang and Jian Huang Presenter: Quefeng The Sparsity
More informationA Greedy Framework for First-Order Optimization
A Greedy Framework for First-Order Optimization Jacob Steinhardt Department of Computer Science Stanford University Stanford, CA 94305 jsteinhardt@cs.stanford.edu Jonathan Huggins Department of EECS Massachusetts
More informationThe volume growth of complete gradient shrinking Ricci solitons
arxiv:0904.0798v [math.dg] Apr 009 The volume growth of complete gradient shrinking Ricci solitons Ovidiu Munteanu Abstract We prove that any gradient shrinking Ricci soliton has at most Euclidean volume
More informationMethods for sparse analysis of high-dimensional data, II
Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional
More informationAPPLICATIONS OF DIFFERENTIABILITY IN R n.
APPLICATIONS OF DIFFERENTIABILITY IN R n. MATANIA BEN-ARTZI April 2015 Functions here are defined on a subset T R n and take values in R m, where m can be smaller, equal or greater than n. The (open) ball
More informationMethods for sparse analysis of high-dimensional data, II
Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional
More informationCONVEX RECOVERY OF A STRUCTURED SIGNAL FROM INDEPENDENT RANDOM LINEAR MEASUREMENTS
CONVEX RECOVERY OF A STRUCTURED SIGNAL FROM INDEPENDENT RANDOM LINEAR MEASUREMENTS JOEL A. TROPP ABSTRACT. This chapter develops a theoretical analysis of the convex programming method for recovering a
More informationarxiv: v1 [math.pr] 22 Dec 2018
arxiv:1812.09618v1 [math.pr] 22 Dec 2018 Operator norm upper bound for sub-gaussian tailed random matrices Eric Benhamou Jamal Atif Rida Laraki December 27, 2018 Abstract This paper investigates an upper
More informationA note on the convex infimum convolution inequality
A note on the convex infimum convolution inequality Naomi Feldheim, Arnaud Marsiglietti, Piotr Nayar, Jing Wang Abstract We characterize the symmetric measures which satisfy the one dimensional convex
More informationarxiv: v1 [math.st] 27 Sep 2018
Robust covariance estimation under L 4 L 2 norm equivalence. arxiv:809.0462v [math.st] 27 Sep 208 Shahar Mendelson ikita Zhivotovskiy Abstract Let X be a centered random vector taking values in R d and
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini April 27, 2018 1 / 80 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationExistence, stability and instability for Einstein-scalar field Lichnerowicz equations by Emmanuel Hebey
Existence, stability and instability for Einstein-scalar field Lichnerowicz equations by Emmanuel Hebey Joint works with Olivier Druet and with Frank Pacard and Dan Pollack Two hours lectures IAS, October
More informationarxiv: v3 [cs.it] 25 Jul 2014
Simultaneously Structured Models with Application to Sparse and Low-rank Matrices Samet Oymak c, Amin Jalali w, Maryam Fazel w, Yonina C. Eldar t, Babak Hassibi c arxiv:11.3753v3 [cs.it] 5 Jul 014 c California
More informationLecture 13 October 6, Covering Numbers and Maurey s Empirical Method
CS 395T: Sublinear Algorithms Fall 2016 Prof. Eric Price Lecture 13 October 6, 2016 Scribe: Kiyeon Jeon and Loc Hoang 1 Overview In the last lecture we covered the lower bound for p th moment (p > 2) and
More informationSparse Recovery with Pre-Gaussian Random Matrices
Sparse Recovery with Pre-Gaussian Random Matrices Simon Foucart Laboratoire Jacques-Louis Lions Université Pierre et Marie Curie Paris, 75013, France Ming-Jun Lai Department of Mathematics University of
More informationStatistical learning with Lipschitz and convex loss functions
Statistical learning with Lipschitz and convex loss functions Geoffrey Chinot, Guillaume Lecué and Matthieu Lerasle October 3, 08 Abstract We obtain risk bounds for Empirical Risk Minimizers ERM and minmax
More informationPointwise estimates for marginals of convex bodies
Journal of Functional Analysis 54 8) 75 93 www.elsevier.com/locate/jfa Pointwise estimates for marginals of convex bodies R. Eldan a,1,b.klartag b,, a School of Mathematical Sciences, Tel-Aviv University,
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57
More informationCompressive Sensing with Random Matrices
Compressive Sensing with Random Matrices Lucas Connell University of Georgia 9 November 017 Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November 017 1 / 18 Overview
More informationTools from Lebesgue integration
Tools from Lebesgue integration E.P. van den Ban Fall 2005 Introduction In these notes we describe some of the basic tools from the theory of Lebesgue integration. Definitions and results will be given
More informationContents: 1. Minimization. 2. The theorem of Lions-Stampacchia for variational inequalities. 3. Γ -Convergence. 4. Duality mapping.
Minimization Contents: 1. Minimization. 2. The theorem of Lions-Stampacchia for variational inequalities. 3. Γ -Convergence. 4. Duality mapping. 1 Minimization A Topological Result. Let S be a topological
More informationRecent Developments in Compressed Sensing
Recent Developments in Compressed Sensing M. Vidyasagar Distinguished Professor, IIT Hyderabad m.vidyasagar@iith.ac.in, www.iith.ac.in/ m vidyasagar/ ISL Seminar, Stanford University, 19 April 2018 Outline
More informationOptimization methods
Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,
More informationConvex inequalities, isoperimetry and spectral gap III
Convex inequalities, isoperimetry and spectral gap III Jesús Bastero (Universidad de Zaragoza) CIDAMA Antequera, September 11, 2014 Part III. K-L-S spectral gap conjecture KLS estimate, through Milman's
More informationSignal Recovery from Permuted Observations
EE381V Course Project Signal Recovery from Permuted Observations 1 Problem Shanshan Wu (sw33323) May 8th, 2015 We start with the following problem: let s R n be an unknown n-dimensional real-valued signal,
More informationPseudo-Poincaré Inequalities and Applications to Sobolev Inequalities
Pseudo-Poincaré Inequalities and Applications to Sobolev Inequalities Laurent Saloff-Coste Abstract Most smoothing procedures are via averaging. Pseudo-Poincaré inequalities give a basic L p -norm control
More informationAN EXAMPLE OF FUNCTIONAL WHICH IS WEAKLY LOWER SEMICONTINUOUS ON W 1,p FOR EVERY p > 2 BUT NOT ON H0
AN EXAMPLE OF FUNCTIONAL WHICH IS WEAKLY LOWER SEMICONTINUOUS ON W,p FOR EVERY p > BUT NOT ON H FERNANDO FARRONI, RAFFAELLA GIOVA AND FRANÇOIS MURAT Abstract. In this note we give an example of functional
More informationConfidence Intervals for Low-dimensional Parameters with High-dimensional Data
Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology
More informationNear Optimal Signal Recovery From Random Projections: Universal Encoding Strategies?
Near Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? Emmanuel Candes and Terence Tao Applied and Computational Mathematics, Caltech, Pasadena, CA 91125 Department of Mathematics,
More informationOptimisation Combinatoire et Convexe.
Optimisation Combinatoire et Convexe. Low complexity models, l 1 penalties. A. d Aspremont. M1 ENS. 1/36 Today Sparsity, low complexity models. l 1 -recovery results: three approaches. Extensions: matrix
More informationA New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables
A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider the problem of
More informationTHE SMALLEST SINGULAR VALUE OF A RANDOM RECTANGULAR MATRIX
THE SMALLEST SINGULAR VALUE OF A RANDOM RECTANGULAR MATRIX MARK RUDELSON AND ROMAN VERSHYNIN Abstract. We prove an optimal estimate on the smallest singular value of a random subgaussian matrix, valid
More informationRANDOM FIELDS AND GEOMETRY. Robert Adler and Jonathan Taylor
RANDOM FIELDS AND GEOMETRY from the book of the same name by Robert Adler and Jonathan Taylor IE&M, Technion, Israel, Statistics, Stanford, US. ie.technion.ac.il/adler.phtml www-stat.stanford.edu/ jtaylor
More informationThe adaptive and the thresholded Lasso for potentially misspecified models (and a lower bound for the Lasso)
Electronic Journal of Statistics Vol. 0 (2010) ISSN: 1935-7524 The adaptive the thresholded Lasso for potentially misspecified models ( a lower bound for the Lasso) Sara van de Geer Peter Bühlmann Seminar
More informationDaniel M. Oberlin Department of Mathematics, Florida State University. January 2005
PACKING SPHERES AND FRACTAL STRICHARTZ ESTIMATES IN R d FOR d 3 Daniel M. Oberlin Department of Mathematics, Florida State University January 005 Fix a dimension d and for x R d and r > 0, let Sx, r) stand
More informationA Unified Approach to Proximal Algorithms using Bregman Distance
A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department
More informationOn Acceleration with Noise-Corrupted Gradients. + m k 1 (x). By the definition of Bregman divergence:
A Omitted Proofs from Section 3 Proof of Lemma 3 Let m x) = a i On Acceleration with Noise-Corrupted Gradients fxi ), u x i D ψ u, x 0 ) denote the function under the minimum in the lower bound By Proposition
More informationZ Algorithmic Superpower Randomization October 15th, Lecture 12
15.859-Z Algorithmic Superpower Randomization October 15th, 014 Lecture 1 Lecturer: Bernhard Haeupler Scribe: Goran Žužić Today s lecture is about finding sparse solutions to linear systems. The problem
More informationMATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous:
MATH 51H Section 4 October 16, 2015 1 Continuity Recall what it means for a function between metric spaces to be continuous: Definition. Let (X, d X ), (Y, d Y ) be metric spaces. A function f : X Y is
More informationarxiv:math/ v1 [math.fa] 30 Sep 2004
arxiv:math/04000v [math.fa] 30 Sep 2004 Small ball probability and Dvoretzy Theorem B. Klartag, R. Vershynin Abstract Large deviation estimates are by now a standard tool in the Asymptotic Convex Geometry,
More information4 Sobolev spaces, trace theorem and normal derivative
4 Sobolev spaces, trace theorem and normal derivative Throughout, n will be a sufficiently smooth, bounded domain. We use the standard Sobolev spaces H 0 ( n ) := L 2 ( n ), H 0 () := L 2 (), H k ( n ),
More informationON THE REGULARITY OF SAMPLE PATHS OF SUB-ELLIPTIC DIFFUSIONS ON MANIFOLDS
Bendikov, A. and Saloff-Coste, L. Osaka J. Math. 4 (5), 677 7 ON THE REGULARITY OF SAMPLE PATHS OF SUB-ELLIPTIC DIFFUSIONS ON MANIFOLDS ALEXANDER BENDIKOV and LAURENT SALOFF-COSTE (Received March 4, 4)
More informationThe Canonical Gaussian Measure on R
The Canonical Gaussian Measure on R 1. Introduction The main goal of this course is to study Gaussian measures. The simplest example of a Gaussian measure is the canonical Gaussian measure P on R where
More informationMath 259: Introduction to Analytic Number Theory Functions of finite order: product formula and logarithmic derivative
Math 259: Introduction to Analytic Number Theory Functions of finite order: product formula and logarithmic derivative This chapter is another review of standard material in complex analysis. See for instance
More informationInégalités de dispersion via le semi-groupe de la chaleur
Inégalités de dispersion via le semi-groupe de la chaleur Valentin Samoyeau, Advisor: Frédéric Bernicot. Laboratoire de Mathématiques Jean Leray, Université de Nantes January 28, 2016 1 Introduction Schrödinger
More informationMultivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma
Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Suppose again we have n sample points x,..., x n R p. The data-point x i R p can be thought of as the i-th row X i of an n p-dimensional
More informationLecture 6: September 19
36-755: Advanced Statistical Theory I Fall 2016 Lecture 6: September 19 Lecturer: Alessandro Rinaldo Scribe: YJ Choe Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have
More informationConvex Analysis and Economic Theory AY Elementary properties of convex functions
Division of the Humanities and Social Sciences Ec 181 KC Border Convex Analysis and Economic Theory AY 2018 2019 Topic 6: Convex functions I 6.1 Elementary properties of convex functions We may occasionally
More informationAn Algorithmist s Toolkit Nov. 10, Lecture 17
8.409 An Algorithmist s Toolkit Nov. 0, 009 Lecturer: Jonathan Kelner Lecture 7 Johnson-Lindenstrauss Theorem. Recap We first recap a theorem (isoperimetric inequality) and a lemma (concentration) from
More informationLECTURE 7. Least Squares and Variants. Optimization Models EE 127 / EE 227AT. Outline. Least Squares. Notes. Notes. Notes. Notes.
Optimization Models EE 127 / EE 227AT Laurent El Ghaoui EECS department UC Berkeley Spring 2015 Sp 15 1 / 23 LECTURE 7 Least Squares and Variants If others would but reflect on mathematical truths as deeply
More informationOptimization for Compressed Sensing
Optimization for Compressed Sensing Robert J. Vanderbei 2014 March 21 Dept. of Industrial & Systems Engineering University of Florida http://www.princeton.edu/ rvdb Lasso Regression The problem is to solve
More informationComposite Objective Mirror Descent
Composite Objective Mirror Descent John C. Duchi 1,3 Shai Shalev-Shwartz 2 Yoram Singer 3 Ambuj Tewari 4 1 University of California, Berkeley 2 Hebrew University of Jerusalem, Israel 3 Google Research
More information1 Sparsity and l 1 relaxation
6.883 Learning with Combinatorial Structure Note for Lecture 2 Author: Chiyuan Zhang Sparsity and l relaxation Last time we talked about sparsity and characterized when an l relaxation could recover the
More informationRademacher Averages and Phase Transitions in Glivenko Cantelli Classes
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 1, JANUARY 2002 251 Rademacher Averages Phase Transitions in Glivenko Cantelli Classes Shahar Mendelson Abstract We introduce a new parameter which
More informationarxiv: v2 [math.st] 30 Nov 2017
Robust machine learning by median-of-means : theory and practice G. Lecué and M. Lerasle December 4, 2017 arxiv:1711.10306v2 [math.st] 30 Nov 2017 Abstract We introduce new estimators for robust machine
More informationConvexity of chance constraints with dependent random variables: the use of copulae
Convexity of chance constraints with dependent random variables: the use of copulae René Henrion 1 and Cyrille Strugarek 2 1 Weierstrass Institute for Applied Analysis and Stochastics, 10117 Berlin, Germany.
More information(Part 1) High-dimensional statistics May / 41
Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2
More informationAn example of a convex body without symmetric projections.
An example of a convex body without symmetric projections. E. D. Gluskin A. E. Litvak N. Tomczak-Jaegermann Abstract Many crucial results of the asymptotic theory of symmetric convex bodies were extended
More informationStrengthened Sobolev inequalities for a random subspace of functions
Strengthened Sobolev inequalities for a random subspace of functions Rachel Ward University of Texas at Austin April 2013 2 Discrete Sobolev inequalities Proposition (Sobolev inequality for discrete images)
More informationOptimization and Optimal Control in Banach Spaces
Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,
More informationREPRESENTABLE BANACH SPACES AND UNIFORMLY GÂTEAUX-SMOOTH NORMS. Julien Frontisi
Serdica Math. J. 22 (1996), 33-38 REPRESENTABLE BANACH SPACES AND UNIFORMLY GÂTEAUX-SMOOTH NORMS Julien Frontisi Communicated by G. Godefroy Abstract. It is proved that a representable non-separable Banach
More informationROW PRODUCTS OF RANDOM MATRICES
ROW PRODUCTS OF RANDOM MATRICES MARK RUDELSON Abstract Let 1,, K be d n matrices We define the row product of these matrices as a d K n matrix, whose rows are entry-wise products of rows of 1,, K This
More informationContents 1. Introduction 1 2. Main results 3 3. Proof of the main inequalities 7 4. Application to random dynamical systems 11 References 16
WEIGHTED CSISZÁR-KULLBACK-PINSKER INEQUALITIES AND APPLICATIONS TO TRANSPORTATION INEQUALITIES FRANÇOIS BOLLEY AND CÉDRIC VILLANI Abstract. We strengthen the usual Csiszár-Kullback-Pinsker inequality by
More informationarxiv: v1 [math.pr] 11 Feb 2019
A Short Note on Concentration Inequalities for Random Vectors with SubGaussian Norm arxiv:190.03736v1 math.pr] 11 Feb 019 Chi Jin University of California, Berkeley chijin@cs.berkeley.edu Rong Ge Duke
More informationTHE LASSO, CORRELATED DESIGN, AND IMPROVED ORACLE INEQUALITIES. By Sara van de Geer and Johannes Lederer. ETH Zürich
Submitted to the Annals of Applied Statistics arxiv: math.pr/0000000 THE LASSO, CORRELATED DESIGN, AND IMPROVED ORACLE INEQUALITIES By Sara van de Geer and Johannes Lederer ETH Zürich We study high-dimensional
More informationMath 259: Introduction to Analytic Number Theory Functions of finite order: product formula and logarithmic derivative
Math 259: Introduction to Analytic Number Theory Functions of finite order: product formula and logarithmic derivative This chapter is another review of standard material in complex analysis. See for instance
More informationSpectral Gap and Concentration for Some Spherically Symmetric Probability Measures
Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures S.G. Bobkov School of Mathematics, University of Minnesota, 127 Vincent Hall, 26 Church St. S.E., Minneapolis, MN 55455,
More information