SUPPLEMENTARY MATERIAL TO REGULARIZATION AND THE SMALL-BALL METHOD I: SPARSE RECOVERY

Size: px
Start display at page:

Download "SUPPLEMENTARY MATERIAL TO REGULARIZATION AND THE SMALL-BALL METHOD I: SPARSE RECOVERY"

Transcription

1 arxiv: arxiv: SUPPLEMETARY MATERIAL TO REGULARIZATIO AD THE SMALL-BALL METHOD I: SPARSE RECOVERY By Guillaume Lecué, and Shahar Mendelson, CREST, CRS, Université Paris Saclay. Technion I.I.T. and Mathematical Sciences Institute, AU. An inspection of Theorem 3. reveals no mention of an isotropicity assumption. There is no choice of a Euclidean structure, and in fact, the statement itself is not even finite dimensional. All that isotropicity has been used for was to bound the complexity function r and the sparsity function in the three applications the LASSO in Theorem.4, SLOPE in Theorem.6 and the trace norm regularization in Theorem.7. We may apply Theorem 3. to situations that do not involve an isotropic vector and here we give an example of how this may be done. To simplify our presentation we will only consider l and SLOPE regularization, which may both be written as Ψt = d β j t j, j= where β β d > 0 and t t d 0 is the nondecreasing rearrangement of t j. As mentioned previously, the LASSO case is recovered for β = = β d = and the SLOPE norm is obtained for β j = C loged/j for some constant C. We also denote by B Ψ resp. S Ψ the unit ball resp. sphere associated with the Ψ-norm. Let Σ R d d be the covariance matrix of X and set D = {x R d : Σ / x } to be the corresponding ellipsoid. aturally, if X is not isotropic than Σ is not the identity matrix. Supported by Labex ECODEC AR - -LABEX-0047 and by the Chaire Economie et Gestion des ouvelles Données, under the auspices of Institut Louis Bachelier, Havas-Media and Paris-Dauphine. Supported by the Israel Science Foundation.

2 G. LECUÉ AD S. MEDELSO In order to apply Theorem 3., we need to bound from above the expectation of the remum of the Gaussian process indexed by ρb Ψ rd: 0. l ρb Ψ rd = E Σ / G, w where G is a standard Gaussian vector in R d. We also need to solve the sparsity equation that is, find ρ > 0 for which ρ 4ρ /5 where, for every ρ > 0, ρ = inf h, g h ρs Ψ rd g Γ t ρ and Γ t ρ is the collection of all subgradients of Ψ of vectors in t + ρ/0b Ψ. We will show that the same results that have been obtained for the LASSO and SLOPE in Theorem.4 and Theorem.6 actually hold under the following assumption. Assumption 0.. Let j {,..., d} and denote by Σ / j Σ /. Let s {,..., d} and set B s = s j= β j/ j.. There exists σ > 0 such that for all j {,..., d}, Σ / j. For all x 0B s S Ψ D, Σ / x J s x J. the j-th row of σ. SLOPE. We first control the Gaussian mean width in 0. when Ψ is the SLOPE norm. Lemma 0.. Set β j = C loged/j and let Σ be a d d symmetric nonnegative matrix for which max j σ. If D = {x R d : Σ / x Σ / } then E E Proof. ote that j Σ / G, w min { ρ C 3 6σ + r 8 π, r } d Σ / G, w re G, Σ / w / = re G r E G r d. w D

3 REGULARIZATIO AD THE SMALL-BALL METHOD 3 ext, let H : R d R be defined by Hu = Σ / u, w : w ρb Ψ rd and recall that G is the standard Gaussian vector in R d. It is straightforward to verify that HG Σ / G, w ρ w ρb Ψ C max ξ j j d loged/j where we set ξ j d j= = Σ/ G and ξ j d j= is the non-increasing rearrangement of ξ j d j=. Observe that for u, v Rd, Hu Hv Σ / u v, w r u v, w = r u v, w B d implying that H is a Lipschitz function with constant r; thus, it follows from p. in Chapter of [] that π 0. EHG MedHG + r, where MedHG is the median of HG. Hence, to obtain the claimed bound on E w ρbψ Σ / G, w it suffices to establish a suitable upper estimate on the median of max j d ξ j loged/j. With that in mind, let ξ,..., ξ be mean-zero Gaussian variables and assume that for every j =,..., d, 0.3 E expξ j /L e for some L > 0. ote that in our case, ξ,..., ξ satisfying 0.3 for L = 3σ/8. By Jensen s inequality, E exp hence, j j k= 0.4 P r ξ k L j j k= j ξ k L j E exp k= log ed/j ξ k L j d k= ξ E exp k L ed j ; exp log ed/j = j ed. Let q 0 be the integer that satisfies q d < q+. It follows from 0.4 that with probability at least q l ed = q > ed, l=0

4 4 G. LECUÉ AD S. MEDELSO for every l = 0,, q, ξ l l l ξ j L loged/ l. max j d j= Moreover, for l j < l+, we have ξ j ξ and loged/ l loged/j; l also for q j d, we have ξ j ξ and loged/ l 3 loged/j. q Therefore, ξ j 0.5 P r 6L > loged/j, proving the requested bound on MedHG. Observe that up to constant σ, we actually recover the same result as in Equation 5.; therefore, one may choose the same complexity function r as in the proof of Theorem.6. Let us turn to a lower bound on the sparsity function. Lemma 0.. There exists an absolute constant 0 < c < 80 for which the following holds. Let s {,..., d} and set B s = β j/ j. Assume that for every x 80B s S ψ D one has Σ / x / J s x J. Let ρ > 0 and assume further that there is a s-sparse vector in t + ρ/0b Ψ. If 80B s ρ/rρ then ρ = inf h ρs Ψ rρd g Γ t ρ 4ρ h, g 5. Proof. Let h ρs Ψ rρd and denote by h j the non-increasing rearrangement of h j. It follows from the proof of Lemma 4. that 7ρ h, g g Γ t ρ 0 β j h j. Let h,s be the s-sparse vector with coordinates given by h j for j s and 0 otherwise. We have h rρ ρ S Ψ D S Ψ D ρ 80B s

5 REGULARIZATIO AD THE SMALL-BALL METHOD 5 implying that Σ / h h,s. Furthermore, since h j h / j for every j s, we have β j h h j,s Σ B s B / s h B s rρ. Hence, if ρ 80rρB s then ρ 4ρ/5. We thus recover the same condition as in Lemma 4.3, implying that Theorem.6 actually holds under the weaker Assumption 0.: let X be an L-subgaussian random vector whose covariance matrix satisfies Assumption 0.. The SLOPE estimator with regularization parameter λ ξ Lq / satisfies, with probability at least δ exp c 0 L 8, Ψˆt t s ed c 3 ξ Lq log and Σ / ˆt t s c 3 ξ s ed L q log s when c 4 s loged/s and when there is a s-sparse vector close enough to t. The LASSO. Here, for every j d, β j = ; B Ψ = B d ; and B s = s j= / j s. Lemma 0.3. Let Σ be a d d symmetric nonnegative matrix for which Σ / max j j σ. If D = {x R d : Σ / x } then every ρ > 0 and r > 0, E Σ / G, w { min r d, ρσ } loged. w ρb d rd Proof. ote that E Σ / G, w re G, Σ / w / = re G r E G r d. w ρb d rd w D ext, set ξ j d j= = Σ/ G and let ξ j d j= be the non-increasing rearrangement of ξ j. Therefore, ξ,..., ξ d are mean-zero Gaussian variables and satisfy E expξj /L e for L = 3σ/8. It is evident that ξ E ξ d L log E exp ξ L log j E exp L log ed, and therefore, E Σ / G, w E w ρb d rd w ρb d j= Σ / G, w ρe ξ / 3ρσL log ed. 8

6 6 G. LECUÉ AD S. MEDELSO Lemma 0.3 leads to a slightly different result than in the isotropic case Lemma 5.3, and as a consequence, r has to be slightly modified. A straightforward computation shows that rmρ ξ Lq d loged L,q,δ min, ρσ ξ L q and r Qρ L { ρ σ 0 if L d log cld otherwise. and still rρ = max{r M ρ, r Q ρ}. Finally, let us prove the sparsity condition. Lemma 0.4. There exists an absolute constant 0 < c < 80 for which the following holds. Let s {,..., d} and set B s = / j. Assume that for every x 0B s Sl d D one has Σ / x / J s x J. Let ρ > 0 and assume further that there is a s-sparse vector in t + ρ/0b d. If 0B s ρ/rρ then ρ = inf h ρsl d rρd g Γ t ρ 4ρ h, g 5. Proof. Let h ρsl d rρd and denote by h j the non-increasing rearrangement of h j. It follows from the proof of Lemma 4. that g Γ t ρ h, g ρ h j. Let h,s be the s-sparse vector with coordinates given by h j for j s and 0 otherwise. Observe that h rρ ρ Sld D Sl d ρ D, 40B s and therefore Σ / h h,s. It follows that h j s h h Σ B s B / s h B s rρ, and in particular, if ρ 0rρB s then ρ 4ρ/5.

7 REGULARIZATIO AD THE SMALL-BALL METHOD 7 Using the estimate on r and Lemma 0.4, it is evident that when L,q,δ sσ loged, one has ρ 4ρ /5 for loged ρ L,q,δ ξ Lq s and if there is a s-sparse vector in t + ρ /0B d. Finally, one may choose the regularization parameter by setting λ r ρ ρ loged L,q,δ ξ Lq s. It follows that if X is an L-subgaussian random vector that satisfies Assumption 0. then with probability larger than δ exp c 0 L 8, ˆt t loged ρ = c δ ξ Lq s and Σ / ˆt t s loged rρ = c L, δ ξ Lq. References. [] Michel Ledoux. The concentration of measure phenomenon, volume 89 of Mathematical Surveys and Monographs. American Mathematical Society, Providence, RI, 00. ESAE, 3, avenue Pierre Larousse, 945 MALAKOFF. France. Department of Mathematics, Technion, I.I.T., Haifa, Israel

REGULARIZATION AND THE SMALL-BALL METHOD I: SPARSE RECOVERY

REGULARIZATION AND THE SMALL-BALL METHOD I: SPARSE RECOVERY Submitted to the Annals of Applied Statistics arxiv: arxiv:1601.05584 REGULARIZATIO AD THE SMALL-BALL METHOD I: SPARSE RECOVERY By Guillaume Lecué, and Shahar Mendelson, CREST, CRS, Université Paris Saclay.

More information

arxiv: v2 [math.st] 7 Feb 2017

arxiv: v2 [math.st] 7 Feb 2017 Estimation bounds and sharp oracle inequalities of regularized procedures with Lipschitz loss functions Pierre Alquier,3,4, Vincent Cottet,3,4, Guillaume Lecué 2,3,4 CREST, ESAE, Université Paris Saclay

More information

Minimax rate of convergence and the performance of ERM in phase recovery

Minimax rate of convergence and the performance of ERM in phase recovery Minimax rate of convergence and the performance of ERM in phase recovery Guillaume Lecué,3 Shahar Mendelson,4,5 ovember 0, 03 Abstract We study the performance of Empirical Risk Minimization in noisy phase

More information

Learning subgaussian classes : Upper and minimax bounds

Learning subgaussian classes : Upper and minimax bounds Learning subgaussian classes : Upper and minimax bounds Guillaume Lecué 1,3 Shahar Mendelson,4,5 May 0, 013 Abstract We obtain sharp oracle inequalities for the empirical risk minimization procedure in

More information

On the singular values of random matrices

On the singular values of random matrices On the singular values of random matrices Shahar Mendelson Grigoris Paouris Abstract We present an approach that allows one to bound the largest and smallest singular values of an N n random matrix with

More information

Reconstruction from Anisotropic Random Measurements

Reconstruction from Anisotropic Random Measurements Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013

More information

Learning from MOM s principles : Le Cam s approach

Learning from MOM s principles : Le Cam s approach Learning from MOM s principles : Le Cam s approach Guillaume Lecué a, Matthieu Lerasle b a CREST, CNRS, Université aris Saclay b Laboratoire de Mathmatiques d Orsay, Univ aris-sud, CNRS, Université aris

More information

Sparse recovery under weak moment assumptions

Sparse recovery under weak moment assumptions Sparse recovery under weak moment assumptions Guillaume Lecué 1,3 Shahar Mendelson 2,4,5 February 23, 2015 Abstract We prove that iid random vectors that satisfy a rather weak moment assumption can be

More information

Uniform uncertainty principle for Bernoulli and subgaussian ensembles

Uniform uncertainty principle for Bernoulli and subgaussian ensembles arxiv:math.st/0608665 v1 27 Aug 2006 Uniform uncertainty principle for Bernoulli and subgaussian ensembles Shahar MENDELSON 1 Alain PAJOR 2 Nicole TOMCZAK-JAEGERMANN 3 1 Introduction August 21, 2006 In

More information

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011 LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS S. G. Bobkov and F. L. Nazarov September 25, 20 Abstract We study large deviations of linear functionals on an isotropic

More information

arxiv: v2 [math.st] 29 Nov 2017

arxiv: v2 [math.st] 29 Nov 2017 arxiv:1701.04112v2 [math.st] 29 Nov 2017 Regularization, sparse recovery, and median-of-means tournaments Gábor Lugosi Shahar Mendelson November 30, 2017 Abstract We introduce a regularized risk minimization

More information

Majorizing measures and proportional subsets of bounded orthonormal systems

Majorizing measures and proportional subsets of bounded orthonormal systems Majorizing measures and proportional subsets of bounded orthonormal systems Olivier GUÉDON Shahar MENDELSON1 Alain PAJOR Nicole TOMCZAK-JAEGERMANN Abstract In this article we prove that for any orthonormal

More information

l 1 -Regularized Linear Regression: Persistence and Oracle Inequalities

l 1 -Regularized Linear Regression: Persistence and Oracle Inequalities l -Regularized Linear Regression: Persistence and Oracle Inequalities Peter Bartlett EECS and Statistics UC Berkeley slides at http://www.stat.berkeley.edu/ bartlett Joint work with Shahar Mendelson and

More information

arxiv: v1 [math.st] 21 Sep 2016

arxiv: v1 [math.st] 21 Sep 2016 Bounds on the prediction error of penalized least squares estimators with convex penalty Pierre C. Bellec and Alexandre B. Tsybakov arxiv:609.06675v math.st] Sep 06 Abstract This paper considers the penalized

More information

Convex Optimization & Parsimony of L p-balls representation

Convex Optimization & Parsimony of L p-balls representation Convex Optimization & Parsimony of L p -balls representation LAAS-CNRS and Institute of Mathematics, Toulouse, France IMA, January 2016 Motivation Unit balls associated with nonnegative homogeneous polynomials

More information

IEOR 265 Lecture 3 Sparse Linear Regression

IEOR 265 Lecture 3 Sparse Linear Regression IOR 65 Lecture 3 Sparse Linear Regression 1 M Bound Recall from last lecture that the reason we are interested in complexity measures of sets is because of the following result, which is known as the M

More information

Geometry of log-concave Ensembles of random matrices

Geometry of log-concave Ensembles of random matrices Geometry of log-concave Ensembles of random matrices Nicole Tomczak-Jaegermann Joint work with Radosław Adamczak, Rafał Latała, Alexander Litvak, Alain Pajor Cortona, June 2011 Nicole Tomczak-Jaegermann

More information

Sparse and Low Rank Recovery via Null Space Properties

Sparse and Low Rank Recovery via Null Space Properties Sparse and Low Rank Recovery via Null Space Properties Holger Rauhut Lehrstuhl C für Mathematik (Analysis), RWTH Aachen Convexity, probability and discrete structures, a geometric viewpoint Marne-la-Vallée,

More information

arxiv: v2 [math.st] 12 Feb 2008

arxiv: v2 [math.st] 12 Feb 2008 arxiv:080.460v2 [math.st] 2 Feb 2008 Electronic Journal of Statistics Vol. 2 2008 90 02 ISSN: 935-7524 DOI: 0.24/08-EJS77 Sup-norm convergence rate and sign concentration property of Lasso and Dantzig

More information

Supremum of simple stochastic processes

Supremum of simple stochastic processes Subspace embeddings Daniel Hsu COMS 4772 1 Supremum of simple stochastic processes 2 Recap: JL lemma JL lemma. For any ε (0, 1/2), point set S R d of cardinality 16 ln n S = n, and k N such that k, there

More information

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun

More information

Gaussian Graphical Models and Graphical Lasso

Gaussian Graphical Models and Graphical Lasso ELE 538B: Sparsity, Structure and Inference Gaussian Graphical Models and Graphical Lasso Yuxin Chen Princeton University, Spring 2017 Multivariate Gaussians Consider a random vector x N (0, Σ) with pdf

More information

The Stability of Low-Rank Matrix Reconstruction: a Constrained Singular Value Perspective

The Stability of Low-Rank Matrix Reconstruction: a Constrained Singular Value Perspective Forty-Eighth Annual Allerton Conference Allerton House UIUC Illinois USA September 9 - October 1 010 The Stability of Low-Rank Matrix Reconstruction: a Constrained Singular Value Perspective Gongguo Tang

More information

19.1 Problem setup: Sparse linear regression

19.1 Problem setup: Sparse linear regression ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 19: Minimax rates for sparse linear regression Lecturer: Yihong Wu Scribe: Subhadeep Paul, April 13/14, 2016 In

More information

Risk minimization by median-of-means tournaments

Risk minimization by median-of-means tournaments Risk minimization by median-of-means tournaments Gábor Lugosi Shahar Mendelson August 2, 206 Abstract We consider the classical statistical learning/regression problem, when the value of a real random

More information

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28 Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:

More information

Inverse Brascamp-Lieb inequalities along the Heat equation

Inverse Brascamp-Lieb inequalities along the Heat equation Inverse Brascamp-Lieb inequalities along the Heat equation Franck Barthe and Dario Cordero-Erausquin October 8, 003 Abstract Adapting Borell s proof of Ehrhard s inequality for general sets, we provide

More information

Approximately Gaussian marginals and the hyperplane conjecture

Approximately Gaussian marginals and the hyperplane conjecture Approximately Gaussian marginals and the hyperplane conjecture Tel-Aviv University Conference on Asymptotic Geometric Analysis, Euler Institute, St. Petersburg, July 2010 Lecture based on a joint work

More information

Subspaces and orthogonal decompositions generated by bounded orthogonal systems

Subspaces and orthogonal decompositions generated by bounded orthogonal systems Subspaces and orthogonal decompositions generated by bounded orthogonal systems Olivier GUÉDON Shahar MENDELSON Alain PAJOR Nicole TOMCZAK-JAEGERMANN August 3, 006 Abstract We investigate properties of

More information

On the minimum of several random variables

On the minimum of several random variables On the minimum of several random variables Yehoram Gordon Alexander Litvak Carsten Schütt Elisabeth Werner Abstract For a given sequence of real numbers a,..., a n we denote the k-th smallest one by k-

More information

1 Regression with High Dimensional Data

1 Regression with High Dimensional Data 6.883 Learning with Combinatorial Structure ote for Lecture 11 Instructor: Prof. Stefanie Jegelka Scribe: Xuhong Zhang 1 Regression with High Dimensional Data Consider the following regression problem:

More information

A SIMPLE TOOL FOR BOUNDING THE DEVIATION OF RANDOM MATRICES ON GEOMETRIC SETS

A SIMPLE TOOL FOR BOUNDING THE DEVIATION OF RANDOM MATRICES ON GEOMETRIC SETS A SIMPLE TOOL FOR BOUNDING THE DEVIATION OF RANDOM MATRICES ON GEOMETRIC SETS CHRISTOPHER LIAW, ABBAS MEHRABIAN, YANIV PLAN, AND ROMAN VERSHYNIN Abstract. Let A be an isotropic, sub-gaussian m n matrix.

More information

Risk Bounds for Lévy Processes in the PAC-Learning Framework

Risk Bounds for Lévy Processes in the PAC-Learning Framework Risk Bounds for Lévy Processes in the PAC-Learning Framework Chao Zhang School of Computer Engineering anyang Technological University Dacheng Tao School of Computer Engineering anyang Technological University

More information

The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression

The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression Cun-hui Zhang and Jian Huang Presenter: Quefeng Li Feb. 26, 2010 un-hui Zhang and Jian Huang Presenter: Quefeng The Sparsity

More information

A Greedy Framework for First-Order Optimization

A Greedy Framework for First-Order Optimization A Greedy Framework for First-Order Optimization Jacob Steinhardt Department of Computer Science Stanford University Stanford, CA 94305 jsteinhardt@cs.stanford.edu Jonathan Huggins Department of EECS Massachusetts

More information

The volume growth of complete gradient shrinking Ricci solitons

The volume growth of complete gradient shrinking Ricci solitons arxiv:0904.0798v [math.dg] Apr 009 The volume growth of complete gradient shrinking Ricci solitons Ovidiu Munteanu Abstract We prove that any gradient shrinking Ricci soliton has at most Euclidean volume

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional

More information

APPLICATIONS OF DIFFERENTIABILITY IN R n.

APPLICATIONS OF DIFFERENTIABILITY IN R n. APPLICATIONS OF DIFFERENTIABILITY IN R n. MATANIA BEN-ARTZI April 2015 Functions here are defined on a subset T R n and take values in R m, where m can be smaller, equal or greater than n. The (open) ball

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional

More information

CONVEX RECOVERY OF A STRUCTURED SIGNAL FROM INDEPENDENT RANDOM LINEAR MEASUREMENTS

CONVEX RECOVERY OF A STRUCTURED SIGNAL FROM INDEPENDENT RANDOM LINEAR MEASUREMENTS CONVEX RECOVERY OF A STRUCTURED SIGNAL FROM INDEPENDENT RANDOM LINEAR MEASUREMENTS JOEL A. TROPP ABSTRACT. This chapter develops a theoretical analysis of the convex programming method for recovering a

More information

arxiv: v1 [math.pr] 22 Dec 2018

arxiv: v1 [math.pr] 22 Dec 2018 arxiv:1812.09618v1 [math.pr] 22 Dec 2018 Operator norm upper bound for sub-gaussian tailed random matrices Eric Benhamou Jamal Atif Rida Laraki December 27, 2018 Abstract This paper investigates an upper

More information

A note on the convex infimum convolution inequality

A note on the convex infimum convolution inequality A note on the convex infimum convolution inequality Naomi Feldheim, Arnaud Marsiglietti, Piotr Nayar, Jing Wang Abstract We characterize the symmetric measures which satisfy the one dimensional convex

More information

arxiv: v1 [math.st] 27 Sep 2018

arxiv: v1 [math.st] 27 Sep 2018 Robust covariance estimation under L 4 L 2 norm equivalence. arxiv:809.0462v [math.st] 27 Sep 208 Shahar Mendelson ikita Zhivotovskiy Abstract Let X be a centered random vector taking values in R d and

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini April 27, 2018 1 / 80 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

Existence, stability and instability for Einstein-scalar field Lichnerowicz equations by Emmanuel Hebey

Existence, stability and instability for Einstein-scalar field Lichnerowicz equations by Emmanuel Hebey Existence, stability and instability for Einstein-scalar field Lichnerowicz equations by Emmanuel Hebey Joint works with Olivier Druet and with Frank Pacard and Dan Pollack Two hours lectures IAS, October

More information

arxiv: v3 [cs.it] 25 Jul 2014

arxiv: v3 [cs.it] 25 Jul 2014 Simultaneously Structured Models with Application to Sparse and Low-rank Matrices Samet Oymak c, Amin Jalali w, Maryam Fazel w, Yonina C. Eldar t, Babak Hassibi c arxiv:11.3753v3 [cs.it] 5 Jul 014 c California

More information

Lecture 13 October 6, Covering Numbers and Maurey s Empirical Method

Lecture 13 October 6, Covering Numbers and Maurey s Empirical Method CS 395T: Sublinear Algorithms Fall 2016 Prof. Eric Price Lecture 13 October 6, 2016 Scribe: Kiyeon Jeon and Loc Hoang 1 Overview In the last lecture we covered the lower bound for p th moment (p > 2) and

More information

Sparse Recovery with Pre-Gaussian Random Matrices

Sparse Recovery with Pre-Gaussian Random Matrices Sparse Recovery with Pre-Gaussian Random Matrices Simon Foucart Laboratoire Jacques-Louis Lions Université Pierre et Marie Curie Paris, 75013, France Ming-Jun Lai Department of Mathematics University of

More information

Statistical learning with Lipschitz and convex loss functions

Statistical learning with Lipschitz and convex loss functions Statistical learning with Lipschitz and convex loss functions Geoffrey Chinot, Guillaume Lecué and Matthieu Lerasle October 3, 08 Abstract We obtain risk bounds for Empirical Risk Minimizers ERM and minmax

More information

Pointwise estimates for marginals of convex bodies

Pointwise estimates for marginals of convex bodies Journal of Functional Analysis 54 8) 75 93 www.elsevier.com/locate/jfa Pointwise estimates for marginals of convex bodies R. Eldan a,1,b.klartag b,, a School of Mathematical Sciences, Tel-Aviv University,

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57

More information

Compressive Sensing with Random Matrices

Compressive Sensing with Random Matrices Compressive Sensing with Random Matrices Lucas Connell University of Georgia 9 November 017 Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November 017 1 / 18 Overview

More information

Tools from Lebesgue integration

Tools from Lebesgue integration Tools from Lebesgue integration E.P. van den Ban Fall 2005 Introduction In these notes we describe some of the basic tools from the theory of Lebesgue integration. Definitions and results will be given

More information

Contents: 1. Minimization. 2. The theorem of Lions-Stampacchia for variational inequalities. 3. Γ -Convergence. 4. Duality mapping.

Contents: 1. Minimization. 2. The theorem of Lions-Stampacchia for variational inequalities. 3. Γ -Convergence. 4. Duality mapping. Minimization Contents: 1. Minimization. 2. The theorem of Lions-Stampacchia for variational inequalities. 3. Γ -Convergence. 4. Duality mapping. 1 Minimization A Topological Result. Let S be a topological

More information

Recent Developments in Compressed Sensing

Recent Developments in Compressed Sensing Recent Developments in Compressed Sensing M. Vidyasagar Distinguished Professor, IIT Hyderabad m.vidyasagar@iith.ac.in, www.iith.ac.in/ m vidyasagar/ ISL Seminar, Stanford University, 19 April 2018 Outline

More information

Optimization methods

Optimization methods Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,

More information

Convex inequalities, isoperimetry and spectral gap III

Convex inequalities, isoperimetry and spectral gap III Convex inequalities, isoperimetry and spectral gap III Jesús Bastero (Universidad de Zaragoza) CIDAMA Antequera, September 11, 2014 Part III. K-L-S spectral gap conjecture KLS estimate, through Milman's

More information

Signal Recovery from Permuted Observations

Signal Recovery from Permuted Observations EE381V Course Project Signal Recovery from Permuted Observations 1 Problem Shanshan Wu (sw33323) May 8th, 2015 We start with the following problem: let s R n be an unknown n-dimensional real-valued signal,

More information

Pseudo-Poincaré Inequalities and Applications to Sobolev Inequalities

Pseudo-Poincaré Inequalities and Applications to Sobolev Inequalities Pseudo-Poincaré Inequalities and Applications to Sobolev Inequalities Laurent Saloff-Coste Abstract Most smoothing procedures are via averaging. Pseudo-Poincaré inequalities give a basic L p -norm control

More information

AN EXAMPLE OF FUNCTIONAL WHICH IS WEAKLY LOWER SEMICONTINUOUS ON W 1,p FOR EVERY p > 2 BUT NOT ON H0

AN EXAMPLE OF FUNCTIONAL WHICH IS WEAKLY LOWER SEMICONTINUOUS ON W 1,p FOR EVERY p > 2 BUT NOT ON H0 AN EXAMPLE OF FUNCTIONAL WHICH IS WEAKLY LOWER SEMICONTINUOUS ON W,p FOR EVERY p > BUT NOT ON H FERNANDO FARRONI, RAFFAELLA GIOVA AND FRANÇOIS MURAT Abstract. In this note we give an example of functional

More information

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology

More information

Near Optimal Signal Recovery From Random Projections: Universal Encoding Strategies?

Near Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? Near Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? Emmanuel Candes and Terence Tao Applied and Computational Mathematics, Caltech, Pasadena, CA 91125 Department of Mathematics,

More information

Optimisation Combinatoire et Convexe.

Optimisation Combinatoire et Convexe. Optimisation Combinatoire et Convexe. Low complexity models, l 1 penalties. A. d Aspremont. M1 ENS. 1/36 Today Sparsity, low complexity models. l 1 -recovery results: three approaches. Extensions: matrix

More information

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider the problem of

More information

THE SMALLEST SINGULAR VALUE OF A RANDOM RECTANGULAR MATRIX

THE SMALLEST SINGULAR VALUE OF A RANDOM RECTANGULAR MATRIX THE SMALLEST SINGULAR VALUE OF A RANDOM RECTANGULAR MATRIX MARK RUDELSON AND ROMAN VERSHYNIN Abstract. We prove an optimal estimate on the smallest singular value of a random subgaussian matrix, valid

More information

RANDOM FIELDS AND GEOMETRY. Robert Adler and Jonathan Taylor

RANDOM FIELDS AND GEOMETRY. Robert Adler and Jonathan Taylor RANDOM FIELDS AND GEOMETRY from the book of the same name by Robert Adler and Jonathan Taylor IE&M, Technion, Israel, Statistics, Stanford, US. ie.technion.ac.il/adler.phtml www-stat.stanford.edu/ jtaylor

More information

The adaptive and the thresholded Lasso for potentially misspecified models (and a lower bound for the Lasso)

The adaptive and the thresholded Lasso for potentially misspecified models (and a lower bound for the Lasso) Electronic Journal of Statistics Vol. 0 (2010) ISSN: 1935-7524 The adaptive the thresholded Lasso for potentially misspecified models ( a lower bound for the Lasso) Sara van de Geer Peter Bühlmann Seminar

More information

Daniel M. Oberlin Department of Mathematics, Florida State University. January 2005

Daniel M. Oberlin Department of Mathematics, Florida State University. January 2005 PACKING SPHERES AND FRACTAL STRICHARTZ ESTIMATES IN R d FOR d 3 Daniel M. Oberlin Department of Mathematics, Florida State University January 005 Fix a dimension d and for x R d and r > 0, let Sx, r) stand

More information

A Unified Approach to Proximal Algorithms using Bregman Distance

A Unified Approach to Proximal Algorithms using Bregman Distance A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department

More information

On Acceleration with Noise-Corrupted Gradients. + m k 1 (x). By the definition of Bregman divergence:

On Acceleration with Noise-Corrupted Gradients. + m k 1 (x). By the definition of Bregman divergence: A Omitted Proofs from Section 3 Proof of Lemma 3 Let m x) = a i On Acceleration with Noise-Corrupted Gradients fxi ), u x i D ψ u, x 0 ) denote the function under the minimum in the lower bound By Proposition

More information

Z Algorithmic Superpower Randomization October 15th, Lecture 12

Z Algorithmic Superpower Randomization October 15th, Lecture 12 15.859-Z Algorithmic Superpower Randomization October 15th, 014 Lecture 1 Lecturer: Bernhard Haeupler Scribe: Goran Žužić Today s lecture is about finding sparse solutions to linear systems. The problem

More information

MATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous:

MATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous: MATH 51H Section 4 October 16, 2015 1 Continuity Recall what it means for a function between metric spaces to be continuous: Definition. Let (X, d X ), (Y, d Y ) be metric spaces. A function f : X Y is

More information

arxiv:math/ v1 [math.fa] 30 Sep 2004

arxiv:math/ v1 [math.fa] 30 Sep 2004 arxiv:math/04000v [math.fa] 30 Sep 2004 Small ball probability and Dvoretzy Theorem B. Klartag, R. Vershynin Abstract Large deviation estimates are by now a standard tool in the Asymptotic Convex Geometry,

More information

4 Sobolev spaces, trace theorem and normal derivative

4 Sobolev spaces, trace theorem and normal derivative 4 Sobolev spaces, trace theorem and normal derivative Throughout, n will be a sufficiently smooth, bounded domain. We use the standard Sobolev spaces H 0 ( n ) := L 2 ( n ), H 0 () := L 2 (), H k ( n ),

More information

ON THE REGULARITY OF SAMPLE PATHS OF SUB-ELLIPTIC DIFFUSIONS ON MANIFOLDS

ON THE REGULARITY OF SAMPLE PATHS OF SUB-ELLIPTIC DIFFUSIONS ON MANIFOLDS Bendikov, A. and Saloff-Coste, L. Osaka J. Math. 4 (5), 677 7 ON THE REGULARITY OF SAMPLE PATHS OF SUB-ELLIPTIC DIFFUSIONS ON MANIFOLDS ALEXANDER BENDIKOV and LAURENT SALOFF-COSTE (Received March 4, 4)

More information

The Canonical Gaussian Measure on R

The Canonical Gaussian Measure on R The Canonical Gaussian Measure on R 1. Introduction The main goal of this course is to study Gaussian measures. The simplest example of a Gaussian measure is the canonical Gaussian measure P on R where

More information

Math 259: Introduction to Analytic Number Theory Functions of finite order: product formula and logarithmic derivative

Math 259: Introduction to Analytic Number Theory Functions of finite order: product formula and logarithmic derivative Math 259: Introduction to Analytic Number Theory Functions of finite order: product formula and logarithmic derivative This chapter is another review of standard material in complex analysis. See for instance

More information

Inégalités de dispersion via le semi-groupe de la chaleur

Inégalités de dispersion via le semi-groupe de la chaleur Inégalités de dispersion via le semi-groupe de la chaleur Valentin Samoyeau, Advisor: Frédéric Bernicot. Laboratoire de Mathématiques Jean Leray, Université de Nantes January 28, 2016 1 Introduction Schrödinger

More information

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Suppose again we have n sample points x,..., x n R p. The data-point x i R p can be thought of as the i-th row X i of an n p-dimensional

More information

Lecture 6: September 19

Lecture 6: September 19 36-755: Advanced Statistical Theory I Fall 2016 Lecture 6: September 19 Lecturer: Alessandro Rinaldo Scribe: YJ Choe Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have

More information

Convex Analysis and Economic Theory AY Elementary properties of convex functions

Convex Analysis and Economic Theory AY Elementary properties of convex functions Division of the Humanities and Social Sciences Ec 181 KC Border Convex Analysis and Economic Theory AY 2018 2019 Topic 6: Convex functions I 6.1 Elementary properties of convex functions We may occasionally

More information

An Algorithmist s Toolkit Nov. 10, Lecture 17

An Algorithmist s Toolkit Nov. 10, Lecture 17 8.409 An Algorithmist s Toolkit Nov. 0, 009 Lecturer: Jonathan Kelner Lecture 7 Johnson-Lindenstrauss Theorem. Recap We first recap a theorem (isoperimetric inequality) and a lemma (concentration) from

More information

LECTURE 7. Least Squares and Variants. Optimization Models EE 127 / EE 227AT. Outline. Least Squares. Notes. Notes. Notes. Notes.

LECTURE 7. Least Squares and Variants. Optimization Models EE 127 / EE 227AT. Outline. Least Squares. Notes. Notes. Notes. Notes. Optimization Models EE 127 / EE 227AT Laurent El Ghaoui EECS department UC Berkeley Spring 2015 Sp 15 1 / 23 LECTURE 7 Least Squares and Variants If others would but reflect on mathematical truths as deeply

More information

Optimization for Compressed Sensing

Optimization for Compressed Sensing Optimization for Compressed Sensing Robert J. Vanderbei 2014 March 21 Dept. of Industrial & Systems Engineering University of Florida http://www.princeton.edu/ rvdb Lasso Regression The problem is to solve

More information

Composite Objective Mirror Descent

Composite Objective Mirror Descent Composite Objective Mirror Descent John C. Duchi 1,3 Shai Shalev-Shwartz 2 Yoram Singer 3 Ambuj Tewari 4 1 University of California, Berkeley 2 Hebrew University of Jerusalem, Israel 3 Google Research

More information

1 Sparsity and l 1 relaxation

1 Sparsity and l 1 relaxation 6.883 Learning with Combinatorial Structure Note for Lecture 2 Author: Chiyuan Zhang Sparsity and l relaxation Last time we talked about sparsity and characterized when an l relaxation could recover the

More information

Rademacher Averages and Phase Transitions in Glivenko Cantelli Classes

Rademacher Averages and Phase Transitions in Glivenko Cantelli Classes IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 1, JANUARY 2002 251 Rademacher Averages Phase Transitions in Glivenko Cantelli Classes Shahar Mendelson Abstract We introduce a new parameter which

More information

arxiv: v2 [math.st] 30 Nov 2017

arxiv: v2 [math.st] 30 Nov 2017 Robust machine learning by median-of-means : theory and practice G. Lecué and M. Lerasle December 4, 2017 arxiv:1711.10306v2 [math.st] 30 Nov 2017 Abstract We introduce new estimators for robust machine

More information

Convexity of chance constraints with dependent random variables: the use of copulae

Convexity of chance constraints with dependent random variables: the use of copulae Convexity of chance constraints with dependent random variables: the use of copulae René Henrion 1 and Cyrille Strugarek 2 1 Weierstrass Institute for Applied Analysis and Stochastics, 10117 Berlin, Germany.

More information

(Part 1) High-dimensional statistics May / 41

(Part 1) High-dimensional statistics May / 41 Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2

More information

An example of a convex body without symmetric projections.

An example of a convex body without symmetric projections. An example of a convex body without symmetric projections. E. D. Gluskin A. E. Litvak N. Tomczak-Jaegermann Abstract Many crucial results of the asymptotic theory of symmetric convex bodies were extended

More information

Strengthened Sobolev inequalities for a random subspace of functions

Strengthened Sobolev inequalities for a random subspace of functions Strengthened Sobolev inequalities for a random subspace of functions Rachel Ward University of Texas at Austin April 2013 2 Discrete Sobolev inequalities Proposition (Sobolev inequality for discrete images)

More information

Optimization and Optimal Control in Banach Spaces

Optimization and Optimal Control in Banach Spaces Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,

More information

REPRESENTABLE BANACH SPACES AND UNIFORMLY GÂTEAUX-SMOOTH NORMS. Julien Frontisi

REPRESENTABLE BANACH SPACES AND UNIFORMLY GÂTEAUX-SMOOTH NORMS. Julien Frontisi Serdica Math. J. 22 (1996), 33-38 REPRESENTABLE BANACH SPACES AND UNIFORMLY GÂTEAUX-SMOOTH NORMS Julien Frontisi Communicated by G. Godefroy Abstract. It is proved that a representable non-separable Banach

More information

ROW PRODUCTS OF RANDOM MATRICES

ROW PRODUCTS OF RANDOM MATRICES ROW PRODUCTS OF RANDOM MATRICES MARK RUDELSON Abstract Let 1,, K be d n matrices We define the row product of these matrices as a d K n matrix, whose rows are entry-wise products of rows of 1,, K This

More information

Contents 1. Introduction 1 2. Main results 3 3. Proof of the main inequalities 7 4. Application to random dynamical systems 11 References 16

Contents 1. Introduction 1 2. Main results 3 3. Proof of the main inequalities 7 4. Application to random dynamical systems 11 References 16 WEIGHTED CSISZÁR-KULLBACK-PINSKER INEQUALITIES AND APPLICATIONS TO TRANSPORTATION INEQUALITIES FRANÇOIS BOLLEY AND CÉDRIC VILLANI Abstract. We strengthen the usual Csiszár-Kullback-Pinsker inequality by

More information

arxiv: v1 [math.pr] 11 Feb 2019

arxiv: v1 [math.pr] 11 Feb 2019 A Short Note on Concentration Inequalities for Random Vectors with SubGaussian Norm arxiv:190.03736v1 math.pr] 11 Feb 019 Chi Jin University of California, Berkeley chijin@cs.berkeley.edu Rong Ge Duke

More information

THE LASSO, CORRELATED DESIGN, AND IMPROVED ORACLE INEQUALITIES. By Sara van de Geer and Johannes Lederer. ETH Zürich

THE LASSO, CORRELATED DESIGN, AND IMPROVED ORACLE INEQUALITIES. By Sara van de Geer and Johannes Lederer. ETH Zürich Submitted to the Annals of Applied Statistics arxiv: math.pr/0000000 THE LASSO, CORRELATED DESIGN, AND IMPROVED ORACLE INEQUALITIES By Sara van de Geer and Johannes Lederer ETH Zürich We study high-dimensional

More information

Math 259: Introduction to Analytic Number Theory Functions of finite order: product formula and logarithmic derivative

Math 259: Introduction to Analytic Number Theory Functions of finite order: product formula and logarithmic derivative Math 259: Introduction to Analytic Number Theory Functions of finite order: product formula and logarithmic derivative This chapter is another review of standard material in complex analysis. See for instance

More information

Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures

Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures S.G. Bobkov School of Mathematics, University of Minnesota, 127 Vincent Hall, 26 Church St. S.E., Minneapolis, MN 55455,

More information