Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas
|
|
- Maryann Thomas
- 5 years ago
- Views:
Transcription
1 Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas Department of Mathematics Department of Statistical Science Cornell University London, January 7, 2016
2 Joint work with Yue Zhao Supported by NSF Grant DMS
3 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3
4 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3
5 Copula Let the continuous random vector X = (X 1,..., X d ) R d have marginal distribution functions F j (x) = P { X j x }. The copula function C(u 1,..., u d ) is the joint cumulative distribution function of the transformation U = (F 1 (X 1 ),..., F d (X d ) ), i.e., for u = (u 1,..., u d ) [0, 1] d, C(u 1,..., u d ) = P {F 1 (X 1 ) u 1,..., F d (X d ) u d }.
6 Copula The copula C always describes a distribution on the unit square. Copula is invariant under strictly increasing transformations T j of the marginals X j allows separate specifications of the marginals and the dependence structure. In this talk we mostly study semiparametric copula models. The dependence structure is given by a parametric copula, while the marginals are left unspecified (nonparametric).
7 Elliptical distribution A random vector X R d has an elliptical distribution if its characteristic function E(exp(it X)) can be written as e it µ Ψ(t Σt) with parameters µ R d, covariance matrix Σ R d d, and characteristic generator Ψ : [0, ) R. For instance, when Ψ(t) = e t/2, X N(µ, Σ).
8 Semiparametric elliptical distributions and the copula correlation matrix Elliptical copulas are the copulas of elliptical distributions. Elliptical copula is characterized by copula parameters: Ψ and the copula correlation matrix Σ, with Σ ij = Σ ij Σ ii Σ jj.
9 Semiparametric Gaussian graphical copula model Let X N(0, Σ), and Ω = Σ 1 be the precision matrix. If [Ω] kl = 0, then X k X l X \{k,l}. Accurate estimator of Ω can be obtained by feeding the sample covariance matrix Σ into the graphical lasso, CLIME, graphical Dantzig selector, etc. Assume instead that we merely have f (X) N(0, Σ), for some strictly increasing, unknown function f (i.e., X has a Gaussian copula with copula correlation matrix Σ), and again we let Ω = Σ 1. Now, if [Ω] kl = 0, then f (X k ) f (X l ) f (X) \{k,l} = X k X l X \{k,l}.
10 Semiparametric Gaussian graphical copula model Using a rank-based estimator Σ rank-based (using Kendall s tau or Spearman s rho) of Σ, Liu et al 2012, Xue & Zou 2012 recovers the same optimal parametric rate (obtained under the Gaussian model) under the semiparametric model. This crucially depends on estimating Σ rank-based Σ max at the rate of log(d)/n (w.h.p.), which is in turn based on Hoeffding s classical bound for scalar U-statistic.
11 Kendall s tau Let X R d be an independent copy of X. Kendall s tau between the kth and lth coordinates is [ ( τ kl = E sgn (X k X k )(X l X )] l ) Let T be the matrix of Kendall s tau: [T ] kl = τ kl.
12 Kendall s tau Kendall s tau statistic is, for samples X 1,..., X n that are independent copies of X, τ kl = 2 n(n 1) 1 i<j d ) sgn ((X k i X j k )(X l i X j l ). Let T be the matrix of Kendall s tau statistics: [ T ] kl = τ kl.
13 Kendall s tau matrix as plug-in estimator For elliptical copulas, independent of the characteristic generator, ( π ) Σ = sin 2 T with the sine function acting component-wise (Kruskal 1958, Kendall & Gibbons 1990, Fang, Fang & Kotz 2002). Hence, the natural plug-in estimator for Σ is ( π Σ = sin 2 T ).
14 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3
15 For the rest of the talk, we assume X R d follows a semiparametric elliptical distribution, The copula correlation matrix of X is Σ R d d, The plug-in estimator Σ = Σ n of Σ is constructed from n independent copies of X, (e.g., Demarta & McNeil 2004, Klüppelberg & Kuhn 2009)
16 Some recent advanced in elliptical/gaussian copulas Klüppelberg & Kuhn 2009 study the property of Σ in the asymptotic regime, i.e., d fixed, n. Liu et al 2012, Xue & Zou 2012 study estimation of the precision matrix Σ 1 of Gaussian copulas in high dimensions. Sharp bound on Σ Σ max is a key step of the study.
17 Proposed Research: Part I First, we establish a sharp bound of the operator norm Σ Σ 2. (See also Han & Liu 2013.) It has often been observed (e.g., Demarta & McNeil 2004, Klüppelberg & Kuhn 2009) that the plug-in estimator Σ is not always positive semidefinite. A bound on Σ Σ 2 quantifies the extent to which the non-positive semidefinite problem may happen. If Σ is not positive semidefinite, we can project it onto the set of correlation matrices in some specific way to obtain Σ +. Furthermore, Σ + Σ 2 2 Σ Σ 2.
18 Proposed Research: Part II Later, we study a factor model for Σ. The factor model assumes that Σ = Θ + V for a low-rank (or nearly low-rank) matrix Θ and a diagonal matrix V. X and [L, V 1/2 ]Y have the same copula for some elliptical distribution with Σ = I r+d. Here Θ = LL for some L R d r. Within this context, the effective number of parameters is O(r d), for r = rank(θ), instead of O(d 2 ) under the generic model. We propose a refined estimator Σ for Σ using a penalized least square procedure. The bound on Σ Σ 2 serves to scale the penalty term.
19 Bounding T T 2 Bounding Σ Σ 2 Outline for bounding Σ Σ 2 Bounding T T 2 (matrix counterpart of Hoeffding 1963 classical bound for scalar U-statistic). Bounding Σ Σ 2 in terms of T T 2 (matrix counterpart of the Lipschitz property).
20 Bounding T T 2 Bounding Σ Σ 2 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3
21 Bounding T T 2 Bounding Σ Σ 2 Matrix U-statistic Theorem With probability at least 1 α, we have with { T T 2 < max T 2 f n,d,α, fn,d,α 2 T 2 fn,d,α 2 + f n,d,α 2 /4 + f n,d,α 2 { } /2 < max T 2 f n,d,α, fn,d,α 2 + fn,d,α 2 16 f n,d,α = 3 d log(2α 1 d). n The above bounds hold for all n, d, α. }
22 Bounding T T 2 Bounding Σ Σ 2 Matrix U-statistic: proof We need a Bernstein-type inequality, { e.g., } Tropp (2012), to bound the tail probability P T T 2 t. That theorem, as many other results on bounding the tail probability of the operator norm of a sum of random matrices, requires that the summands be independent. The U-statistics T does not satisfy this condition. Many results on tail bound rely on the Laplace transform technique to convert the tail probability into an expectation of some convex function of T T. A technique pointed out by Hoeffding (1963) allows us to convert the problem of bounding T T 2 into a problem involving a sum of independent random matrices.
23 Bounding T T 2 Bounding Σ Σ 2 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3
24 Bounding T T 2 Bounding Σ Σ 2 Matrix Lipschitz property Theorem We have, with probability at least 1 α, that Σ Σ 2 π T T 2 + 3π2 16 f 2 n,d,α. Combining earlier bound for T T 2, we have, with probability at least 1 2α, that Σ Σ 2 < π max { T 2 f n,d,α, f 2 n,d,α } + 3π2 16 f 2 n,d,α. In addition, 2 π Σ 2 T 2 Σ 2.
25 Bounding T T 2 Bounding Σ Σ 2 Potential improvement We can obtain from the corollary that, for n d, Σ 2 d log(d) E Σ Σ 2. n In contrast, for Gaussian X and Σ its sample covariance matrix, Koltchinskii & Lounici 2015 shows that, for n d, Σ 2 d E Σ Σ 2. n
26 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3
27 A factor model for Σ From now on, we assume that the copula correlation matrix Σ satisfies a factor model: Σ = Θ + V for some low-rank (or nearly low-rank), positive semidefinite, symmetric matrix Θ and diagonal matrix V.
28 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3
29 Prelude: an elementary factor model for Σ In an elementary factor model for Σ = Θ + V, we assume that rank(θ) = r < d, and V = σ 2 I d. Let the plug-in estimator Σ have the eigen-decomposition d λ k û k û k, λ1 λ d. k=1
30 Prelude: an elementary factor model for Σ Construct the following closed-form estimators for r, σ 2, Θ and Σ, based on some threshold µ: r = σ 2 = Θ = d k=1 1 d r r k=1 1 { λk λ } d + µ, λ j, j> r ( λ k σ 2 )û k û k Σ = Θ diag( Θ) + I.
31 Prelude: an elementary factor model for Σ, continued Theorem Assume λ r (Θ) 2µ. On the event { } 2 Σ Σ 2 < µ, we have r = r, Σ Σ 2 F Θ Θ 2 F 8r Σ Σ 2 2, σ 2 σ 2 Σ Σ 2. Here, F denotes the Frobenius norm.
32 Prelude: an elementary factor model for Σ, continued Corollary Set µ = 8 T 2 fn,d,α 2 + f n,d,α 4 /4 + 8f n,d,α 2 µ = 8 T 2 f n,d,α + 12fn,d,α 2. Assume λ r (Θ) 2 µ, and T 2 fn,d,α 2. Then, Σ Σ 2 F 2r µ2, hold with probability exceeding 1 2α.
33 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3
34 The refined estimator Σ of Σ = Θ + V In the factor model we can write Σ = Θ diag(θ) + I d := Θ o + I d. This motivates us to estimate the low-rank matrix Θ by Θ as the solution to the convex program { } 1 Θ = argmin Θ R d d 2 Θ o Σ o 2 F + µ Θ. Here, denotes the nuclear norm. Finally, we let the refined estimator Σ of the copula correlation matrix Σ be Σ = Θ o + I d.
35 Comparison of Σ with existing estimators Saunderson et al 2012 study a factor model for Σ in the noiseless setting using minimum trace factor analysis: Σ = Θ + V (Θ, V ) = argmin tr(θ ) subject to Θ 0 V diagonal If Θ has column space U such that its coherence max P Ue i 2 < 1/ 2, 1 i d then the above algorithm recovers Θ, V exactly.
36 Comparison of Σ with existing estimators, continued For Θ with effective low-rank and without sparse corruption, Lounici 2013 proposes { } 1 Θ = argmin Θ R d d 2 Θ Θ 2 F + µ Θ for an unbiased initial estimator Θ of Θ (which we lack). For Σ = Θ + S with low-rank matrix Θ and generic sparse matrix S, Chandrasekaran et al 2012, Hsu et al 2011 propose ( Θ, S) = argmin Θ,S R d d This formulation requires λ > 0. { } 1 2 Θ + S Σ 2 F + µ Θ + λ S l1.
37 Comparison of Σ with existing estimators, continued We are in between the above scenarios. We will extend the result of the former and relax the condition of the latter with more precise analysis.
38 Oracle inequality for the refined estimator Σ Theorem Let Θ r be best rank-r approximation to Θ, with reduced SVD Θ r = U r D r U r. Set µ = 10 T 2 f 2 n,d,α + f 4 n,d,α /4 + 10f 2 n,d,α, µ = 10 max [ T 2 f n,d,α, f 2 n,d,α Then, with probability at least 1 2α, ] + 15f 2 n,d,α. Σ Σ 2 F j>r λ 2 j (Θ) + 18r µ2 for all r with γ r = U r U r max 1/9.
39 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3
40 Simulation result The object of interest is the ratio Σ Σ 2 F, which our Σ Σ 2 F theory predicts to be roughly proportional to d/r as n. We (somewhat arbitrarily) set C = 0.1, α = 0.1, and forgo the log(d) factor when choosing the regularization parameter. We estimate this ratio by simulation.
41 Simulation setup Σ 2 = (Larger is better) 10 2 k b '! 'k 2 F =je '! 'k 2 F d=128, r=1 d=32, r=1 d=128, r=4 d=8, r=1 d=32, r= d=8, r= Sample size
42 Summary We have obtained a sharp bound on the operator norm of Σ Σ, for the plug-in estimator Σ of the copula correlation matrix Σ. We have applied the above bound to the setting of a factor model of Σ to obtain a refined estimator Σ; a sharp oracle inequality for Σ is established.
43 Thanks!
The Nonparanormal skeptic
The Nonpara skeptic Han Liu Johns Hopkins University, 615 N. Wolfe Street, Baltimore, MD 21205 USA Fang Han Johns Hopkins University, 615 N. Wolfe Street, Baltimore, MD 21205 USA Ming Yuan Georgia Institute
More informationDependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline.
MFM Practitioner Module: Risk & Asset Allocation September 11, 2013 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y
More informationRobust Sparse Principal Component Regression under the High Dimensional Elliptical Model
Robust Sparse Principal Component Regression under the High Dimensional Elliptical Model Fang Han Department of Biostatistics Johns Hopkins University Baltimore, MD 21210 fhan@jhsph.edu Han Liu Department
More informationDependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.
Practitioner Course: Portfolio Optimization September 10, 2008 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y ) (x,
More informationLearning discrete graphical models via generalized inverse covariance matrices
Learning discrete graphical models via generalized inverse covariance matrices Duzhe Wang, Yiming Lv, Yongjoon Kim, Young Lee Department of Statistics University of Wisconsin-Madison {dwang282, lv23, ykim676,
More informationHigh Dimensional Semiparametric Gaussian Copula Graphical Models
High Dimeional Semiparametric Gaussian Copula Graphical Models Han Liu, Fang Han, Ming Yuan, John Lafferty and Larry Wasserman February 9, 2012 arxiv:1202.2169v1 [stat.ml] 10 Feb 2012 Contents Abstract:
More informationHigh-dimensional covariance estimation based on Gaussian graphical models
High-dimensional covariance estimation based on Gaussian graphical models Shuheng Zhou Department of Statistics, The University of Michigan, Ann Arbor IMA workshop on High Dimensional Phenomena Sept. 26,
More informationFinancial Econometrics and Volatility Models Copulas
Financial Econometrics and Volatility Models Copulas Eric Zivot Updated: May 10, 2010 Reading MFTS, chapter 19 FMUND, chapters 6 and 7 Introduction Capturing co-movement between financial asset returns
More informationAn iterative hard thresholding estimator for low rank matrix recovery
An iterative hard thresholding estimator for low rank matrix recovery Alexandra Carpentier - based on a joint work with Arlene K.Y. Kim Statistical Laboratory, Department of Pure Mathematics and Mathematical
More informationHigh Dimensional Inverse Covariate Matrix Estimation via Linear Programming
High Dimensional Inverse Covariate Matrix Estimation via Linear Programming Ming Yuan October 24, 2011 Gaussian Graphical Model X = (X 1,..., X p ) indep. N(µ, Σ) Inverse covariance matrix Σ 1 = Ω = (ω
More informationRobust and sparse Gaussian graphical modelling under cell-wise contamination
Robust and sparse Gaussian graphical modelling under cell-wise contamination Shota Katayama 1, Hironori Fujisawa 2 and Mathias Drton 3 1 Tokyo Institute of Technology, Japan 2 The Institute of Statistical
More informationCombining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation
UIUC CSL Mar. 24 Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation Yuejie Chi Department of ECE and BMI Ohio State University Joint work with Yuxin Chen (Stanford).
More informationSolving Corrupted Quadratic Equations, Provably
Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin
More informationHigh Dimensional Covariance and Precision Matrix Estimation
High Dimensional Covariance and Precision Matrix Estimation Wei Wang Washington University in St. Louis Thursday 23 rd February, 2017 Wei Wang (Washington University in St. Louis) High Dimensional Covariance
More informationMulti-stage convex relaxation approach for low-rank structured PSD matrix recovery
Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun
More informationCausal Inference: Discussion
Causal Inference: Discussion Mladen Kolar The University of Chicago Booth School of Business Sept 23, 2016 Types of machine learning problems Based on the information available: Supervised learning Reinforcement
More informationEfficient estimation of a semiparametric dynamic copula model
Efficient estimation of a semiparametric dynamic copula model Christian Hafner Olga Reznikova Institute of Statistics Université catholique de Louvain Louvain-la-Neuve, Blgium 30 January 2009 Young Researchers
More informationVariations on Nonparametric Additive Models: Computational and Statistical Aspects
Variations on Nonparametric Additive Models: Computational and Statistical Aspects John Lafferty Department of Statistics & Department of Computer Science University of Chicago Collaborators Sivaraman
More informationRobust Principal Component Analysis
ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M
More informationAn efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss
An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss arxiv:1811.04545v1 [stat.co] 12 Nov 2018 Cheng Wang School of Mathematical Sciences, Shanghai Jiao
More informationSparse Covariance Matrix Estimation with Eigenvalue Constraints
Sparse Covariance Matrix Estimation with Eigenvalue Constraints Han Liu and Lie Wang 2 and Tuo Zhao 3 Department of Operations Research and Financial Engineering, Princeton University 2 Department of Mathematics,
More informationMatrix concentration inequalities
ELE 538B: Mathematics of High-Dimensional Data Matrix concentration inequalities Yuxin Chen Princeton University, Fall 2018 Recap: matrix Bernstein inequality Consider a sequence of independent random
More informationRecovery of Simultaneously Structured Models using Convex Optimization
Recovery of Simultaneously Structured Models using Convex Optimization Maryam Fazel University of Washington Joint work with: Amin Jalali (UW), Samet Oymak and Babak Hassibi (Caltech) Yonina Eldar (Technion)
More informationModel Selection and Geometry
Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model
More informationConstruction and estimation of high dimensional copulas
Construction and estimation of high dimensional copulas Gildas Mazo PhD work supervised by S. Girard and F. Forbes Mistis, Inria and laboratoire Jean Kuntzmann, Grenoble, France Séminaire Statistiques,
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57
More informationIntrinsic Volumes of Convex Cones Theory and Applications
Intrinsic Volumes of Convex Cones Theory and Applications M\cr NA Manchester Numerical Analysis Martin Lotz School of Mathematics The University of Manchester with the collaboration of Dennis Amelunxen,
More informationProperties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation
Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation Adam J. Rothman School of Statistics University of Minnesota October 8, 2014, joint work with Liliana
More informationReconstruction from Anisotropic Random Measurements
Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013
More informationSTATISTICAL MACHINE LEARNING FOR STRUCTURED AND HIGH DIMENSIONAL DATA
AFRL-OSR-VA-TR-2014-0234 STATISTICAL MACHINE LEARNING FOR STRUCTURED AND HIGH DIMENSIONAL DATA Larry Wasserman CARNEGIE MELLON UNIVERSITY 0 Final Report DISTRIBUTION A: Distribution approved for public
More informationLeast squares under convex constraint
Stanford University Questions Let Z be an n-dimensional standard Gaussian random vector. Let µ be a point in R n and let Y = Z + µ. We are interested in estimating µ from the data vector Y, under the assumption
More informationFast and Robust Phase Retrieval
Fast and Robust Phase Retrieval Aditya Viswanathan aditya@math.msu.edu CCAM Lunch Seminar Purdue University April 18 2014 0 / 27 Joint work with Yang Wang Mark Iwen Research supported in part by National
More informationThe deterministic Lasso
The deterministic Lasso Sara van de Geer Seminar für Statistik, ETH Zürich Abstract We study high-dimensional generalized linear models and empirical risk minimization using the Lasso An oracle inequality
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationsparse and low-rank tensor recovery Cubic-Sketching
Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru
More informationModelling Dependent Credit Risks
Modelling Dependent Credit Risks Filip Lindskog, RiskLab, ETH Zürich 30 November 2000 Home page:http://www.math.ethz.ch/ lindskog E-mail:lindskog@math.ethz.ch RiskLab:http://www.risklab.ch Modelling Dependent
More informationConfidence Intervals for Low-dimensional Parameters with High-dimensional Data
Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology
More informationStein s Method for Matrix Concentration
Stein s Method for Matrix Concentration Lester Mackey Collaborators: Michael I. Jordan, Richard Y. Chen, Brendan Farrell, and Joel A. Tropp University of California, Berkeley California Institute of Technology
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE7C (Spring 08): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee7c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee7c@berkeley.edu October
More information(Part 1) High-dimensional statistics May / 41
Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2
More informationELE 538B: Mathematics of High-Dimensional Data. Spectral methods. Yuxin Chen Princeton University, Fall 2018
ELE 538B: Mathematics of High-Dimensional Data Spectral methods Yuxin Chen Princeton University, Fall 2018 Outline A motivating application: graph clustering Distance and angles between two subspaces Eigen-space
More informationarxiv: v1 [math.st] 13 Feb 2012
Sparse Matrix Inversion with Scaled Lasso Tingni Sun and Cun-Hui Zhang Rutgers University arxiv:1202.2723v1 [math.st] 13 Feb 2012 Address: Department of Statistics and Biostatistics, Hill Center, Busch
More informationUpper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1
Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1 Feng Wei 2 University of Michigan July 29, 2016 1 This presentation is based a project under the supervision of M. Rudelson.
More informationA measure of radial asymmetry for bivariate copulas based on Sobolev norm
A measure of radial asymmetry for bivariate copulas based on Sobolev norm Ahmad Alikhani-Vafa Ali Dolati Abstract The modified Sobolev norm is used to construct an index for measuring the degree of radial
More informationFundamentals of Matrices
Maschinelles Lernen II Fundamentals of Matrices Christoph Sawade/Niels Landwehr/Blaine Nelson Tobias Scheffer Matrix Examples Recap: Data Linear Model: f i x = w i T x Let X = x x n be the data matrix
More informationCan we do statistical inference in a non-asymptotic way? 1
Can we do statistical inference in a non-asymptotic way? 1 Guang Cheng 2 Statistics@Purdue www.science.purdue.edu/bigdata/ ONR Review Meeting@Duke Oct 11, 2017 1 Acknowledge NSF, ONR and Simons Foundation.
More informationProbabilistic Engineering Mechanics. An innovating analysis of the Nataf transformation from the copula viewpoint
Probabilistic Engineering Mechanics 4 9 3 3 Contents lists available at ScienceDirect Probabilistic Engineering Mechanics journal homepage: www.elsevier.com/locate/probengmech An innovating analysis of
More informationECA: High Dimensional Elliptical Component Analysis in non-gaussian Distributions
: High Dimensional Elliptical Component Analysis in non-gaussian Distributions Fang Han and Han Liu arxiv:1310.3561v4 [stat.ml] 3 Oct 016 Abstract We present a robust alternative to principal component
More informationMultivariate Normal-Laplace Distribution and Processes
CHAPTER 4 Multivariate Normal-Laplace Distribution and Processes The normal-laplace distribution, which results from the convolution of independent normal and Laplace random variables is introduced by
More informationAdaptive one-bit matrix completion
Adaptive one-bit matrix completion Joseph Salmon Télécom Paristech, Institut Mines-Télécom Joint work with Jean Lafond (Télécom Paristech) Olga Klopp (Crest / MODAL X, Université Paris Ouest) Éric Moulines
More informationJournal of Multivariate Analysis. Consistency of sparse PCA in High Dimension, Low Sample Size contexts
Journal of Multivariate Analysis 5 (03) 37 333 Contents lists available at SciVerse ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Consistency of sparse PCA
More informationStatistical Machine Learning for Structured and High Dimensional Data
Statistical Machine Learning for Structured and High Dimensional Data (FA9550-09- 1-0373) PI: Larry Wasserman (CMU) Co- PI: John Lafferty (UChicago and CMU) AFOSR Program Review (Jan 28-31, 2013, Washington,
More informationNumerical schemes of resolution of stochastic optimal control HJB equation
Numerical schemes of resolution of stochastic optimal control HJB equation Elisabeth Ottenwaelter Journée des doctorants 7 mars 2007 Équipe COMMANDS Elisabeth Ottenwaelter (INRIA et CMAP) Numerical schemes
More informationarxiv: v2 [math.st] 12 Feb 2008
arxiv:080.460v2 [math.st] 2 Feb 2008 Electronic Journal of Statistics Vol. 2 2008 90 02 ISSN: 935-7524 DOI: 0.24/08-EJS77 Sup-norm convergence rate and sign concentration property of Lasso and Dantzig
More informationthe long tau-path for detecting monotone association in an unspecified subpopulation
the long tau-path for detecting monotone association in an unspecified subpopulation Joe Verducci Current Challenges in Statistical Learning Workshop Banff International Research Station Tuesday, December
More informationA Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression
A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression Noah Simon Jerome Friedman Trevor Hastie November 5, 013 Abstract In this paper we purpose a blockwise descent
More informationσ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =
Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,
More informationRegularization Methods for Prediction in Dynamic Graphs and e-marketing Applications
Regularization Methods for Prediction in Dynamic Graphs and e-marketing Applications Emile Richard CMLA-ENS Cachan 1000mercis PhD defense Advisors: Th. Evgeniou (INSEAD), N. Vayatis (CMLA-ENS Cachan) November
More informationA New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables
A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider the problem of
More informationComputational and Statistical Aspects of Statistical Machine Learning. John Lafferty Department of Statistics Retreat Gleacher Center
Computational and Statistical Aspects of Statistical Machine Learning John Lafferty Department of Statistics Retreat Gleacher Center Outline Modern nonparametric inference for high dimensional data Nonparametric
More informationLecture 5 : Projections
Lecture 5 : Projections EE227C. Lecturer: Professor Martin Wainwright. Scribe: Alvin Wan Up until now, we have seen convergence rates of unconstrained gradient descent. Now, we consider a constrained minimization
More informationModelling and Estimation of Stochastic Dependence
Modelling and Estimation of Stochastic Dependence Uwe Schmock Based on joint work with Dr. Barbara Dengler Financial and Actuarial Mathematics and Christian Doppler Laboratory for Portfolio Risk Management
More informationProbabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms
Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationSensitivity Analysis for the Problem of Matrix Joint Diagonalization
Sensitivity Analysis for the Problem of Matrix Joint Diagonalization Bijan Afsari bijan@math.umd.edu AMSC Department Sensitivity Analysis for the Problem of Matrix Joint Diagonalization p. 1/1 Outline
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More informationMeasuring relationships among multiple responses
Measuring relationships among multiple responses Linear association (correlation, relatedness, shared information) between pair-wise responses is an important property used in almost all multivariate analyses.
More informationRobust estimation, efficiency, and Lasso debiasing
Robust estimation, efficiency, and Lasso debiasing Po-Ling Loh University of Wisconsin - Madison Departments of ECE & Statistics WHOA-PSI workshop Washington University in St. Louis Aug 12, 2017 Po-Ling
More information1 Data Arrays and Decompositions
1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is
More informationConvex sets, conic matrix factorizations and conic rank lower bounds
Convex sets, conic matrix factorizations and conic rank lower bounds Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute
More informationPhase Retrieval from Local Measurements in Two Dimensions
Phase Retrieval from Local Measurements in Two Dimensions Mark Iwen a, Brian Preskitt b, Rayan Saab b, and Aditya Viswanathan c a Department of Mathematics, and Department of Computational Mathematics,
More informationMarginal Specifications and a Gaussian Copula Estimation
Marginal Specifications and a Gaussian Copula Estimation Kazim Azam Abstract Multivariate analysis involving random variables of different type like count, continuous or mixture of both is frequently required
More informationStable Process. 2. Multivariate Stable Distributions. July, 2006
Stable Process 2. Multivariate Stable Distributions July, 2006 1. Stable random vectors. 2. Characteristic functions. 3. Strictly stable and symmetric stable random vectors. 4. Sub-Gaussian random vectors.
More informationProximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725
Proximal Gradient Descent and Acceleration Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: subgradient method Consider the problem min f(x) with f convex, and dom(f) = R n. Subgradient method:
More informationx. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).
.8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics
More informationHigh-dimensional Covariance Estimation Based On Gaussian Graphical Models
High-dimensional Covariance Estimation Based On Gaussian Graphical Models Shuheng Zhou, Philipp Rutimann, Min Xu and Peter Buhlmann February 3, 2012 Problem definition Want to estimate the covariance matrix
More informationforms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms
Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.
More informationThe Use of Copulas to Model Conditional Expectation for Multivariate Data
The Use of Copulas to Model Conditional Expectation for Multivariate Data Kääri, Meelis; Selart, Anne; Kääri, Ene University of Tartu, Institute of Mathematical Statistics J. Liivi 2 50409 Tartu, Estonia
More informationTractable Upper Bounds on the Restricted Isometry Constant
Tractable Upper Bounds on the Restricted Isometry Constant Alex d Aspremont, Francis Bach, Laurent El Ghaoui Princeton University, École Normale Supérieure, U.C. Berkeley. Support from NSF, DHS and Google.
More information1.1 Basis of Statistical Decision Theory
ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 1: Introduction Lecturer: Yihong Wu Scribe: AmirEmad Ghassami, Jan 21, 2016 [Ed. Jan 31] Outline: Introduction of
More informationLinear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations
Linear Algebra in Computer Vision CSED441:Introduction to Computer Vision (2017F Lecture2: Basic Linear Algebra & Probability Bohyung Han CSE, POSTECH bhhan@postech.ac.kr Mathematics in vector space Linear
More informationLecture Notes 5: Multiresolution Analysis
Optimization-based data analysis Fall 2017 Lecture Notes 5: Multiresolution Analysis 1 Frames A frame is a generalization of an orthonormal basis. The inner products between the vectors in a frame and
More information8.1 Concentration inequality for Gaussian random matrix (cont d)
MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration
More informationBayesian Sparse Linear Regression with Unknown Symmetric Error
Bayesian Sparse Linear Regression with Unknown Symmetric Error Minwoo Chae 1 Joint work with Lizhen Lin 2 David B. Dunson 3 1 Department of Mathematics, The University of Texas at Austin 2 Department of
More informationJoint Gaussian Graphical Model Review Series I
Joint Gaussian Graphical Model Review Series I Probability Foundations Beilun Wang Advisor: Yanjun Qi 1 Department of Computer Science, University of Virginia http://jointggm.org/ June 23rd, 2017 Beilun
More informationRecovering any low-rank matrix, provably
Recovering any low-rank matrix, provably Rachel Ward University of Texas at Austin October, 2014 Joint work with Yudong Chen (U.C. Berkeley), Srinadh Bhojanapalli and Sujay Sanghavi (U.T. Austin) Matrix
More informationProbability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur
Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur Lecture No. # 29 Introduction to Copulas Hello and welcome to this
More informationA direct formulation for sparse PCA using semidefinite programming
A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley A. d Aspremont, INFORMS, Denver,
More informationNonconvex penalties: Signal-to-noise ratio and algorithms
Nonconvex penalties: Signal-to-noise ratio and algorithms Patrick Breheny March 21 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/22 Introduction In today s lecture, we will return to nonconvex
More informationO Combining cross-validation and plug-in methods - for kernel density bandwidth selection O
O Combining cross-validation and plug-in methods - for kernel density selection O Carlos Tenreiro CMUC and DMUC, University of Coimbra PhD Program UC UP February 18, 2011 1 Overview The nonparametric problem
More informationMulti-task Quantile Regression under the Transnormal Model
Multi-task Quantile Regression under the Transnormal Model Jianqing Fan, Lingzhou Xue and Hui Zou Princeton University, Pennsylvania State University and University of Minnesota Abstract We consider estimating
More informationA direct formulation for sparse PCA using semidefinite programming
A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon
More informationA priori bounds on the condition numbers in interior-point methods
A priori bounds on the condition numbers in interior-point methods Florian Jarre, Mathematisches Institut, Heinrich-Heine Universität Düsseldorf, Germany. Abstract Interior-point methods are known to be
More informationVariable Selection for Highly Correlated Predictors
Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu arxiv:1709.04840v1 [stat.me] 14 Sep 2017 Abstract Penalty-based variable selection methods are powerful in selecting relevant covariates
More informationOverlapping Variable Clustering with Statistical Guarantees and LOVE
with Statistical Guarantees and LOVE Department of Statistical Science Cornell University WHOA-PSI, St. Louis, August 2017 Joint work with Mike Bing, Yang Ning and Marten Wegkamp Cornell University, Department
More informationUsing Friendly Tail Bounds for Sums of Random Matrices
Using Friendly Tail Bounds for Sums of Random Matrices Joel A. Tropp Computing + Mathematical Sciences California Institute of Technology jtropp@cms.caltech.edu Research supported in part by NSF, DARPA,
More informationPermutation-invariant regularization of large covariance matrices. Liza Levina
Liza Levina Permutation-invariant covariance regularization 1/42 Permutation-invariant regularization of large covariance matrices Liza Levina Department of Statistics University of Michigan Joint work
More informationUniversal low-rank matrix recovery from Pauli measurements
Universal low-rank matrix recovery from Pauli measurements Yi-Kai Liu Applied and Computational Mathematics Division National Institute of Standards and Technology Gaithersburg, MD, USA yi-kai.liu@nist.gov
More information10-725/36-725: Convex Optimization Prerequisite Topics
10-725/36-725: Convex Optimization Prerequisite Topics February 3, 2015 This is meant to be a brief, informal refresher of some topics that will form building blocks in this course. The content of the
More informationA PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS
Statistica Sinica 20 2010, 365-378 A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS Liang Peng Georgia Institute of Technology Abstract: Estimating tail dependence functions is important for applications
More information