Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas

Size: px
Start display at page:

Download "Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas"

Transcription

1 Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas Department of Mathematics Department of Statistical Science Cornell University London, January 7, 2016

2 Joint work with Yue Zhao Supported by NSF Grant DMS

3 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3

4 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3

5 Copula Let the continuous random vector X = (X 1,..., X d ) R d have marginal distribution functions F j (x) = P { X j x }. The copula function C(u 1,..., u d ) is the joint cumulative distribution function of the transformation U = (F 1 (X 1 ),..., F d (X d ) ), i.e., for u = (u 1,..., u d ) [0, 1] d, C(u 1,..., u d ) = P {F 1 (X 1 ) u 1,..., F d (X d ) u d }.

6 Copula The copula C always describes a distribution on the unit square. Copula is invariant under strictly increasing transformations T j of the marginals X j allows separate specifications of the marginals and the dependence structure. In this talk we mostly study semiparametric copula models. The dependence structure is given by a parametric copula, while the marginals are left unspecified (nonparametric).

7 Elliptical distribution A random vector X R d has an elliptical distribution if its characteristic function E(exp(it X)) can be written as e it µ Ψ(t Σt) with parameters µ R d, covariance matrix Σ R d d, and characteristic generator Ψ : [0, ) R. For instance, when Ψ(t) = e t/2, X N(µ, Σ).

8 Semiparametric elliptical distributions and the copula correlation matrix Elliptical copulas are the copulas of elliptical distributions. Elliptical copula is characterized by copula parameters: Ψ and the copula correlation matrix Σ, with Σ ij = Σ ij Σ ii Σ jj.

9 Semiparametric Gaussian graphical copula model Let X N(0, Σ), and Ω = Σ 1 be the precision matrix. If [Ω] kl = 0, then X k X l X \{k,l}. Accurate estimator of Ω can be obtained by feeding the sample covariance matrix Σ into the graphical lasso, CLIME, graphical Dantzig selector, etc. Assume instead that we merely have f (X) N(0, Σ), for some strictly increasing, unknown function f (i.e., X has a Gaussian copula with copula correlation matrix Σ), and again we let Ω = Σ 1. Now, if [Ω] kl = 0, then f (X k ) f (X l ) f (X) \{k,l} = X k X l X \{k,l}.

10 Semiparametric Gaussian graphical copula model Using a rank-based estimator Σ rank-based (using Kendall s tau or Spearman s rho) of Σ, Liu et al 2012, Xue & Zou 2012 recovers the same optimal parametric rate (obtained under the Gaussian model) under the semiparametric model. This crucially depends on estimating Σ rank-based Σ max at the rate of log(d)/n (w.h.p.), which is in turn based on Hoeffding s classical bound for scalar U-statistic.

11 Kendall s tau Let X R d be an independent copy of X. Kendall s tau between the kth and lth coordinates is [ ( τ kl = E sgn (X k X k )(X l X )] l ) Let T be the matrix of Kendall s tau: [T ] kl = τ kl.

12 Kendall s tau Kendall s tau statistic is, for samples X 1,..., X n that are independent copies of X, τ kl = 2 n(n 1) 1 i<j d ) sgn ((X k i X j k )(X l i X j l ). Let T be the matrix of Kendall s tau statistics: [ T ] kl = τ kl.

13 Kendall s tau matrix as plug-in estimator For elliptical copulas, independent of the characteristic generator, ( π ) Σ = sin 2 T with the sine function acting component-wise (Kruskal 1958, Kendall & Gibbons 1990, Fang, Fang & Kotz 2002). Hence, the natural plug-in estimator for Σ is ( π Σ = sin 2 T ).

14 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3

15 For the rest of the talk, we assume X R d follows a semiparametric elliptical distribution, The copula correlation matrix of X is Σ R d d, The plug-in estimator Σ = Σ n of Σ is constructed from n independent copies of X, (e.g., Demarta & McNeil 2004, Klüppelberg & Kuhn 2009)

16 Some recent advanced in elliptical/gaussian copulas Klüppelberg & Kuhn 2009 study the property of Σ in the asymptotic regime, i.e., d fixed, n. Liu et al 2012, Xue & Zou 2012 study estimation of the precision matrix Σ 1 of Gaussian copulas in high dimensions. Sharp bound on Σ Σ max is a key step of the study.

17 Proposed Research: Part I First, we establish a sharp bound of the operator norm Σ Σ 2. (See also Han & Liu 2013.) It has often been observed (e.g., Demarta & McNeil 2004, Klüppelberg & Kuhn 2009) that the plug-in estimator Σ is not always positive semidefinite. A bound on Σ Σ 2 quantifies the extent to which the non-positive semidefinite problem may happen. If Σ is not positive semidefinite, we can project it onto the set of correlation matrices in some specific way to obtain Σ +. Furthermore, Σ + Σ 2 2 Σ Σ 2.

18 Proposed Research: Part II Later, we study a factor model for Σ. The factor model assumes that Σ = Θ + V for a low-rank (or nearly low-rank) matrix Θ and a diagonal matrix V. X and [L, V 1/2 ]Y have the same copula for some elliptical distribution with Σ = I r+d. Here Θ = LL for some L R d r. Within this context, the effective number of parameters is O(r d), for r = rank(θ), instead of O(d 2 ) under the generic model. We propose a refined estimator Σ for Σ using a penalized least square procedure. The bound on Σ Σ 2 serves to scale the penalty term.

19 Bounding T T 2 Bounding Σ Σ 2 Outline for bounding Σ Σ 2 Bounding T T 2 (matrix counterpart of Hoeffding 1963 classical bound for scalar U-statistic). Bounding Σ Σ 2 in terms of T T 2 (matrix counterpart of the Lipschitz property).

20 Bounding T T 2 Bounding Σ Σ 2 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3

21 Bounding T T 2 Bounding Σ Σ 2 Matrix U-statistic Theorem With probability at least 1 α, we have with { T T 2 < max T 2 f n,d,α, fn,d,α 2 T 2 fn,d,α 2 + f n,d,α 2 /4 + f n,d,α 2 { } /2 < max T 2 f n,d,α, fn,d,α 2 + fn,d,α 2 16 f n,d,α = 3 d log(2α 1 d). n The above bounds hold for all n, d, α. }

22 Bounding T T 2 Bounding Σ Σ 2 Matrix U-statistic: proof We need a Bernstein-type inequality, { e.g., } Tropp (2012), to bound the tail probability P T T 2 t. That theorem, as many other results on bounding the tail probability of the operator norm of a sum of random matrices, requires that the summands be independent. The U-statistics T does not satisfy this condition. Many results on tail bound rely on the Laplace transform technique to convert the tail probability into an expectation of some convex function of T T. A technique pointed out by Hoeffding (1963) allows us to convert the problem of bounding T T 2 into a problem involving a sum of independent random matrices.

23 Bounding T T 2 Bounding Σ Σ 2 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3

24 Bounding T T 2 Bounding Σ Σ 2 Matrix Lipschitz property Theorem We have, with probability at least 1 α, that Σ Σ 2 π T T 2 + 3π2 16 f 2 n,d,α. Combining earlier bound for T T 2, we have, with probability at least 1 2α, that Σ Σ 2 < π max { T 2 f n,d,α, f 2 n,d,α } + 3π2 16 f 2 n,d,α. In addition, 2 π Σ 2 T 2 Σ 2.

25 Bounding T T 2 Bounding Σ Σ 2 Potential improvement We can obtain from the corollary that, for n d, Σ 2 d log(d) E Σ Σ 2. n In contrast, for Gaussian X and Σ its sample covariance matrix, Koltchinskii & Lounici 2015 shows that, for n d, Σ 2 d E Σ Σ 2. n

26 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3

27 A factor model for Σ From now on, we assume that the copula correlation matrix Σ satisfies a factor model: Σ = Θ + V for some low-rank (or nearly low-rank), positive semidefinite, symmetric matrix Θ and diagonal matrix V.

28 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3

29 Prelude: an elementary factor model for Σ In an elementary factor model for Σ = Θ + V, we assume that rank(θ) = r < d, and V = σ 2 I d. Let the plug-in estimator Σ have the eigen-decomposition d λ k û k û k, λ1 λ d. k=1

30 Prelude: an elementary factor model for Σ Construct the following closed-form estimators for r, σ 2, Θ and Σ, based on some threshold µ: r = σ 2 = Θ = d k=1 1 d r r k=1 1 { λk λ } d + µ, λ j, j> r ( λ k σ 2 )û k û k Σ = Θ diag( Θ) + I.

31 Prelude: an elementary factor model for Σ, continued Theorem Assume λ r (Θ) 2µ. On the event { } 2 Σ Σ 2 < µ, we have r = r, Σ Σ 2 F Θ Θ 2 F 8r Σ Σ 2 2, σ 2 σ 2 Σ Σ 2. Here, F denotes the Frobenius norm.

32 Prelude: an elementary factor model for Σ, continued Corollary Set µ = 8 T 2 fn,d,α 2 + f n,d,α 4 /4 + 8f n,d,α 2 µ = 8 T 2 f n,d,α + 12fn,d,α 2. Assume λ r (Θ) 2 µ, and T 2 fn,d,α 2. Then, Σ Σ 2 F 2r µ2, hold with probability exceeding 1 2α.

33 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3

34 The refined estimator Σ of Σ = Θ + V In the factor model we can write Σ = Θ diag(θ) + I d := Θ o + I d. This motivates us to estimate the low-rank matrix Θ by Θ as the solution to the convex program { } 1 Θ = argmin Θ R d d 2 Θ o Σ o 2 F + µ Θ. Here, denotes the nuclear norm. Finally, we let the refined estimator Σ of the copula correlation matrix Σ be Σ = Θ o + I d.

35 Comparison of Σ with existing estimators Saunderson et al 2012 study a factor model for Σ in the noiseless setting using minimum trace factor analysis: Σ = Θ + V (Θ, V ) = argmin tr(θ ) subject to Θ 0 V diagonal If Θ has column space U such that its coherence max P Ue i 2 < 1/ 2, 1 i d then the above algorithm recovers Θ, V exactly.

36 Comparison of Σ with existing estimators, continued For Θ with effective low-rank and without sparse corruption, Lounici 2013 proposes { } 1 Θ = argmin Θ R d d 2 Θ Θ 2 F + µ Θ for an unbiased initial estimator Θ of Θ (which we lack). For Σ = Θ + S with low-rank matrix Θ and generic sparse matrix S, Chandrasekaran et al 2012, Hsu et al 2011 propose ( Θ, S) = argmin Θ,S R d d This formulation requires λ > 0. { } 1 2 Θ + S Σ 2 F + µ Θ + λ S l1.

37 Comparison of Σ with existing estimators, continued We are in between the above scenarios. We will extend the result of the former and relax the condition of the latter with more precise analysis.

38 Oracle inequality for the refined estimator Σ Theorem Let Θ r be best rank-r approximation to Θ, with reduced SVD Θ r = U r D r U r. Set µ = 10 T 2 f 2 n,d,α + f 4 n,d,α /4 + 10f 2 n,d,α, µ = 10 max [ T 2 f n,d,α, f 2 n,d,α Then, with probability at least 1 2α, ] + 15f 2 n,d,α. Σ Σ 2 F j>r λ 2 j (Θ) + 18r µ2 for all r with γ r = U r U r max 1/9.

39 Outline 1 Introduction 2 Bounding T T 2 Bounding Σ Σ 2 3

40 Simulation result The object of interest is the ratio Σ Σ 2 F, which our Σ Σ 2 F theory predicts to be roughly proportional to d/r as n. We (somewhat arbitrarily) set C = 0.1, α = 0.1, and forgo the log(d) factor when choosing the regularization parameter. We estimate this ratio by simulation.

41 Simulation setup Σ 2 = (Larger is better) 10 2 k b '! 'k 2 F =je '! 'k 2 F d=128, r=1 d=32, r=1 d=128, r=4 d=8, r=1 d=32, r= d=8, r= Sample size

42 Summary We have obtained a sharp bound on the operator norm of Σ Σ, for the plug-in estimator Σ of the copula correlation matrix Σ. We have applied the above bound to the setting of a factor model of Σ to obtain a refined estimator Σ; a sharp oracle inequality for Σ is established.

43 Thanks!

The Nonparanormal skeptic

The Nonparanormal skeptic The Nonpara skeptic Han Liu Johns Hopkins University, 615 N. Wolfe Street, Baltimore, MD 21205 USA Fang Han Johns Hopkins University, 615 N. Wolfe Street, Baltimore, MD 21205 USA Ming Yuan Georgia Institute

More information

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline.

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline. MFM Practitioner Module: Risk & Asset Allocation September 11, 2013 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y

More information

Robust Sparse Principal Component Regression under the High Dimensional Elliptical Model

Robust Sparse Principal Component Regression under the High Dimensional Elliptical Model Robust Sparse Principal Component Regression under the High Dimensional Elliptical Model Fang Han Department of Biostatistics Johns Hopkins University Baltimore, MD 21210 fhan@jhsph.edu Han Liu Department

More information

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline. Practitioner Course: Portfolio Optimization September 10, 2008 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y ) (x,

More information

Learning discrete graphical models via generalized inverse covariance matrices

Learning discrete graphical models via generalized inverse covariance matrices Learning discrete graphical models via generalized inverse covariance matrices Duzhe Wang, Yiming Lv, Yongjoon Kim, Young Lee Department of Statistics University of Wisconsin-Madison {dwang282, lv23, ykim676,

More information

High Dimensional Semiparametric Gaussian Copula Graphical Models

High Dimensional Semiparametric Gaussian Copula Graphical Models High Dimeional Semiparametric Gaussian Copula Graphical Models Han Liu, Fang Han, Ming Yuan, John Lafferty and Larry Wasserman February 9, 2012 arxiv:1202.2169v1 [stat.ml] 10 Feb 2012 Contents Abstract:

More information

High-dimensional covariance estimation based on Gaussian graphical models

High-dimensional covariance estimation based on Gaussian graphical models High-dimensional covariance estimation based on Gaussian graphical models Shuheng Zhou Department of Statistics, The University of Michigan, Ann Arbor IMA workshop on High Dimensional Phenomena Sept. 26,

More information

Financial Econometrics and Volatility Models Copulas

Financial Econometrics and Volatility Models Copulas Financial Econometrics and Volatility Models Copulas Eric Zivot Updated: May 10, 2010 Reading MFTS, chapter 19 FMUND, chapters 6 and 7 Introduction Capturing co-movement between financial asset returns

More information

An iterative hard thresholding estimator for low rank matrix recovery

An iterative hard thresholding estimator for low rank matrix recovery An iterative hard thresholding estimator for low rank matrix recovery Alexandra Carpentier - based on a joint work with Arlene K.Y. Kim Statistical Laboratory, Department of Pure Mathematics and Mathematical

More information

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming High Dimensional Inverse Covariate Matrix Estimation via Linear Programming Ming Yuan October 24, 2011 Gaussian Graphical Model X = (X 1,..., X p ) indep. N(µ, Σ) Inverse covariance matrix Σ 1 = Ω = (ω

More information

Robust and sparse Gaussian graphical modelling under cell-wise contamination

Robust and sparse Gaussian graphical modelling under cell-wise contamination Robust and sparse Gaussian graphical modelling under cell-wise contamination Shota Katayama 1, Hironori Fujisawa 2 and Mathias Drton 3 1 Tokyo Institute of Technology, Japan 2 The Institute of Statistical

More information

Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation

Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation UIUC CSL Mar. 24 Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation Yuejie Chi Department of ECE and BMI Ohio State University Joint work with Yuxin Chen (Stanford).

More information

Solving Corrupted Quadratic Equations, Provably

Solving Corrupted Quadratic Equations, Provably Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin

More information

High Dimensional Covariance and Precision Matrix Estimation

High Dimensional Covariance and Precision Matrix Estimation High Dimensional Covariance and Precision Matrix Estimation Wei Wang Washington University in St. Louis Thursday 23 rd February, 2017 Wei Wang (Washington University in St. Louis) High Dimensional Covariance

More information

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun

More information

Causal Inference: Discussion

Causal Inference: Discussion Causal Inference: Discussion Mladen Kolar The University of Chicago Booth School of Business Sept 23, 2016 Types of machine learning problems Based on the information available: Supervised learning Reinforcement

More information

Efficient estimation of a semiparametric dynamic copula model

Efficient estimation of a semiparametric dynamic copula model Efficient estimation of a semiparametric dynamic copula model Christian Hafner Olga Reznikova Institute of Statistics Université catholique de Louvain Louvain-la-Neuve, Blgium 30 January 2009 Young Researchers

More information

Variations on Nonparametric Additive Models: Computational and Statistical Aspects

Variations on Nonparametric Additive Models: Computational and Statistical Aspects Variations on Nonparametric Additive Models: Computational and Statistical Aspects John Lafferty Department of Statistics & Department of Computer Science University of Chicago Collaborators Sivaraman

More information

Robust Principal Component Analysis

Robust Principal Component Analysis ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M

More information

An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss

An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss arxiv:1811.04545v1 [stat.co] 12 Nov 2018 Cheng Wang School of Mathematical Sciences, Shanghai Jiao

More information

Sparse Covariance Matrix Estimation with Eigenvalue Constraints

Sparse Covariance Matrix Estimation with Eigenvalue Constraints Sparse Covariance Matrix Estimation with Eigenvalue Constraints Han Liu and Lie Wang 2 and Tuo Zhao 3 Department of Operations Research and Financial Engineering, Princeton University 2 Department of Mathematics,

More information

Matrix concentration inequalities

Matrix concentration inequalities ELE 538B: Mathematics of High-Dimensional Data Matrix concentration inequalities Yuxin Chen Princeton University, Fall 2018 Recap: matrix Bernstein inequality Consider a sequence of independent random

More information

Recovery of Simultaneously Structured Models using Convex Optimization

Recovery of Simultaneously Structured Models using Convex Optimization Recovery of Simultaneously Structured Models using Convex Optimization Maryam Fazel University of Washington Joint work with: Amin Jalali (UW), Samet Oymak and Babak Hassibi (Caltech) Yonina Eldar (Technion)

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

Construction and estimation of high dimensional copulas

Construction and estimation of high dimensional copulas Construction and estimation of high dimensional copulas Gildas Mazo PhD work supervised by S. Girard and F. Forbes Mistis, Inria and laboratoire Jean Kuntzmann, Grenoble, France Séminaire Statistiques,

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57

More information

Intrinsic Volumes of Convex Cones Theory and Applications

Intrinsic Volumes of Convex Cones Theory and Applications Intrinsic Volumes of Convex Cones Theory and Applications M\cr NA Manchester Numerical Analysis Martin Lotz School of Mathematics The University of Manchester with the collaboration of Dennis Amelunxen,

More information

Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation

Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation Adam J. Rothman School of Statistics University of Minnesota October 8, 2014, joint work with Liliana

More information

Reconstruction from Anisotropic Random Measurements

Reconstruction from Anisotropic Random Measurements Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013

More information

STATISTICAL MACHINE LEARNING FOR STRUCTURED AND HIGH DIMENSIONAL DATA

STATISTICAL MACHINE LEARNING FOR STRUCTURED AND HIGH DIMENSIONAL DATA AFRL-OSR-VA-TR-2014-0234 STATISTICAL MACHINE LEARNING FOR STRUCTURED AND HIGH DIMENSIONAL DATA Larry Wasserman CARNEGIE MELLON UNIVERSITY 0 Final Report DISTRIBUTION A: Distribution approved for public

More information

Least squares under convex constraint

Least squares under convex constraint Stanford University Questions Let Z be an n-dimensional standard Gaussian random vector. Let µ be a point in R n and let Y = Z + µ. We are interested in estimating µ from the data vector Y, under the assumption

More information

Fast and Robust Phase Retrieval

Fast and Robust Phase Retrieval Fast and Robust Phase Retrieval Aditya Viswanathan aditya@math.msu.edu CCAM Lunch Seminar Purdue University April 18 2014 0 / 27 Joint work with Yang Wang Mark Iwen Research supported in part by National

More information

The deterministic Lasso

The deterministic Lasso The deterministic Lasso Sara van de Geer Seminar für Statistik, ETH Zürich Abstract We study high-dimensional generalized linear models and empirical risk minimization using the Lasso An oracle inequality

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

sparse and low-rank tensor recovery Cubic-Sketching

sparse and low-rank tensor recovery Cubic-Sketching Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru

More information

Modelling Dependent Credit Risks

Modelling Dependent Credit Risks Modelling Dependent Credit Risks Filip Lindskog, RiskLab, ETH Zürich 30 November 2000 Home page:http://www.math.ethz.ch/ lindskog E-mail:lindskog@math.ethz.ch RiskLab:http://www.risklab.ch Modelling Dependent

More information

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology

More information

Stein s Method for Matrix Concentration

Stein s Method for Matrix Concentration Stein s Method for Matrix Concentration Lester Mackey Collaborators: Michael I. Jordan, Richard Y. Chen, Brendan Farrell, and Joel A. Tropp University of California, Berkeley California Institute of Technology

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE7C (Spring 08): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee7c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee7c@berkeley.edu October

More information

(Part 1) High-dimensional statistics May / 41

(Part 1) High-dimensional statistics May / 41 Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2

More information

ELE 538B: Mathematics of High-Dimensional Data. Spectral methods. Yuxin Chen Princeton University, Fall 2018

ELE 538B: Mathematics of High-Dimensional Data. Spectral methods. Yuxin Chen Princeton University, Fall 2018 ELE 538B: Mathematics of High-Dimensional Data Spectral methods Yuxin Chen Princeton University, Fall 2018 Outline A motivating application: graph clustering Distance and angles between two subspaces Eigen-space

More information

arxiv: v1 [math.st] 13 Feb 2012

arxiv: v1 [math.st] 13 Feb 2012 Sparse Matrix Inversion with Scaled Lasso Tingni Sun and Cun-Hui Zhang Rutgers University arxiv:1202.2723v1 [math.st] 13 Feb 2012 Address: Department of Statistics and Biostatistics, Hill Center, Busch

More information

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1 Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1 Feng Wei 2 University of Michigan July 29, 2016 1 This presentation is based a project under the supervision of M. Rudelson.

More information

A measure of radial asymmetry for bivariate copulas based on Sobolev norm

A measure of radial asymmetry for bivariate copulas based on Sobolev norm A measure of radial asymmetry for bivariate copulas based on Sobolev norm Ahmad Alikhani-Vafa Ali Dolati Abstract The modified Sobolev norm is used to construct an index for measuring the degree of radial

More information

Fundamentals of Matrices

Fundamentals of Matrices Maschinelles Lernen II Fundamentals of Matrices Christoph Sawade/Niels Landwehr/Blaine Nelson Tobias Scheffer Matrix Examples Recap: Data Linear Model: f i x = w i T x Let X = x x n be the data matrix

More information

Can we do statistical inference in a non-asymptotic way? 1

Can we do statistical inference in a non-asymptotic way? 1 Can we do statistical inference in a non-asymptotic way? 1 Guang Cheng 2 Statistics@Purdue www.science.purdue.edu/bigdata/ ONR Review Meeting@Duke Oct 11, 2017 1 Acknowledge NSF, ONR and Simons Foundation.

More information

Probabilistic Engineering Mechanics. An innovating analysis of the Nataf transformation from the copula viewpoint

Probabilistic Engineering Mechanics. An innovating analysis of the Nataf transformation from the copula viewpoint Probabilistic Engineering Mechanics 4 9 3 3 Contents lists available at ScienceDirect Probabilistic Engineering Mechanics journal homepage: www.elsevier.com/locate/probengmech An innovating analysis of

More information

ECA: High Dimensional Elliptical Component Analysis in non-gaussian Distributions

ECA: High Dimensional Elliptical Component Analysis in non-gaussian Distributions : High Dimensional Elliptical Component Analysis in non-gaussian Distributions Fang Han and Han Liu arxiv:1310.3561v4 [stat.ml] 3 Oct 016 Abstract We present a robust alternative to principal component

More information

Multivariate Normal-Laplace Distribution and Processes

Multivariate Normal-Laplace Distribution and Processes CHAPTER 4 Multivariate Normal-Laplace Distribution and Processes The normal-laplace distribution, which results from the convolution of independent normal and Laplace random variables is introduced by

More information

Adaptive one-bit matrix completion

Adaptive one-bit matrix completion Adaptive one-bit matrix completion Joseph Salmon Télécom Paristech, Institut Mines-Télécom Joint work with Jean Lafond (Télécom Paristech) Olga Klopp (Crest / MODAL X, Université Paris Ouest) Éric Moulines

More information

Journal of Multivariate Analysis. Consistency of sparse PCA in High Dimension, Low Sample Size contexts

Journal of Multivariate Analysis. Consistency of sparse PCA in High Dimension, Low Sample Size contexts Journal of Multivariate Analysis 5 (03) 37 333 Contents lists available at SciVerse ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Consistency of sparse PCA

More information

Statistical Machine Learning for Structured and High Dimensional Data

Statistical Machine Learning for Structured and High Dimensional Data Statistical Machine Learning for Structured and High Dimensional Data (FA9550-09- 1-0373) PI: Larry Wasserman (CMU) Co- PI: John Lafferty (UChicago and CMU) AFOSR Program Review (Jan 28-31, 2013, Washington,

More information

Numerical schemes of resolution of stochastic optimal control HJB equation

Numerical schemes of resolution of stochastic optimal control HJB equation Numerical schemes of resolution of stochastic optimal control HJB equation Elisabeth Ottenwaelter Journée des doctorants 7 mars 2007 Équipe COMMANDS Elisabeth Ottenwaelter (INRIA et CMAP) Numerical schemes

More information

arxiv: v2 [math.st] 12 Feb 2008

arxiv: v2 [math.st] 12 Feb 2008 arxiv:080.460v2 [math.st] 2 Feb 2008 Electronic Journal of Statistics Vol. 2 2008 90 02 ISSN: 935-7524 DOI: 0.24/08-EJS77 Sup-norm convergence rate and sign concentration property of Lasso and Dantzig

More information

the long tau-path for detecting monotone association in an unspecified subpopulation

the long tau-path for detecting monotone association in an unspecified subpopulation the long tau-path for detecting monotone association in an unspecified subpopulation Joe Verducci Current Challenges in Statistical Learning Workshop Banff International Research Station Tuesday, December

More information

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression Noah Simon Jerome Friedman Trevor Hastie November 5, 013 Abstract In this paper we purpose a blockwise descent

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

Regularization Methods for Prediction in Dynamic Graphs and e-marketing Applications

Regularization Methods for Prediction in Dynamic Graphs and e-marketing Applications Regularization Methods for Prediction in Dynamic Graphs and e-marketing Applications Emile Richard CMLA-ENS Cachan 1000mercis PhD defense Advisors: Th. Evgeniou (INSEAD), N. Vayatis (CMLA-ENS Cachan) November

More information

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider the problem of

More information

Computational and Statistical Aspects of Statistical Machine Learning. John Lafferty Department of Statistics Retreat Gleacher Center

Computational and Statistical Aspects of Statistical Machine Learning. John Lafferty Department of Statistics Retreat Gleacher Center Computational and Statistical Aspects of Statistical Machine Learning John Lafferty Department of Statistics Retreat Gleacher Center Outline Modern nonparametric inference for high dimensional data Nonparametric

More information

Lecture 5 : Projections

Lecture 5 : Projections Lecture 5 : Projections EE227C. Lecturer: Professor Martin Wainwright. Scribe: Alvin Wan Up until now, we have seen convergence rates of unconstrained gradient descent. Now, we consider a constrained minimization

More information

Modelling and Estimation of Stochastic Dependence

Modelling and Estimation of Stochastic Dependence Modelling and Estimation of Stochastic Dependence Uwe Schmock Based on joint work with Dr. Barbara Dengler Financial and Actuarial Mathematics and Christian Doppler Laboratory for Portfolio Risk Management

More information

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Sensitivity Analysis for the Problem of Matrix Joint Diagonalization

Sensitivity Analysis for the Problem of Matrix Joint Diagonalization Sensitivity Analysis for the Problem of Matrix Joint Diagonalization Bijan Afsari bijan@math.umd.edu AMSC Department Sensitivity Analysis for the Problem of Matrix Joint Diagonalization p. 1/1 Outline

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

Measuring relationships among multiple responses

Measuring relationships among multiple responses Measuring relationships among multiple responses Linear association (correlation, relatedness, shared information) between pair-wise responses is an important property used in almost all multivariate analyses.

More information

Robust estimation, efficiency, and Lasso debiasing

Robust estimation, efficiency, and Lasso debiasing Robust estimation, efficiency, and Lasso debiasing Po-Ling Loh University of Wisconsin - Madison Departments of ECE & Statistics WHOA-PSI workshop Washington University in St. Louis Aug 12, 2017 Po-Ling

More information

1 Data Arrays and Decompositions

1 Data Arrays and Decompositions 1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is

More information

Convex sets, conic matrix factorizations and conic rank lower bounds

Convex sets, conic matrix factorizations and conic rank lower bounds Convex sets, conic matrix factorizations and conic rank lower bounds Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute

More information

Phase Retrieval from Local Measurements in Two Dimensions

Phase Retrieval from Local Measurements in Two Dimensions Phase Retrieval from Local Measurements in Two Dimensions Mark Iwen a, Brian Preskitt b, Rayan Saab b, and Aditya Viswanathan c a Department of Mathematics, and Department of Computational Mathematics,

More information

Marginal Specifications and a Gaussian Copula Estimation

Marginal Specifications and a Gaussian Copula Estimation Marginal Specifications and a Gaussian Copula Estimation Kazim Azam Abstract Multivariate analysis involving random variables of different type like count, continuous or mixture of both is frequently required

More information

Stable Process. 2. Multivariate Stable Distributions. July, 2006

Stable Process. 2. Multivariate Stable Distributions. July, 2006 Stable Process 2. Multivariate Stable Distributions July, 2006 1. Stable random vectors. 2. Characteristic functions. 3. Strictly stable and symmetric stable random vectors. 4. Sub-Gaussian random vectors.

More information

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725 Proximal Gradient Descent and Acceleration Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: subgradient method Consider the problem min f(x) with f convex, and dom(f) = R n. Subgradient method:

More information

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ). .8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics

More information

High-dimensional Covariance Estimation Based On Gaussian Graphical Models

High-dimensional Covariance Estimation Based On Gaussian Graphical Models High-dimensional Covariance Estimation Based On Gaussian Graphical Models Shuheng Zhou, Philipp Rutimann, Min Xu and Peter Buhlmann February 3, 2012 Problem definition Want to estimate the covariance matrix

More information

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.

More information

The Use of Copulas to Model Conditional Expectation for Multivariate Data

The Use of Copulas to Model Conditional Expectation for Multivariate Data The Use of Copulas to Model Conditional Expectation for Multivariate Data Kääri, Meelis; Selart, Anne; Kääri, Ene University of Tartu, Institute of Mathematical Statistics J. Liivi 2 50409 Tartu, Estonia

More information

Tractable Upper Bounds on the Restricted Isometry Constant

Tractable Upper Bounds on the Restricted Isometry Constant Tractable Upper Bounds on the Restricted Isometry Constant Alex d Aspremont, Francis Bach, Laurent El Ghaoui Princeton University, École Normale Supérieure, U.C. Berkeley. Support from NSF, DHS and Google.

More information

1.1 Basis of Statistical Decision Theory

1.1 Basis of Statistical Decision Theory ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 1: Introduction Lecturer: Yihong Wu Scribe: AmirEmad Ghassami, Jan 21, 2016 [Ed. Jan 31] Outline: Introduction of

More information

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations Linear Algebra in Computer Vision CSED441:Introduction to Computer Vision (2017F Lecture2: Basic Linear Algebra & Probability Bohyung Han CSE, POSTECH bhhan@postech.ac.kr Mathematics in vector space Linear

More information

Lecture Notes 5: Multiresolution Analysis

Lecture Notes 5: Multiresolution Analysis Optimization-based data analysis Fall 2017 Lecture Notes 5: Multiresolution Analysis 1 Frames A frame is a generalization of an orthonormal basis. The inner products between the vectors in a frame and

More information

8.1 Concentration inequality for Gaussian random matrix (cont d)

8.1 Concentration inequality for Gaussian random matrix (cont d) MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration

More information

Bayesian Sparse Linear Regression with Unknown Symmetric Error

Bayesian Sparse Linear Regression with Unknown Symmetric Error Bayesian Sparse Linear Regression with Unknown Symmetric Error Minwoo Chae 1 Joint work with Lizhen Lin 2 David B. Dunson 3 1 Department of Mathematics, The University of Texas at Austin 2 Department of

More information

Joint Gaussian Graphical Model Review Series I

Joint Gaussian Graphical Model Review Series I Joint Gaussian Graphical Model Review Series I Probability Foundations Beilun Wang Advisor: Yanjun Qi 1 Department of Computer Science, University of Virginia http://jointggm.org/ June 23rd, 2017 Beilun

More information

Recovering any low-rank matrix, provably

Recovering any low-rank matrix, provably Recovering any low-rank matrix, provably Rachel Ward University of Texas at Austin October, 2014 Joint work with Yudong Chen (U.C. Berkeley), Srinadh Bhojanapalli and Sujay Sanghavi (U.T. Austin) Matrix

More information

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur Lecture No. # 29 Introduction to Copulas Hello and welcome to this

More information

A direct formulation for sparse PCA using semidefinite programming

A direct formulation for sparse PCA using semidefinite programming A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley A. d Aspremont, INFORMS, Denver,

More information

Nonconvex penalties: Signal-to-noise ratio and algorithms

Nonconvex penalties: Signal-to-noise ratio and algorithms Nonconvex penalties: Signal-to-noise ratio and algorithms Patrick Breheny March 21 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/22 Introduction In today s lecture, we will return to nonconvex

More information

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O O Combining cross-validation and plug-in methods - for kernel density selection O Carlos Tenreiro CMUC and DMUC, University of Coimbra PhD Program UC UP February 18, 2011 1 Overview The nonparametric problem

More information

Multi-task Quantile Regression under the Transnormal Model

Multi-task Quantile Regression under the Transnormal Model Multi-task Quantile Regression under the Transnormal Model Jianqing Fan, Lingzhou Xue and Hui Zou Princeton University, Pennsylvania State University and University of Minnesota Abstract We consider estimating

More information

A direct formulation for sparse PCA using semidefinite programming

A direct formulation for sparse PCA using semidefinite programming A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon

More information

A priori bounds on the condition numbers in interior-point methods

A priori bounds on the condition numbers in interior-point methods A priori bounds on the condition numbers in interior-point methods Florian Jarre, Mathematisches Institut, Heinrich-Heine Universität Düsseldorf, Germany. Abstract Interior-point methods are known to be

More information

Variable Selection for Highly Correlated Predictors

Variable Selection for Highly Correlated Predictors Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu arxiv:1709.04840v1 [stat.me] 14 Sep 2017 Abstract Penalty-based variable selection methods are powerful in selecting relevant covariates

More information

Overlapping Variable Clustering with Statistical Guarantees and LOVE

Overlapping Variable Clustering with Statistical Guarantees and LOVE with Statistical Guarantees and LOVE Department of Statistical Science Cornell University WHOA-PSI, St. Louis, August 2017 Joint work with Mike Bing, Yang Ning and Marten Wegkamp Cornell University, Department

More information

Using Friendly Tail Bounds for Sums of Random Matrices

Using Friendly Tail Bounds for Sums of Random Matrices Using Friendly Tail Bounds for Sums of Random Matrices Joel A. Tropp Computing + Mathematical Sciences California Institute of Technology jtropp@cms.caltech.edu Research supported in part by NSF, DARPA,

More information

Permutation-invariant regularization of large covariance matrices. Liza Levina

Permutation-invariant regularization of large covariance matrices. Liza Levina Liza Levina Permutation-invariant covariance regularization 1/42 Permutation-invariant regularization of large covariance matrices Liza Levina Department of Statistics University of Michigan Joint work

More information

Universal low-rank matrix recovery from Pauli measurements

Universal low-rank matrix recovery from Pauli measurements Universal low-rank matrix recovery from Pauli measurements Yi-Kai Liu Applied and Computational Mathematics Division National Institute of Standards and Technology Gaithersburg, MD, USA yi-kai.liu@nist.gov

More information

10-725/36-725: Convex Optimization Prerequisite Topics

10-725/36-725: Convex Optimization Prerequisite Topics 10-725/36-725: Convex Optimization Prerequisite Topics February 3, 2015 This is meant to be a brief, informal refresher of some topics that will form building blocks in this course. The content of the

More information

A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS

A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS Statistica Sinica 20 2010, 365-378 A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS Liang Peng Georgia Institute of Technology Abstract: Estimating tail dependence functions is important for applications

More information