Multivariate quantiles and conditional depth

Size: px
Start display at page:

Download "Multivariate quantiles and conditional depth"

Transcription

1 M. Hallin a,b, Z. Lu c, D. Paindaveine a, and M. Šiman d a Université libre de Bruxelles, Belgium b Princenton University, USA c University of Adelaide, Australia d Institute of Information Theory and Automation of the ASCR, Czech Republic Paris, février 2013

2 Partie 1: La profondeur et les quantiles multivariés

3 1 Statistical depth Halfspace depth More depths Possible applications 2 3 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles

4 Halfspace depth More depths Possible applications 1 Statistical depth Halfspace depth More depths Possible applications 2 3 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles

5 Halfspace depth More depths Possible applications Let C n = {y 1,..., y n}, y i R d. Statistical depth measures centrality of any y R d w.r.t. C n D(, C n) : R d [0, 1] y D(y, C n) The larger D(y, C n), the more central y w.r.t. C n. The (Tukey 1975) halfspace depth D H (y, C n) := 1 min u S d 1 n #{ y i C n : u (y i y) 0 }, where S d 1 := {u R d : u = 1} is the unit sphere in R d.

6 Halfspace depth More depths Possible applications The (Tukey 1975) halfspace depth D H (y, C n) := 1 min u S d 1 n #{ y i C n : u (y i y) 0 }, where S d 1 := {u R d : u = 1} is the unit sphere in R d

7 Halfspace depth More depths Possible applications The (Tukey 1975) halfspace depth D H (y, C n) := 1 min u S d 1 n #{ y i C n : u (y i y) 0 }, where S d 1 := {u R d : u = 1} is the unit sphere in R d

8 Halfspace depth More depths Possible applications The (Tukey 1975) halfspace depth D H (y, C n) := 1 min u S d 1 n #{ y i C n : u (y i y) 0 }, where S d 1 := {u R d : u = 1} is the unit sphere in R d

9 Halfspace depth More depths Possible applications The (Tukey 1975) halfspace depth D H (y, C n) := 1 min u S d 1 n #{ y i C n : u (y i y) 0 }, where S d 1 := {u R d : u = 1} is the unit sphere in R d

10 Halfspace depth More depths Possible applications The (Tukey 1975) halfspace depth D H (y, C n) := 1 min u S d 1 n #{ y i C n : u (y i y) 0 }, where S d 1 := {u R d : u = 1} is the unit sphere in R d

11 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ >

12 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

13 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

14 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

15 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

16 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

17 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

18 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

19 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

20 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

21 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

22 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

23 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

24 Halfspace depth More depths Possible applications depth regions R τ (C n) := { y R d : D(y, C n) τ }, for τ > The R τ s are nested, convex, and (for τ > 0) compact.

25 Halfspace depth More depths Possible applications From C n = {y 1,..., y n} to an arbitrary probability measure P Writing P (n) = 1 n n i=1 δy i for the empirical measure associated with C n = {y 1,..., y n}, 1 D H (y, C n) := min u S d 1 n #{ y i C n : u (y i y) 0 } = min u S d 1 P (n) [u (Y y) 0]. Extension to an arbitrary P : D H (y, P) = inf P[u (Y y) 0]. u S d 1 This population halfspace depth" was studied in Rousseeuw and Ruts (1999).

26 Halfspace depth More depths Possible applications This population halfspace depth" was studied in Rousseeuw and Ruts (1999) : D(Ay + b, P AY+b ) = D(y, P Y ) for any d d invertible matrix A, any d-vector b and any distribution P Y (affine-invariance) For any P, 1 d+1 max y R d D H(y, P) max y R d P[{y}] D H (y 0, P) = P[{y 0}] P is angularly symmetric about y 0 R τ (P) = {closed halfspaces H : P[H] > 1 τ}...

27 Halfspace depth More depths Possible applications This population halfspace depth" was studied in Rousseeuw and Ruts (1999) : D(Ay + b, P AY+b ) = D(y, P Y ) for any d d invertible matrix A, any d-vector b and any distribution P Y (affine-invariance) For any P, 1 d+1 max y R d D H(y, P) max y R d P[{y}] D H (y 0, P) = P[{y 0}] P is angularly symmetric about y 0 R τ (P) = {closed halfspaces H : P[H] > 1 τ}... Halfspace depth characterizes empirical measures; see Rousseuw and Struyf (1999)

28 Halfspace depth More depths Possible applications This population halfspace depth" was studied in Rousseeuw and Ruts (1999) : D(Ay + b, P AY+b ) = D(y, P Y ) for any d d invertible matrix A, any d-vector b and any distribution P Y (affine-invariance) For any P, 1 d+1 max y R d D H(y, P) max y R d P[{y}] D H (y 0, P) = P[{y 0}] P is angularly symmetric about y 0 R τ (P) = {closed halfspaces H : P[H] > 1 τ}... Halfspace depth characterizes empirical measures; see Rousseuw and Struyf (1999) compactly supported absolutely continuous distributions; see Koshevoy (2003)

29 Halfspace depth More depths Possible applications This population halfspace depth" was studied in Rousseeuw and Ruts (1999) : D(Ay + b, P AY+b ) = D(y, P Y ) for any d d invertible matrix A, any d-vector b and any distribution P Y (affine-invariance) For any P, 1 d+1 max y R d D H(y, P) max y R d P[{y}] D H (y 0, P) = P[{y 0}] P is angularly symmetric about y 0 R τ (P) = {closed halfspaces H : P[H] > 1 τ}... Halfspace depth characterizes empirical measures; see Rousseuw and Struyf (1999) compactly supported absolutely continuous distributions; see Koshevoy (2003) distributions with smooth depth contours; see Kong and Zuo (2010). General characterization remains an open question.

30 Halfspace depth More depths Possible applications Some depth regions R τ (P), with P = uniform on [0, 1] 2 ( depth >< density)

31 Halfspace depth More depths Possible applications Some depth regions R τ (P), with P having independent Cauchy marginals ( depth >< density)

32 Halfspace depth More depths Possible applications The (Liu 1990, AoS) simplicial depth D S (y, C n) := ( 1 n ) d+1 i 1 <i 2 <...<i d+1 I [ ] y Simpl(y i1, y i2,..., y id+1 ), where Simpl(y 1, y 2,..., y d+1 ) denotes the simplex with vertices y 1, y 2,..., y d

33 Halfspace depth More depths Possible applications The (Vardi and Zhang 2000) spatial depth D Sp (y, C n) := 1 O Sp (y, C n), with O Sp (y, C n) = 1 n n y y i y y i. i=1 The quantity O Sp (y, C n) is a measure of outlyingness of y wrt C n. The deepest point is the y 0 such that O Sp (y 0, C n) = 0. It is called the spatial median, and generalizes the univariate sample median. Milasevic and Ducharme (1987, AoS, ) prove uniqueness of the spatial median in dimension d > 1. Note that this depth is not affine-invariant.

34 Halfspace depth More depths Possible applications Spatial depth Halfspace depth Some depth regions R τ (P), with P = uniform on [0, 1] 2

35 Halfspace depth More depths Possible applications The Mahalanobis depth D M (y, C n) := d, with d := (y ȳ n) Sn 1 (y ȳ n), where ȳ n = 1 n n y i and S n = 1 n i=1 n (y i ȳ n)(y i ȳ n). (more generally, any affine-equivariant location ˆµ and scatter ˆΣ may be used) i=

36 Halfspace depth More depths Possible applications Axiomatic approach (Zuo and Serfling 2000, AoS): A depth function should satisfy the following property (P1) affine-invariance: D(Ay + b, P AY+b ) = D(y, P Y ) for any d d invertible matrix A, any d-vector b and any distribution P Y ; (P2) maximality at center: D(θ, P) = sup y R d D(y, P) holds for any P symmetric about θ (central, angular, or halfspace symmetry); (P3) monotonicity relative to the deepest point: for any P having deepest point θ and any u R d, r D(θ + ru, P) is monotone nonincreasing in r 0; (P4) vanishing at infinity: for any P, D(y, P) 0 as y.

37 Halfspace depth More depths Possible applications Axiomatic approach (Zuo and Serfling 2000, AoS): A depth function should satisfy the following property (P1) affine-invariance: D(Ay + b, P AY+b ) = D(y, P Y ) for any d d invertible matrix A, any d-vector b and any distribution P Y ; (P2) maximality at center: D(θ, P) = sup y R d D(y, P) holds for any P symmetric about θ (central, angular, or halfspace symmetry); (P3) monotonicity relative to the deepest point: for any P having deepest point θ and any u R d, r D(θ + ru, P) is monotone nonincreasing in r 0; (P4) vanishing at infinity: for any P, D(y, P) 0 as y. depth >< density

38 Halfspace depth More depths Possible applications (i) A quite informative descriptive tool

39 Halfspace depth More depths Possible applications Bai and He (1999, AoS) showed that, for P = P Y angularly symmetric about θ, { } n (arg max DH (, P (n) ) θ) L arg max inf Z (u) f u Y (0)u x x R d u S d 1 as n, where, letting H 0,u = {y R d : u y 0}, Z (u) is a centered Gaussian process on S d 1 with covariance function E[Z (u)z (v)] = P[H 0,u H 0,v ] P[H 0,u ]P[H 0,v ] = P[H 0,u H 0,v ] 1 4. Massé (2002) extended this result to the possibly asymmetric case. Still under angular symmetry, the simplicial deepest point was shown to be root-n consistent and asympotically normal in Arcones, Chen and Giné (1994, AoS). Obviously, this made use of U-process theory...

40 Halfspace depth More depths Possible applications The (Liu 1990, AoS) simplicial depth D S (y, C n) := ( 1 n ) d+1 i 1 <i 2 <...<i d+1 I [ ] y Simpl(y i1, y i2,..., y id+1 ), where Simpl(y 1, y 2,..., y d+1 ) denotes the simplex with vertices y 1, y 2,..., y d

41 Halfspace depth More depths Possible applications (ii) Symmetry testing From the results above, it follows that, for any absolutely continuous P, max D H (y, P) = 1 P is angularly symmetric about the point achieving depth 1/2. y R d 2 Hence, a universally consistent test for angular symmetry about y 0 (Rousseeuw and Struyf 2002) may reject the null for large values of T (n) = 1 2 D H(y 0, P (n) ) (quite remarkably, T (n) is distribution-free under the null). Dutta, Ghosh and Chaudhuri (2011, Bernoulli) independently provide the corresponding test about an unspecified (angular) symmetry center.

42 Halfspace depth More depths Possible applications (iii) Classification Consider (X, Y ), with X R d and Y {0, 1}. Given the realization (x, y) of which only x is observed, we have to make a "prediction" m(x) for the corresponding y

43 Halfspace depth More depths Possible applications (iii) Classification Consider (X, Y ), with X R d and Y {0, 1}. Given the realization (x, y) of which only x is observed, we have to make a "prediction" m(x) for the corresponding y. The rule that minimizes P[Y m(x)] (probability of misclassification) is (Bayes rule). [ m(x) = I η(x) > 1 ], with η(x) = P[Y = 1 X = x] 2

44 Halfspace depth More depths Possible applications (iii) Classification Consider (X, Y ), with X R d and Y {0, 1}. Given the realization (x, y) of which only x is observed, we have to make a "prediction" m(x) for the corresponding y. The rule that minimizes P[Y m(x)] (probability of misclassification) is (Bayes rule). [ m(x) = I η(x) > 1 ], with η(x) = P[Y = 1 X = x] 2 If π j = P[Y = j] and X [Y = j] N d (µ j, Σ j ) (j = 0, 1), this boils down to [ ] m(x) = I d Σ1 (x, µ 1 ) < d Σ0 (x, µ 0 ) + C, with C = C Σ0,Σ 1,π 0,π 1 (QDA).

45 Halfspace depth More depths Possible applications The red point is closer to the blue µ than to the orange µ classification into blue

46 Halfspace depth More depths Possible applications Under ellipticity assumptions (f j E(µ j, Σ j, g j )), QDA may be rephrased as [ ] m(x) = I d Σ1 (x, µ 1 ) < d Σ0 (x, µ 0 ) + C [ = I D M (x, P 1 ) > D M (x, P 0 ) + C ], since affine-equivariance implies that µ(p j ) = µ j and Σ(P j ) Σ j under ellipticity. In line with this, Ghosh and Chaudhuri (2005) propose [ ] m(x) = I D(x, P 1 ) > D(x, P 0 ), based on an arbitrary depth concept D. This is the so-called max-depth approach.

47 Halfspace depth More depths Possible applications The red point is deeper in the blue population than in the orange one classification into blue

48 Halfspace depth More depths Possible applications Li, Cuesta-Albertos, and Liu (JASA, 2012) refine the max-depth approach. They identify more appropriate separating lines in the DD plot, reporting points with coordinates ( D(x i, P (n) 1 ), D(x i, P (n) 0 )), i = 1,..., n. Pop 1 depth Pop 0 depth

49 Halfspace depth More depths Possible applications Li, Cuesta-Albertos, and Liu (JASA, 2012) refine the max-depth approach. They identify more appropriate separating lines in the DD plot, reporting points with coordinates ( D(x i, P (n) 1 ), D(x i, P (n) 0 )), i = 1,..., n. Pop 1 depth Pop 0 depth

50 Halfspace depth More depths Possible applications Li, Cuesta-Albertos, and Liu (JASA, 2012) refine the max-depth approach. They identify more appropriate separating lines in the DD plot, reporting points with coordinates ( D(x i, P (n) 1 ), D(x i, P (n) 0 )), i = 1,..., n. Pop 1 depth Pop 0 depth

51 Halfspace depth More depths Possible applications Further recent approaches: Dutta and Ghosh (2012a) consider max-depth classification based on projection depth. Achieving consistency at the elliptical model requires a modification of the procedure that requires estimating densities. Hubert and Van der Veeken (2012) adopt a parallel approach, but based on a modified version of projection depth that can deal with skewed data (standard outlyingness is replaced with "adjusted outlyingess"). Dutta and Ghosh (2012b) assume that the parent distributions are (affine transformations) of L p-spherically symmetric distributions. They propose a max-depth classifier based on an affine-invariant L p depth with an adaptive choice of p( 1) implementation again requires estimating densities. Dutta, Chaudhuri, and Ghosh (2012) define a depth-based classifier that aggregates several posterior probabilities, each of which is based on spatial depth with a fixed level of localization. Paindaveine and Van Bever (2012) define knn depth-based classifiers.

52 Halfspace depth More depths Possible applications Unless a desperately parametric Mahalanobis-type depth is used, the procedure fails if x lies outside the supports of P 0 and P 1 (since 0 = D(x, P 0 ) = D(x, P 1 ) = 0)

53 1 Statistical depth Halfspace depth More depths Possible applications 2 3 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles

54 Consider the univariate case d = 1. If P is an absolutely continuous distribution with cdf F, then D H (y, P) = min(f(y), 1 F (y))

55 Consider the univariate case d = 1. If P is an absolutely continuous distribution with cdf F, then D H (y, P) = min(f(y), 1 F (y)), so that R τ (P) = [ F 1 (τ), F 1 (1 τ) ], τ (0, 1/2]

56 Consider the univariate case d = 1. If P is an absolutely continuous distribution with cdf F, then D H (y, P) = min(f(y), 1 F (y)), so that R τ (P) = [ F 1 (τ), F 1 (1 τ) ], τ (0, 1/2] Hence, for d = 1, There is a strong connection between depth and quantiles The τ-depth region is associated with a pair of quantiles (orders τ, 1 τ) The most central region is the singleton R 1 (P) = {F 1 ( 1 )} = {Med(P)} 2 2 The R τ (P) s are L 1 -, hence robust, functionals

57 Consider the univariate case d = 1. If P is an absolutely continuous distribution with cdf F, then D H (y, P) = min(f(y), 1 F (y)), so that R τ (P) = [ F 1 (τ), F 1 (1 τ) ], τ (0, 1/2] Hence, for d = 1, There is a strong connection between depth and quantiles The τ-depth region is associated with a pair of quantiles (orders τ, 1 τ) The most central region is the singleton R 1 (P) = {F 1 ( 1 )} = {Med(P)} 2 2 The R τ (P) s are L 1 -, hence robust, functionals Does such a connection exist for d > 1?

58 Note that the end points of R τ (P) = [ F 1 (τ), F 1 (1 τ) ] may be regarded as The quantile of order τ in direction u = 1 The quantile of order τ in direction u = 1

59 Note that the end points of R τ (P) = [ F 1 (τ), F 1 (1 τ) ] may be regarded as The quantile of order τ in direction u = 1 The quantile of order τ in direction u = 1 This suggests directional quantiles, indexed by an order τ (0, 1) a unit vector u S 0 = { 1, 1}.

60 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles 1 Statistical depth Halfspace depth More depths Possible applications 2 3 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles

61 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Many multivariate quantiles are indexed by an order p [0, 1) a unit vector u S d 1 = {y R d : y = 1}, or equivalently, by p = pu B 1 (0) = {y R d : y < 1}. In particular, a d-dimensional quantile function, in Serfling (2006), is a mapping of the form Q(, P) : B 1 (0) R d p Q(p, P), where p is an "outlyingness" parameter and u indeed represents a direction in some sense (e.g., direction from M(P) = Med(P) = Q(0, P) to y = Q(p, P)).

62 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Example of quantile function: the spatial quantile function, defined through Q(p, P) = arg min y R d E P [ Y y + p (Y y)], that coincides with the solution (in y) of [ y Y ] p = E P ; y Y Dudley and Koltchinskii (1992), Chaudhuri (1996, JASA), and Koltchinskii (1997, AoS). For d = 1, Q(pu, P) = F 1 (τ), with τ = (pu + 1)/2. Orthogonal-equivariance, but no affine-equivariance. An affine-equivariant version was defined in Serfling (2010). For p = 0, Q(p, P) is the spatial median... The deepest point for the spatial depth. More generally, D Sp (y, P) = 1 p(y, P) = 1 Q 1 (y, P). (connection between spatial depth and spatial quantiles).

63 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Depth-induced quantile functions (Serfling 2006) : Let D(, P) be a depth, with depth regions R τ (P) and deepest point M(P), say. Denote as R p (P) the depth region with P-probability p. One can then associate with D(, P) the quantile function defined by { {M(P) + λu : λ > 0} R p (P) if p 0 Q(p, P) = Q(pu, P) = M(P) if p =

64 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Depth-induced quantile functions (Serfling 2006) : Let D(, P) be a depth, with depth regions R τ (P) and deepest point M(P), say. Denote as R p (P) the depth region with P-probability p. One can then associate with D(, P) the quantile function defined by { {M(P) + λu : λ > 0} R p (P) if p 0 Q(p, P) = Q(pu, P) = M(P) if p = 0. Parallel to spatial quantiles, these depth-induced quantiles are points in R d. We argue that (directional) quantiles should not be points, but surfaces, such as hyperplanes... Critical values.

65 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles The depth region R τ (P) rewrites R τ (P) = [ F 1 (τ), F 1 (1 τ) ], τ (0, 1/2] = [ F 1 (τ), ) (, F 1 (1 τ) ] =: H + τ,1(p) H + τ, 1(P), hence is the intersection of two directional τ-upper quantile halfspaces. Quantile (halfspace) at level τ =.25 in direction u = 1 Quantile (halfspace) at level τ =.25 in direction u = 1

66 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Let Y be a random d-vector. Let τ = τu, where τ = τ (0, 1) and u S d 1. Kong and Mizera (arxiv 2008, SS 2012) define the τ-quantile of Y as the point q KM,τ := F 1 u Y (τ)u. Or equivalently, one may consider the quantile hyperplane" π KM,τ := { } y R d : u (y q KM,τ ) = 0. Note that π KM,τ is orthogonal to u at q KM,τ.

67 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles

68 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles π (n) KM,τu, with u = (0, 1), τ =.1 (n = 500)

69 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles π (n) KM,τu, with u = (0, 1), τ {.1,.2,.3,.4} (n = 500)

70 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles π (n) KM,τu, with u {(0, 1), (1, 0) }, τ {.1,.2,.3,.4} (n = 500)

71 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles π (n) KM,τu, with u {(0, 1), (1, 0), ( 1 1 2, 2 ) }, τ {.1,.2,.3,.4} (n = 500)

72 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles π (n) KM,τu, with u {(0, ±1), (±1, 0), (± 1 2, ± 1 2 ) }, τ =.1 (n = 500)

73 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles π (n) KM,τu, with 512 u S1, τ =.1 (n = 500)

74 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles π (n) KM,τu, with 512 u S1, τ =.1 (n = 500)

75 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Defining H (n)+ KM,τ := {y R d : u (y q (n) KM,τ ) 0}, we have that, for any τ > 0, u S d 1H (n)+ KM,τu = Rτ (P(n) ). This establishes a connection, in the general multivariate case, between halfspace depth regions and "directional upper quantile halfspaces".

76 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles The depth region R τ (P) rewrites R τ (P) = [ F 1 (τ), F 1 (1 τ) ], τ (0, 1/2] = [ F 1 (τ), ) (, F 1 (1 τ) ] =: H + τ,1(p) H + τ, 1(P), hence is the intersection of two directional τ-upper quantile halfspaces. Quantile (halfspace) at level τ =.25 in direction u = 1 Quantile (halfspace) at level τ =.25 in direction u = 1

77 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Defining H (n)+ KM,τ := {y R d : u (y q (n) KM,τ ) 0}, we have that, for any τ > 0, u S d 1H (n)+ KM,τu = Rτ (P(n) ). This establishes a connection, in the general multivariate case, between halfspace depth regions and "directional upper quantile halfspaces". This connection, however, is quite an indirect one : - The depth regions R τ (P (n) ) cannot be constructed from these quantile hyperplanes, as this would involve an infinite number of u values. - This does not allow for importing to depth the analytical and computational tools related with the L 1 nature of quantiles (Bahadur representations and L 1 simplex algorithms).

78 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles... The L 1 nature of quantile of a random variable Y. Under second-order moment assumptions, it is well-known that E[Y ] = arg min a R E[(Y a)2 ]. Replacing L 2 losses with L 1 losses yields More generally, for any τ (0, 1), Med[Y ] = F 1 (1/2) = arg min E[ Y a ]. a R F 1 (τ) = arg min E[ρτ (Y a)], a R with ρ τ (z) := { (1 τ) z if z < 0 τ z if z 0.

79 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Often covariates are available and it makes sense to consider Y s quantiles conditionally upon the covariates in X = (X 1,..., X p) = (1, W ). n = 1000 observations from Y = (1 + X) + ε, X U([ 3, 2]) ε N (0, 1)

80 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Koenker and Basset (1978) define the regression τ-quantile of Y given X as a τ X, with a τ := arg min a R p E[ρτ (Y a X)]. If X = (1, W ) = 1, then a τ X = a τ reduces to the (unconditional) τ-quantile F 1 Y (τ) of Y.

81 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles n = 1000 observations from Y = (1 + W ) + ε, W U([ 3, 2]) ε N (0, 1)

82 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles n = 1000 observations from Y = (1 + W ) + (W + 4)ε, W U([ 3, 2]) ε N (0, 1)

83 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles By now, nonlinear or even nonparametric quantile regression has become standard.

84 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Hallin, Paindaveine and Šiman (2010, AoS) propose an alternative concept of directional quantiles. For any τ (0, 1), and any direction u S d 1, we define the (τ u)-quantile hyperplane π τu (π (n) τu in the empirical case) as the Koenker and Bassett regression τ-quantile hyperplane once u has been chosen as the (oriented) "vertical axis" in the computation of L 1 deviations.

85 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles More specifically, decompose y = (u y)u + Γ u(γ uy), y R d where Γ u is such that the columns of (u, Γ u) form an orthogonal basis of R d. Then π τu and π (n) τu are the hyperplanes with equations u y c τuγ uy a τu = 0 and u y c (n) τu Γ uy a τu = 0 minimizing, w.r.t. c R d 1 and a R, E[ρ τ (u Y c Γ uy a)] and n ρ τ (u Y i c Γ uy i a), i=1 respectively.

86 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles We define the (τ u)-quantile hyperplane as the Koenker and Bassett regression τ-quantile hyperplane once u has been chosen as the (oriented) "vertical axis". τ =.1,.2,.3 u = ( )

87 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles We define the (τ u)-quantile hyperplane as the Koenker and Bassett regression τ-quantile hyperplane once u has been chosen as the (oriented) "vertical axis". τ =.1,.2,.3 u = ( )

88 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles For the π (n) τu s, we establish consistency, Bahadur-type representation, and asymptotic normality.

89 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Let and ξ i,τ (a, c) := ( ) ( 1 ) t I [u Y i c Γ u Y i a<0], Y i H τ := R m 1 ( 1 z z zz ) f ((a τ + c τ z)u + Γ uz) dz ( ) t(1 t) t(1 t)e[y V τ := J ] u t(1 t)e[y ] Var[(t I [Y R m \Hτ )Y ] J u, ] where J u := diag(1, Γ u). Theorem Under mild regularity assumptions, we have that, as n, ( ) a τ (n) a τ n = 1 n H 1 τ J u ξ i,τ c τ n (aτ, cτ ) + o P (1) c (n) τ i=1 L N m(0, Hτ 1 V τ H 1 τ ).

90 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles For the π (n) τu s, we establish consistency, Bahadur-type representation, and asymptotic normality. We also provide efficient linear programming methods for the construction of the π (n) τu s and the corresponding inner regions...

91 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles We define the (τ u)-quantile hyperplane as the Koenker and Bassett regression τ-quantile hyperplane once u has been chosen as the (oriented) "vertical axis". τ =.1 u = ( )

92 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles We define the (τ u)-quantile hyperplane as the Koenker and Bassett regression τ-quantile hyperplane once u has been chosen as the (oriented) "vertical axis". u = ( 0 ±1 τ =.1 ) (, ±1 0 ), ( ±1/ 2 ±1/ 2 i.e., 8 equispaced u s )

93 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles We define the (τ u)-quantile hyperplane as the Koenker and Bassett regression τ-quantile hyperplane once u has been chosen as the (oriented) "vertical axis". τ = equispaced u s

94 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles We define the (τ u)-quantile hyperplane as the Koenker and Bassett regression τ-quantile hyperplane once u has been chosen as the (oriented) "vertical axis". τ = equispaced u s

95 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Denoting as H (n)+ τ the hallfspace "above" π (n) τ, we have that, for any τ > 0, u S d 1H (n)+ τu = R τ (P (n) ). This establishes another connection, in the general multivariate case, between halfspace depth regions and "directional upper quantile halfspaces". Unlike the previous connection, however, this intersection is over finitely many u s.

96 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles π (n) KM,τu, with 512 u S1, τ =.1 (n = 500)

97 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles π (n) τu, with 512 u S 1, τ =.1 (n = 500)

98 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles For n = 15 data points only : π (n) τu, with 512 u S 1, τ =.1 (n = 15)

99 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles For n = 15 data points only : π (n) KM,τu, with 512 u S1, τ =.1 (n = 15)

100 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Denoting as H (n)+ τ the hallfspace "above" π (n) τ, we have that, for any τ > 0, u S d 1H (n)+ τu = R τ (P (n) ). This establishes another connection, in the general multivariate case, between halfspace depth regions and "directional upper quantile halfspaces". Unlike the previous connection, however, this intersection is over finitely many u s. We showed that this bridge between the quantile and depth worlds makes it possible to import linear programming computability to depth.

101 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles We showed that the breaks in the process u H (n)+ τu (by means of parametric linear programming). can be identified efficiently Hence, we provide efficient and exact depth contour computations

102 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles We showed that the breaks in the process u H (n)+ τu (by means of parametric linear programming). can be identified efficiently Hence, we provide efficient and exact depth contour computations

103 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Some halfspace depth contours from n = 449 i.i.d. bivariate observations with independent U([.5,.5]), N (0, 1), and t 1 marginals

104 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles P = uniform on [0, 1] 2 P with independent t 1 marginals Some depth regions R τ (P)

105 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles π (n) τu, with u = {(±1, 0, 0), (0, ±1, 0), (0, 0, ±1) }, τ =.1, and the resulting halfspace depth contour from n = 49 i.i.d. observations with independent U([.5,.5]) marginals

106 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Some halfspace depth contours from n = 449 i.i.d. trivariate observations with independent U([.5,.5]), N (0, 1), and t 1 marginals

107 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles This provides a (directional) quantile that is easy to compute that is affine-equivariant from which (halfspace) depth regions can be computed, and that characterizes the distribution of Y What about the regression case?

108 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles

109 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles References Bai, Z.-D., and He, X. (1999). Asymptotic Distributions of the Maximal Depth Estimators for Regression and Multivariate Location. Ann. Statist Chaudhuri, P. (1996). On a geometric notion of quantiles for multivariate data. J. Amer. Statist. Assoc Dudley, R. M., and Koltchinskii, V. I. (1992). The spatial quantiles. Unpublished manuscript. Dutta, S., Ghosh, A. K., and Chaudhuri, P. (2011). Some intriguing properties of TukeyÕs half-space depth. Bernoulli Dutta, S. and Ghosh, A. K. (2012a). On robust classification using projection depth. Ann. Inst. Statist. Math Dutta, S., and Ghosh, A. K. (2012b). "On Classification Based on Lp -Depth with an Adaptive Choice of p." Submitted. Ghosh, A. K. and Chaudhuri, P. (2005). On maximum depth and related classifiers. Scand. J. Statist Hallin, M., Paindaveine, D., and Šiman, M. (2010). and multiple-output regression quantiles: from L 1 optimization to halfspace depth. Ann. Statist. 38, (with discussion). Hallin, M., Lu, Z., Paindaveine, D., and Šiman, M. (2013). Local constant and local bilinear multiple-output quantile regression. Submitted. Hubert, M., and Van der Veeken, S. (2012). "Robust Classification for Skewed Data," Advances in Data Analysis and Classification, to appear. Koenker, R., and Bassett, G. J. (1978). Regression quantiles. Econometrica Koshevoy, G.A. (2003). Lift-zonoid and multivariate depths. In: R. Dutter, P. Filzmoser, U. Gather, P.J. Rousseeuw (Eds.), Developments in Robust Statistics: International Conference on Robust Statistics 2001, Springer-Physica, Heidelberg, Koltchinskii, V. (1997). M-estimation, convexity and quantiles. Ann. Statist Kong, L., and Mizera, I. (2008). Quantile tomography: using quantiles with multivariate data. Statistica Sinica 22, Kong, L., and Mizera, I. (2012). Quantile tomography: using quantiles with multivariate data. Statistica Sinica 22, Kong, L. and Zuo, Y. (2010). Smooth depth contours characterize the underlying distribution. J. Multivariate Anal Li, J., Cuesta-Albertos, J. A., and Liu, R. Y. (2012). DD-Classifier: Nonparametric classification procedures based on DD-plots. J. Amer. Statist. Assoc

110 Serfling s D-O-Q(-R) paradigm, Chaudhuri (1996) Kong and Mizera (2012) HPŠ quantiles Liu, R. Y. (1990). On a notion of data depth based on random simplices. Ann. Statist Liu, R. Y., Parelius, J. M., and Singh, K. (1999). "Multivariate Analysis by Data Depth: Descriptive Statistics, Graphics and Inference" (with discussion), Ann. Statist Massé, J.-C. (2002). Asymptotics for the Tukey Median. J. Multivariate Anal Milasevic, P., and Ducharme, G. R. (1987). Uniqueness of the Spatial Median. Ann. Statist Paindaveine, D., and Šiman, M. (2012). Computing multiple-output regression quantile regions (with M. Siman). Computational Statistics and Data Analysis 56, Paindaveine, D. and Van Bever, G. (2012). Nonparametrically consistent depth-based classifiers. ECARES working paper Rousseeuw, P. J., and Ruts, I. (1999). The depth function of a population distribution. Metrika Rousseeuw, P. J. and Struyf, A. (2002). A depth test for symmetry. In Goodness-of-fit tests and model validity (Paris, 2000), Boston, MA: Birkhäuser Boston, Stat. Ind. Technol., pp Rousseeuw, P. J., and Struyf, A. (2004). Characterizing angular symmetry and regression symmetry. J. Statist. Plann. Inference Serfling, R. (2006). Depth functions in nonparametric multivariate inference. In Data Depth: Robust Multivariate Analysis, Computational Geometry and Applications (R. Y. Liu, R. Serfling, D. L. Souvaine, eds.), pp DIMACS Series in Discrete Mathematics and Theoretical Computer Science, Volume 72, American Mathematical Society. Serfling, R. (2010). Equivariance and invariance properties of multivariate quantile and related functions, and the role of standardization, J. Nonparametr. Stat. 22, Struyf, A., and Rousseeuw, P. J. (2005). Halfspace depth and regression depth characterize the empirical distribution. J. Multivariate Anal Tukey, J. W. (1975). Mathematics and the picturing of data. In Proceedings of the international congress of mathematicians (Vancouver, B. C., 1974), Vol. 2, Quebec: Canad. Math. Congress Vardi, Y. and Zhang, C.-H. (2000). The multivariate L 1 -median and associated data depth. Proceedings National Academy of Science USA Zuo, Y., and Serfling, R. (2000). "General notions of statistical depth function," Ann. Statist

Robust estimation of principal components from depth-based multivariate rank covariance matrix

Robust estimation of principal components from depth-based multivariate rank covariance matrix Robust estimation of principal components from depth-based multivariate rank covariance matrix Subho Majumdar Snigdhansu Chatterjee University of Minnesota, School of Statistics Table of contents Summary

More information

Supervised Classification for Functional Data Using False Discovery Rate and Multivariate Functional Depth

Supervised Classification for Functional Data Using False Discovery Rate and Multivariate Functional Depth Supervised Classification for Functional Data Using False Discovery Rate and Multivariate Functional Depth Chong Ma 1 David B. Hitchcock 2 1 PhD Candidate University of South Carolina 2 Associate Professor

More information

CONTRIBUTIONS TO THE THEORY AND APPLICATIONS OF STATISTICAL DEPTH FUNCTIONS

CONTRIBUTIONS TO THE THEORY AND APPLICATIONS OF STATISTICAL DEPTH FUNCTIONS CONTRIBUTIONS TO THE THEORY AND APPLICATIONS OF STATISTICAL DEPTH FUNCTIONS APPROVED BY SUPERVISORY COMMITTEE: Robert Serfling, Chair Larry Ammann John Van Ness Michael Baron Copyright 1998 Yijun Zuo All

More information

Statistical Depth Function

Statistical Depth Function Machine Learning Journal Club, Gatsby Unit June 27, 2016 Outline L-statistics, order statistics, ranking. Instead of moments: median, dispersion, scale, skewness,... New visualization tools: depth contours,

More information

Nonparametrically Consistent. Depth-based Classifiers

Nonparametrically Consistent. Depth-based Classifiers Nonparametrically Consistent Depth-based Classifiers Davy Paindaveine and Germain Van Bever Université Libre de Bruelles, Brussels, Belgium Abstract We introduce a class of depth-based classification procedures

More information

From Depth to Local Depth : A Focus on Centrality

From Depth to Local Depth : A Focus on Centrality From Depth to Local Depth : A Focus on Centrality Davy Paindaveine and Germain Van Bever Université libre de Bruelles, Brussels, Belgium Abstract Aiming at analysing multimodal or non-convely supported

More information

WEIGHTED HALFSPACE DEPTH

WEIGHTED HALFSPACE DEPTH K Y BERNETIKA VOLUM E 46 2010), NUMBER 1, P AGES 125 148 WEIGHTED HALFSPACE DEPTH Daniel Hlubinka, Lukáš Kotík and Ondřej Vencálek Generalised halfspace depth function is proposed. Basic properties of

More information

DD-Classifier: Nonparametric Classification Procedure Based on DD-plot 1. Abstract

DD-Classifier: Nonparametric Classification Procedure Based on DD-plot 1. Abstract DD-Classifier: Nonparametric Classification Procedure Based on DD-plot 1 Jun Li 2, Juan A. Cuesta-Albertos 3, Regina Y. Liu 4 Abstract Using the DD-plot (depth-versus-depth plot), we introduce a new nonparametric

More information

Quantile Processes for Semi and Nonparametric Regression

Quantile Processes for Semi and Nonparametric Regression Quantile Processes for Semi and Nonparametric Regression Shih-Kang Chao Department of Statistics Purdue University IMS-APRM 2016 A joint work with Stanislav Volgushev and Guang Cheng Quantile Response

More information

Chapter 2 Depth Statistics

Chapter 2 Depth Statistics Chapter 2 Depth Statistics Karl Mosler 2.1 Introduction In 1975, John Tukey proposed a multivariate median which is the deepest point in a given data cloud in R d (Tukey 1975). In measuring the depth of

More information

Jensen s inequality for multivariate medians

Jensen s inequality for multivariate medians J. Math. Anal. Appl. [Journal of Mathematical Analysis and Applications], 370 (2010), 258-269 Jensen s inequality for multivariate medians Milan Merkle 1 Abstract. Given a probability measure µ on Borel

More information

MULTIVARIATE QUANTILES AND MULTIPLE-OUTPUT REGRESSION QUANTILES: FROM L 1 OPTIMIZATION TO HALFSPACE DEPTH

MULTIVARIATE QUANTILES AND MULTIPLE-OUTPUT REGRESSION QUANTILES: FROM L 1 OPTIMIZATION TO HALFSPACE DEPTH The Annals of Statistics 2010, Vol. 38, No. 2, 635 669 DOI: 10.1214/09-AOS723 Institute of Mathematical Statistics, 2010 MULTIVARIATE QUANTILES AND MULTIPLE-OUTPUT REGRESSION QUANTILES: FROM L 1 OPTIMIZATION

More information

Extreme geometric quantiles

Extreme geometric quantiles 1/ 30 Extreme geometric quantiles Stéphane GIRARD (Inria Grenoble Rhône-Alpes) Joint work with Gilles STUPFLER (Aix Marseille Université) ERCIM, Pisa, Italy, December 2014 2/ 30 Outline Extreme multivariate

More information

Jensen s inequality for multivariate medians

Jensen s inequality for multivariate medians Jensen s inequality for multivariate medians Milan Merkle University of Belgrade, Serbia emerkle@etf.rs Given a probability measure µ on Borel sigma-field of R d, and a function f : R d R, the main issue

More information

Fast and Robust Classifiers Adjusted for Skewness

Fast and Robust Classifiers Adjusted for Skewness Fast and Robust Classifiers Adjusted for Skewness Mia Hubert 1 and Stephan Van der Veeken 2 1 Department of Mathematics - LStat, Katholieke Universiteit Leuven Celestijnenlaan 200B, Leuven, Belgium, Mia.Hubert@wis.kuleuven.be

More information

YIJUN ZUO. Education. PhD, Statistics, 05/98, University of Texas at Dallas, (GPA 4.0/4.0)

YIJUN ZUO. Education. PhD, Statistics, 05/98, University of Texas at Dallas, (GPA 4.0/4.0) YIJUN ZUO Department of Statistics and Probability Michigan State University East Lansing, MI 48824 Tel: (517) 432-5413 Fax: (517) 432-5413 Email: zuo@msu.edu URL: www.stt.msu.edu/users/zuo Education PhD,

More information

Tests Using Spatial Median

Tests Using Spatial Median AUSTRIAN JOURNAL OF STATISTICS Volume 35 (2006), Number 2&3, 331 338 Tests Using Spatial Median Ján Somorčík Comenius University, Bratislava, Slovakia Abstract: The multivariate multi-sample location problem

More information

Quantile Regression for Extraordinarily Large Data

Quantile Regression for Extraordinarily Large Data Quantile Regression for Extraordinarily Large Data Shih-Kang Chao Department of Statistics Purdue University November, 2016 A joint work with Stanislav Volgushev and Guang Cheng Quantile regression Two-step

More information

On the Computation of Multivariate Scenario Sets for the Skew-t and Generalized Hyperbolic Families

On the Computation of Multivariate Scenario Sets for the Skew-t and Generalized Hyperbolic Families On the Computation of Multivariate Scenario Sets for the Skew-t and Generalized Hyperbolic Families Emanuele Giorgi 1,2, Alexander J. McNeil 3,4 February 5, 2014 arxiv:1402.0686v1 [math.st] 4 Feb 2014

More information

ON MULTIVARIATE MONOTONIC MEASURES OF LOCATION WITH HIGH BREAKDOWN POINT

ON MULTIVARIATE MONOTONIC MEASURES OF LOCATION WITH HIGH BREAKDOWN POINT Sankhyā : The Indian Journal of Statistics 999, Volume 6, Series A, Pt. 3, pp. 362-380 ON MULTIVARIATE MONOTONIC MEASURES OF LOCATION WITH HIGH BREAKDOWN POINT By SUJIT K. GHOSH North Carolina State University,

More information

Influence Functions for a General Class of Depth-Based Generalized Quantile Functions

Influence Functions for a General Class of Depth-Based Generalized Quantile Functions Influence Functions for a General Class of Depth-Based Generalized Quantile Functions Jin Wang 1 Northern Arizona University and Robert Serfling 2 University of Texas at Dallas June 2005 Final preprint

More information

A Convex Hull Peeling Depth Approach to Nonparametric Massive Multivariate Data Analysis with Applications

A Convex Hull Peeling Depth Approach to Nonparametric Massive Multivariate Data Analysis with Applications A Convex Hull Peeling Depth Approach to Nonparametric Massive Multivariate Data Analysis with Applications Hyunsook Lee. hlee@stat.psu.edu Department of Statistics The Pennsylvania State University Hyunsook

More information

arxiv: v2 [math.st] 10 Oct 2012

arxiv: v2 [math.st] 10 Oct 2012 Depth statistics Karl Mosler arxiv:1207.4988v2 [math.st] 10 Oct 2012 1 Introduction In 1975 John Tukey proposed a multivariate median which is the deepest point in a given data cloud inr d (Tukey, 1975).

More information

Jyh-Jen Horng Shiau 1 and Lin-An Chen 1

Jyh-Jen Horng Shiau 1 and Lin-An Chen 1 Aust. N. Z. J. Stat. 45(3), 2003, 343 352 A MULTIVARIATE PARALLELOGRAM AND ITS APPLICATION TO MULTIVARIATE TRIMMED MEANS Jyh-Jen Horng Shiau 1 and Lin-An Chen 1 National Chiao Tung University Summary This

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

Monte Carlo Studies. The response in a Monte Carlo study is a random variable.

Monte Carlo Studies. The response in a Monte Carlo study is a random variable. Monte Carlo Studies The response in a Monte Carlo study is a random variable. The response in a Monte Carlo study has a variance that comes from the variance of the stochastic elements in the data-generating

More information

Inequalities Relating Addition and Replacement Type Finite Sample Breakdown Points

Inequalities Relating Addition and Replacement Type Finite Sample Breakdown Points Inequalities Relating Addition and Replacement Type Finite Sample Breadown Points Robert Serfling Department of Mathematical Sciences University of Texas at Dallas Richardson, Texas 75083-0688, USA Email:

More information

Monge-Kantorovich depth, quantiles, ranks and signs

Monge-Kantorovich depth, quantiles, ranks and signs Monge-Kantorovich depth, quantiles, ranks and signs Victor Chernozhukov Alfred Galichon Marc Hallin Marc Henry The Institute for Fiscal Studies Department of Economics, UCL cemmap working paper CWP57/15

More information

Robust Classification for Skewed Data

Robust Classification for Skewed Data Advances in Data Analysis and Classification manuscript No. (will be inserted by the editor) Robust Classification for Skewed Data Mia Hubert Stephan Van der Veeken Received: date / Accepted: date Abstract

More information

Quantile methods. Class Notes Manuel Arellano December 1, Let F (r) =Pr(Y r). Forτ (0, 1), theτth population quantile of Y is defined to be

Quantile methods. Class Notes Manuel Arellano December 1, Let F (r) =Pr(Y r). Forτ (0, 1), theτth population quantile of Y is defined to be Quantile methods Class Notes Manuel Arellano December 1, 2009 1 Unconditional quantiles Let F (r) =Pr(Y r). Forτ (0, 1), theτth population quantile of Y is defined to be Q τ (Y ) q τ F 1 (τ) =inf{r : F

More information

Methods of Nonparametric Multivariate Ranking and Selection

Methods of Nonparametric Multivariate Ranking and Selection Syracuse University SURFACE Mathematics - Dissertations Mathematics 8-2013 Methods of Nonparametric Multivariate Ranking and Selection Jeremy Entner Follow this and additional works at: http://surface.syr.edu/mat_etd

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

On corrections of classical multivariate tests for high-dimensional data. Jian-feng. Yao Université de Rennes 1, IRMAR

On corrections of classical multivariate tests for high-dimensional data. Jian-feng. Yao Université de Rennes 1, IRMAR Introduction a two sample problem Marčenko-Pastur distributions and one-sample problems Random Fisher matrices and two-sample problems Testing cova On corrections of classical multivariate tests for high-dimensional

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University JSM, 2015 E. Christou, M. G. Akritas (PSU) SIQR JSM, 2015

More information

On the limiting distributions of multivariate depth-based rank sum. statistics and related tests. By Yijun Zuo 2 and Xuming He 3

On the limiting distributions of multivariate depth-based rank sum. statistics and related tests. By Yijun Zuo 2 and Xuming He 3 1 On the limiting distributions of multivariate depth-based rank sum statistics and related tests By Yijun Zuo 2 and Xuming He 3 Michigan State University and University of Illinois A depth-based rank

More information

A Resampling Method on Pivotal Estimating Functions

A Resampling Method on Pivotal Estimating Functions A Resampling Method on Pivotal Estimating Functions Kun Nie Biostat 277,Winter 2004 March 17, 2004 Outline Introduction A General Resampling Method Examples - Quantile Regression -Rank Regression -Simulation

More information

MULTIVARIATE SYMMETRY AND ASYMME- TRY

MULTIVARIATE SYMMETRY AND ASYMME- TRY MULTIVARIATE SYMMETRY AND ASYMME- TRY (By Robert Serfling, January 2003, for Encyclopedia of Statistical Sciences, 2nd ed.) Abstract. Univariate symmetry has interesting and diverse forms of generalization

More information

APPLICATION OF FUNCTIONAL BASED ON SPATIAL QUANTILES

APPLICATION OF FUNCTIONAL BASED ON SPATIAL QUANTILES Studia Ekonomiczne. Zeszyty Naukowe Uniwersytetu Ekonomicznego w Katowicach ISSN 2083-8611 Nr 247 2015 Informatyka i Ekonometria 4 University of Economics in Katowice aculty of Informatics and Communication

More information

Issues on quantile autoregression

Issues on quantile autoregression Issues on quantile autoregression Jianqing Fan and Yingying Fan We congratulate Koenker and Xiao on their interesting and important contribution to the quantile autoregression (QAR). The paper provides

More information

BY ANDRZEJ S. KOZEK Macquarie University

BY ANDRZEJ S. KOZEK Macquarie University The Annals of Statistics 2003, Vol. 31, No. 4, 1170 1185 Institute of Mathematical Statistics, 2003 ON M-ESTIMATORS AND NORMAL QUANTILES BY ANDRZEJ S. KOZEK Macquarie University This paper explores a class

More information

DIRECTIONAL QUANTILE REGRESSION IN R

DIRECTIONAL QUANTILE REGRESSION IN R K Y B E R N E T I K A V O L U M E 5 3 ( 2 0 1 7 ), N U M B E R 3, P A G E S 4 8 0 4 9 2 DIRECTIONAL QUANTILE REGRESSION IN R Pavel Boček and Miroslav Šiman Recently, the eminently popular standard quantile

More information

Local Quantile Regression

Local Quantile Regression Wolfgang K. Härdle Vladimir Spokoiny Weining Wang Ladislaus von Bortkiewicz Chair of Statistics C.A.S.E. - Center for Applied Statistics and Economics Humboldt-Universität zu Berlin http://ise.wiwi.hu-berlin.de

More information

Convergence of Multivariate Quantile Surfaces

Convergence of Multivariate Quantile Surfaces Convergence of Multivariate Quantile Surfaces Adil Ahidar Institut de Mathématiques de Toulouse - CERFACS August 30, 2013 Adil Ahidar (Institut de Mathématiques de Toulouse Convergence - CERFACS) of Multivariate

More information

ASYMPTOTIC NORMALITY UNDER TWO-PHASE SAMPLING DESIGNS

ASYMPTOTIC NORMALITY UNDER TWO-PHASE SAMPLING DESIGNS Statistica Sinica 17(2007), 1047-1064 ASYMPTOTIC NORMALITY UNDER TWO-PHASE SAMPLING DESIGNS Jiahua Chen and J. N. K. Rao University of British Columbia and Carleton University Abstract: Large sample properties

More information

MULTIVARIATE TECHNIQUES, ROBUSTNESS

MULTIVARIATE TECHNIQUES, ROBUSTNESS MULTIVARIATE TECHNIQUES, ROBUSTNESS Mia Hubert Associate Professor, Department of Mathematics and L-STAT Katholieke Universiteit Leuven, Belgium mia.hubert@wis.kuleuven.be Peter J. Rousseeuw 1 Senior Researcher,

More information

arxiv: v1 [stat.co] 2 Jul 2007

arxiv: v1 [stat.co] 2 Jul 2007 The Random Tukey Depth arxiv:0707.0167v1 [stat.co] 2 Jul 2007 J.A. Cuesta-Albertos and A. Nieto-Reyes Departamento de Matemáticas, Estadística y Computación, Universidad de Cantabria, Spain October 26,

More information

Limit Theorems for Quantile and Depth Regions for Stochastic Processes

Limit Theorems for Quantile and Depth Regions for Stochastic Processes Limit Theorems for Quantile and Depth Regions for Stochastic Processes James Kuelbs and Joel Zinn Abstract. Since contours of multi-dimensional depth functions often characterize the distribution, it has

More information

On corrections of classical multivariate tests for high-dimensional data

On corrections of classical multivariate tests for high-dimensional data On corrections of classical multivariate tests for high-dimensional data Jian-feng Yao with Zhidong Bai, Dandan Jiang, Shurong Zheng Overview Introduction High-dimensional data and new challenge in statistics

More information

A new bridge between fuzzy sets and statistics

A new bridge between fuzzy sets and statistics A new bridge between fuzzy sets and statistics Pedro Terán Departamento de Estadística e I.O. y D.M. Universidad de Oviedo E-33071 Gijón (Spain) Email: teranpedro@uniovi.es Abstract Our basic question

More information

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline. Practitioner Course: Portfolio Optimization September 10, 2008 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y ) (x,

More information

On Fractile Transformation of Covariates in. Regression

On Fractile Transformation of Covariates in. Regression On Fractile Transformation of Covariates in Regression Bodhisattva Sen and Probal Chaudhuri Columbia University New York, NY 27 Indian Statistical Institute 23 B.T. Road, Kolkata 78 Abstract The need for

More information

Introduction to Robust Statistics. Elvezio Ronchetti. Department of Econometrics University of Geneva Switzerland.

Introduction to Robust Statistics. Elvezio Ronchetti. Department of Econometrics University of Geneva Switzerland. Introduction to Robust Statistics Elvezio Ronchetti Department of Econometrics University of Geneva Switzerland Elvezio.Ronchetti@metri.unige.ch http://www.unige.ch/ses/metri/ronchetti/ 1 Outline Introduction

More information

Fiducial Inference and Generalizations

Fiducial Inference and Generalizations Fiducial Inference and Generalizations Jan Hannig Department of Statistics and Operations Research The University of North Carolina at Chapel Hill Hari Iyer Department of Statistics, Colorado State University

More information

Quantile Scale Curves

Quantile Scale Curves This article was downloaded by: [Zhejiang University] On: 24 March 2014, At: 16:06 Publisher: Taylor & Francis Informa Ltd Registered in England and Wales Registered Number: 1072954 Registered office:

More information

Testing Statistical Hypotheses

Testing Statistical Hypotheses E.L. Lehmann Joseph P. Romano Testing Statistical Hypotheses Third Edition 4y Springer Preface vii I Small-Sample Theory 1 1 The General Decision Problem 3 1.1 Statistical Inference and Statistical Decisions

More information

Program Evaluation with High-Dimensional Data

Program Evaluation with High-Dimensional Data Program Evaluation with High-Dimensional Data Alexandre Belloni Duke Victor Chernozhukov MIT Iván Fernández-Val BU Christian Hansen Booth ESWC 215 August 17, 215 Introduction Goal is to perform inference

More information

Small Sample Corrections for LTS and MCD

Small Sample Corrections for LTS and MCD myjournal manuscript No. (will be inserted by the editor) Small Sample Corrections for LTS and MCD G. Pison, S. Van Aelst, and G. Willems Department of Mathematics and Computer Science, Universitaire Instelling

More information

Depth-based Runs Tests for Bivariate Central Symmetry

Depth-based Runs Tests for Bivariate Central Symmetry epth-based Runs Tests for Bivariate Central Symmetry Rainer yckerhoff, Christophe Ley and avy Paindaveine Abstract McWilliams (1990) introduced a nonparametric procedure based on runs for the problem of

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression

More information

arxiv: v1 [stat.me] 27 Apr 2017

arxiv: v1 [stat.me] 27 Apr 2017 A note on Quantile curves based bivariate reliability concepts Sreelakshmi N Sudheesh K Kattumannil arxiv:74.8444v [stat.me] 27 Apr 27 Indian Statistical Institute, Chennai, India Abstract Several attempts

More information

A Bahadur Representation of the Linear Support Vector Machine

A Bahadur Representation of the Linear Support Vector Machine A Bahadur Representation of the Linear Support Vector Machine Yoonkyung Lee Department of Statistics The Ohio State University October 7, 2008 Data Mining and Statistical Learning Study Group Outline Support

More information

On Some Classification Methods for High Dimensional and Functional Data

On Some Classification Methods for High Dimensional and Functional Data On Some Classification Methods for High Dimensional and Functional Data by OLUSOLA SAMUEL MAKINDE A thesis submitted to The University of Birmingham for the degree of Doctor of Philosophy School of Mathematics

More information

Generalized Multivariate Rank Type Test Statistics via Spatial U-Quantiles

Generalized Multivariate Rank Type Test Statistics via Spatial U-Quantiles Generalized Multivariate Rank Type Test Statistics via Spatial U-Quantiles Weihua Zhou 1 University of North Carolina at Charlotte and Robert Serfling 2 University of Texas at Dallas Final revision for

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR

More information

A Bayesian Approach to Multiple-Output Quantile Regression

A Bayesian Approach to Multiple-Output Quantile Regression A Bayesian Approach to Multiple-Output Quantile Regression Michael Guggisberg Department of Economics University of California, Irvine November, 06 Abstract This paper proposes a Bayesian approach to Multiple-Output

More information

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Lecture 25: Review. Statistics 104. April 23, Colin Rundel Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April

More information

INFERENCE APPROACHES FOR INSTRUMENTAL VARIABLE QUANTILE REGRESSION. 1. Introduction

INFERENCE APPROACHES FOR INSTRUMENTAL VARIABLE QUANTILE REGRESSION. 1. Introduction INFERENCE APPROACHES FOR INSTRUMENTAL VARIABLE QUANTILE REGRESSION VICTOR CHERNOZHUKOV CHRISTIAN HANSEN MICHAEL JANSSON Abstract. We consider asymptotic and finite-sample confidence bounds in instrumental

More information

QUANTILE TOMOGRAPHY: USING QUANTILES WITH MULTIVARIATE DATA

QUANTILE TOMOGRAPHY: USING QUANTILES WITH MULTIVARIATE DATA Statistica Sinica 22 (2012), 1589-1610 doi:http://dx.doi.org/10.5705/ss.2010.224 QUANTILE TOMOGRAPHY: USING QUANTILES WITH MULTIVARIATE DATA Linglong Kong and Ivan Mizera University of Alberta Abstract:

More information

A Comparison of Robust Estimators Based on Two Types of Trimming

A Comparison of Robust Estimators Based on Two Types of Trimming Submitted to the Bernoulli A Comparison of Robust Estimators Based on Two Types of Trimming SUBHRA SANKAR DHAR 1, and PROBAL CHAUDHURI 1, 1 Theoretical Statistics and Mathematics Unit, Indian Statistical

More information

Discriminant Analysis with High Dimensional. von Mises-Fisher distribution and

Discriminant Analysis with High Dimensional. von Mises-Fisher distribution and Athens Journal of Sciences December 2014 Discriminant Analysis with High Dimensional von Mises - Fisher Distributions By Mario Romanazzi This paper extends previous work in discriminant analysis with von

More information

PROJECTION-BASED DEPTH FUNCTIONS AND ASSOCIATED MEDIANS 1. BY YIJUN ZUO Arizona State University and Michigan State University

PROJECTION-BASED DEPTH FUNCTIONS AND ASSOCIATED MEDIANS 1. BY YIJUN ZUO Arizona State University and Michigan State University AOS ims v.2003/03/12 Prn:25/04/2003; 16:41 F:aos126.tex; (aiste) p. 1 The Annals of Statistics 2003, Vol. 0, No. 0, 1 31 Institute of Mathematical Statistics, 2003 PROJECTION-BASED DEPTH FUNCTIONS AND

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Independent Component (IC) Models: New Extensions of the Multinormal Model

Independent Component (IC) Models: New Extensions of the Multinormal Model Independent Component (IC) Models: New Extensions of the Multinormal Model Davy Paindaveine (joint with Klaus Nordhausen, Hannu Oja, and Sara Taskinen) School of Public Health, ULB, April 2008 My research

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

Multivariate Functional Halfspace Depth

Multivariate Functional Halfspace Depth Multivariate Functional Halfspace Depth Gerda Claeskens, Mia Hubert, Leen Slaets and Kaveh Vakili 1 October 1, 2013 A depth for multivariate functional data is defined and studied. By the multivariate

More information

Practical tests for randomized complete block designs

Practical tests for randomized complete block designs Journal of Multivariate Analysis 96 (2005) 73 92 www.elsevier.com/locate/jmva Practical tests for randomized complete block designs Ziyad R. Mahfoud a,, Ronald H. Randles b a American University of Beirut,

More information

Approximate Data Depth Revisited. May 22, 2018

Approximate Data Depth Revisited. May 22, 2018 Approximate Data Depth Revisited David Bremner Rasoul Shahsavarifar May, 018 arxiv:1805.07373v1 [cs.cg] 18 May 018 Abstract Halfspace depth and β-skeleton depth are two types of depth functions in nonparametric

More information

An Introduction to Multivariate Statistical Analysis

An Introduction to Multivariate Statistical Analysis An Introduction to Multivariate Statistical Analysis Third Edition T. W. ANDERSON Stanford University Department of Statistics Stanford, CA WILEY- INTERSCIENCE A JOHN WILEY & SONS, INC., PUBLICATION Contents

More information

Median Cross-Validation

Median Cross-Validation Median Cross-Validation Chi-Wai Yu 1, and Bertrand Clarke 2 1 Department of Mathematics Hong Kong University of Science and Technology 2 Department of Medicine University of Miami IISA 2011 Outline Motivational

More information

THE BREAKDOWN POINT EXAMPLES AND COUNTEREXAMPLES

THE BREAKDOWN POINT EXAMPLES AND COUNTEREXAMPLES REVSTAT Statistical Journal Volume 5, Number 1, March 2007, 1 17 THE BREAKDOWN POINT EXAMPLES AND COUNTEREXAMPLES Authors: P.L. Davies University of Duisburg-Essen, Germany, and Technical University Eindhoven,

More information

Does Modeling Lead to More Accurate Classification?

Does Modeling Lead to More Accurate Classification? Does Modeling Lead to More Accurate Classification? A Comparison of the Efficiency of Classification Methods Yoonkyung Lee* Department of Statistics The Ohio State University *joint work with Rui Wang

More information

Statistics 612: L p spaces, metrics on spaces of probabilites, and connections to estimation

Statistics 612: L p spaces, metrics on spaces of probabilites, and connections to estimation Statistics 62: L p spaces, metrics on spaces of probabilites, and connections to estimation Moulinath Banerjee December 6, 2006 L p spaces and Hilbert spaces We first formally define L p spaces. Consider

More information

Supplementary Material for Wang and Serfling paper

Supplementary Material for Wang and Serfling paper Supplementary Material for Wang and Serfling paper March 6, 2017 1 Simulation study Here we provide a simulation study to compare empirically the masking and swamping robustness of our selected outlyingness

More information

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ). .8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Accurate and Powerful Multivariate Outlier Detection

Accurate and Powerful Multivariate Outlier Detection Int. Statistical Inst.: Proc. 58th World Statistical Congress, 11, Dublin (Session CPS66) p.568 Accurate and Powerful Multivariate Outlier Detection Cerioli, Andrea Università di Parma, Dipartimento di

More information

Flexible Estimation of Treatment Effect Parameters

Flexible Estimation of Treatment Effect Parameters Flexible Estimation of Treatment Effect Parameters Thomas MaCurdy a and Xiaohong Chen b and Han Hong c Introduction Many empirical studies of program evaluations are complicated by the presence of both

More information

ISSN X On misclassification probabilities of linear and quadratic classifiers

ISSN X On misclassification probabilities of linear and quadratic classifiers Afrika Statistika Vol. 111, 016, pages 943 953. DOI: http://dx.doi.org/10.1699/as/016.943.85 Afrika Statistika ISSN 316-090X On misclassification probabilities of linear and quadratic classifiers Olusola

More information

A Study of Relative Efficiency and Robustness of Classification Methods

A Study of Relative Efficiency and Robustness of Classification Methods A Study of Relative Efficiency and Robustness of Classification Methods Yoonkyung Lee* Department of Statistics The Ohio State University *joint work with Rui Wang April 28, 2011 Department of Statistics

More information

Bayesian Decision and Bayesian Learning

Bayesian Decision and Bayesian Learning Bayesian Decision and Bayesian Learning Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1 / 30 Bayes Rule p(x ω i

More information

Gaussian discriminant analysis Naive Bayes

Gaussian discriminant analysis Naive Bayes DM825 Introduction to Machine Learning Lecture 7 Gaussian discriminant analysis Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. is 2. Multi-variate

More information

Stahel-Donoho Estimation for High-Dimensional Data

Stahel-Donoho Estimation for High-Dimensional Data Stahel-Donoho Estimation for High-Dimensional Data Stefan Van Aelst KULeuven, Department of Mathematics, Section of Statistics Celestijnenlaan 200B, B-3001 Leuven, Belgium Email: Stefan.VanAelst@wis.kuleuven.be

More information

Mutiple-Output Quantile Regression through Optimal Quantization

Mutiple-Output Quantile Regression through Optimal Quantization Mutiple-Output Quantile Regression through Optimal Quantization Isabelle Charlier 1,2 Davy Paindaveine 1 Jérôme Saracco 2 April 3, 2016 1 Université Libre de Bruxelles, ECARES and Département de Mathématique,

More information

ECAS Summer Course. Fundamentals of Quantile Regression Roger Koenker University of Illinois at Urbana-Champaign

ECAS Summer Course. Fundamentals of Quantile Regression Roger Koenker University of Illinois at Urbana-Champaign ECAS Summer Course 1 Fundamentals of Quantile Regression Roger Koenker University of Illinois at Urbana-Champaign La Roche-en-Ardennes: September, 2005 An Outline 2 Beyond Average Treatment Effects Computation

More information

Klasifikace pomocí hloubky dat nové nápady

Klasifikace pomocí hloubky dat nové nápady Klasifikace pomocí hloubky dat nové nápady Ondřej Vencálek Univerzita Palackého v Olomouci ROBUST Rybník, 26. ledna 2018 Z. Fabián: Vzpomínka na Compstat 2010 aneb večeře na Pařížské radnici, Informační

More information

Lecture 7 October 13

Lecture 7 October 13 STATS 300A: Theory of Statistics Fall 2015 Lecture 7 October 13 Lecturer: Lester Mackey Scribe: Jing Miao and Xiuyuan Lu 7.1 Recap So far, we have investigated various criteria for optimal inference. We

More information

A talk on Oracle inequalities and regularization. by Sara van de Geer

A talk on Oracle inequalities and regularization. by Sara van de Geer A talk on Oracle inequalities and regularization by Sara van de Geer Workshop Regularization in Statistics Banff International Regularization Station September 6-11, 2003 Aim: to compare l 1 and other

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline.

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline. MFM Practitioner Module: Risk & Asset Allocation September 11, 2013 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y

More information

Multivariate Statistics

Multivariate Statistics Multivariate Statistics Chapter 2: Multivariate distributions and inference Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2016/2017 Master in Mathematical

More information