Quantile Regression for Extraordinarily Large Data

Size: px
Start display at page:

Download "Quantile Regression for Extraordinarily Large Data"

Transcription

1 Quantile Regression for Extraordinarily Large Data Shih-Kang Chao Department of Statistics Purdue University November, 2016 A joint work with Stanislav Volgushev and Guang Cheng

2 Quantile regression Two-step algorithm Oracle rules Simulation References With the advance of technology, it is increasingly common that data set is so large that it cannot be stored in a single machine Social media (views, likes, comments, images...) Meteorological and environmental surveillance Transactions in e-commerce Others... Figure: A server room in Council Bluffs, Iowa. Photo: Google/Connie Zhou.

3 Divide and Conquer (D&C) To take advantage of the opportunities in massive data, we need to deal with storing (disc) and computational (memory, processor) bottlenecks. Divide and conquer paradigm: Randomly divide N samples into S groups. n = N/S Problem(N) subproblem (n) subproblem (n) subproblem (n) subproblem (n)

4 Aggregate solutions from subproblems to get a solution for original problem Can be implemented by computational platforms such as Hadoop (White, 2012) Data that are stored in distributed locations can be analyzed in similar manner Problem(N) subproblem (n) subproblem (n) subproblem (n) subproblem (n)

5 Does D&C fit for statistical analysis? Sometimes it does, but sometimes it doesn t... In the following, we give two simple examples.

6 Example 1: sample mean Problem(N) subproblem (n) subproblem (n) subproblem (n) subproblem (n) X n,1 X n,2 X n,3 X n, s=1 It fits! X s = 1 4n 4 s=1 i=1 n X is = 1 N N X i = X N. i=1

7 Example 2: sample median Problem(N) subproblem (n) subproblem (n) X 1 (0.5) X 2 (0.5) subproblem (n) X 3 (0.5) subproblem (n) X 4 (0.5) X(0.5) s = the middle value of ordered n samples in s group; X (0.5) = overall median s=1 X s (0.5)?? = X (0.5)

8 Example 2: sample median Simulation 1: X i N(0, 1); Simulation 2: X i Exponential(1). N = True median v.s. simulated S 1 S s=1 Xs (0.5) τ = 0.5, N (0,1) log N(S) τ = 0.5, EXP (1) log N(S)

9 Example 2: sample median Simulation 1: X i N(0, 1); Simulation 2: X i Exponential(1). N = True median v.s. simulated S 1 S s=1 Xs (0.5) τ = 0.5, N (0,1) log N(S) τ = 0.5, EXP (1) log N(S)

10 Major issues When does the D&C algorithm work? Especially for skewed and heavy tail distribution Statistical inference Asymptotic distribution Inference for the whole distribution? Take advantage of massive size to discover subtle patterns hidden in the whole distribution

11 Outline 1 Quantile regression 2 Two-step algorithm: D&C and quantile projection 3 Oracle rules: linear model and nonparametric model 4 Simulation

12 Quantile Response Y, predictors X. For τ (0, 1), conditional quantile curve Q( ; τ) of Y R conditional on X is defined through P (Y Q(X; τ) X = x) = τ x. Q(x; τ) at τ = 0.1, 0.25, 0.5, 0.75, 0.9 under different models Y = 0.5*X + N(0,1) x y Y = X + (.5+X)*N(0,1) x y Y X=x ~ Beta(1.5 x,.5+x) x y

13 Quantile regression v.s. mean regression Mean regression: Y i = m(x i ) + ε i, E[ε X = x] = 0 m: Regression function, object of interest. ε i : errors. Quantile regression: P (Y Q(x; τ) X = x) = τ No strict distinction between signal and noise. Object of interest: properties of conditional distribution of Y X = x. Contains much richer information than just conditional mean Age Bone Mass Density (BMD) Age Bone Mass Density (BMD)

14 Quantile curves v.s. conditional distribution Q(x 0 ; τ) = F 1 (τ x 0 ), where F Y X (y x) is the conditional distribution function of Y given X BMD Q^( Age x 0 = 13 ; τ) Q^( Age x 0 = 18 ; τ) Q^( Age x 0 = 23 ; τ) CDF F^(y Age x 0 = 13) F^(y Age x 0 = 18) F^(y Age x 0 = 23) τ BMD

15 Quantile Regression as Optimization {(X i, Y i )} N i=1 independent and identical samples in Rd+1 Koenker and Bassett (1978): if Q(x; τ) = β(τ) x, estimate by β or (τ) := arg min b N ρ τ (Y i b X i ) (1.1) i=1 where ρ τ (u) := τu + + (1 τ)u check function. Optimization problem (1.2) is convex (but non-smooth), which can be solved by linear programming Q(x 0 ; τ) := x 0 β or (τ) for any x 0 More generally, we can consider a series approximation model

16 A unified framework: Q(x; τ) Z(x) β(τ) m := dim(z) (it is possible that m ). Solve β or (τ) := arg min b N { ρ τ Yi b Z(X i ) } (1.2) i=1 Examples of Z(x): linear model with fixed/increasing dimension, B-splines, polynomials, trigonometric polynomials Q(x; τ) := Z(x) βor (τ) Need to control the bias Q(x; τ) Z(x) β(τ) Quantile Regression

17 Summary for quantile regression β or (τ) is computationally infeasible when N is so large that cannot be handled with a single machine To infer the whole conditional distribution for fixed x 0, by F Y X (y x 0 ) = 1 0 1{Q(x 0 ; τ) < y}dτ where 1(A) = 1 if A is true. To approximate the integral, we need a to compute Q(x 0 ; τ) for many τ

18 D&C algorithm at fixed τ Problem(N) subproblem (n) subproblem (n) β 1 (τ 1 ) β 2 (τ 1 ) subproblem (n) β 3 (τ 1 ) subproblem (n) β 4 (τ 1 ) β s (τ) := arg min b R m n { ρ τ Yis b Z(X is ) } (2.1) i=1 β(τ) := 1 S S β s (τ). (2.2) s=1 However, this is only for a fixed τ!

19 Quantile projection algorithm To avoid repetitively applying D&C, take B := (B 1,..., B q ) B-spline basis defined on equidistant knots {t 1,..., t G } T with degree r τ N, β(τ) := Ξ B(τ). (2.3) Computation of Ξ: 1 Define a grid of quantile levels {τ 1,..., τ K } on [τ L, τ U ], K > q. For each τ k, compute β(τ k ) 2 Compute for j = 1,..., m α j := arg min α R q K ( βj (τ k ) α B(τ k ) ) 2. (2.4) k=1 3 Set the matrix Ξ := [ α 1 α 2... α m ].

20 Computation of F Y X (y x) Define Q(x; τ) := Z(x) β(τ) = Z(x) Ξ B(τ). Compute τu F Y X (y x 0 ) := τ L + 1{ Q(x 0 ; τ) < y}dτ. (2.5) τ L where 0 < τ L < τ U < 1.

21 Remarks Both S and K can grow with N, and they decide the computational limit as well as the statistical performance Find S and K such that β(τ) and β(τ) are close to β or (τ) in some statistical sense This results in sharp upper bound of S and lower bound of K which impose computational limit The two-step procedure requires only one pass through the entire data Quantile projection requires only {β(τ 1 ),..., β(τ K )} of size m K, without the need to access the raw data set

22 Oracle rules For T = [τ L, τ U ] (0, 1), under regularity conditions, C., Volgushev and Cheng (2016): Chao et al. (2016) a N u ( βor (τ) β(τ) ) N at any fixed τ T (3.1) a N u ( βor (τ) β(τ) ) G(τ) as a process in τ T (3.2) where N is a centered normal distribution and G is a centered Gaussian process The oracle rule holds for β(τ) and β(τ) if β(τ) satisfies (3.1) and β(τ) satisfies (3.2) Inference follows from the oracle rule

23 Two leading models Linear model: m = dim(z(x)) is fixed, and Q(x; τ) = Z(x) β(τ); Univariate spline nonparametric model: m = dim(z(x)) with N. c N (γ N ) := Q(x; τ) Z(x) γ N (τ) 0, γ N (τ) := arg min γ R m E[ (Z γ Q(X; τ)) 2 f(q(x; τ) X) ]. (3.3) Oracle rule: spline model γ N can be defined in many other ways. We do not go into detail here Conditions imposed throughout the talk: J m (τ) := E[ZZ f Y X (Q(X; τ) X)] Assumption (A)

24 Linear model with fixed dimension P 1 (ξ m, M, f, f, f min ): all distributions satisfying (A1)-(A3) with some constants 0 < ξ m, M, f, f < and f min > 0 Theorem 3.1 (Oracle Rule for β(τ)) Suppose that S = o(n 1/2 (log N) 2 ). τ T and u R m, Nu (β(τ) β(τ)) (u J m (τ) 1 E[Z(X)Z(X) ]J m (τ) 1 u) 1/2 N ( 0, τ(1 τ) ), If S N 1/2, then the weak convergence result above fails for some P P 1 (1, d, f, f, f min ). S = o(n 1/2 (log N) 2 ) is sharp: only miss by a factor of (log N) 2

25 Λ ητ c (T ): Hölder class with smoothness η τ > 0. l (T ): set of all uniformly bounded real functions on T Theorem 3.2 (Oracle Rule for β( )) Suppose that τ Q(x 0 ; τ) Λ ητ c (T ) for a x 0 X. If S = o(n 1/2 (log N) 2 ), and K G N 1/(2ητ ) with r τ η τ then the projection estimator β(τ) defined in (2.3) satisfies N ( Z(x0 ) β( ) Q(x0 ; ) ) G( ) in l (T ), (3.4) where G is a centered Gaussian process. Covariance function If G N 1/(2ητ ) the weak convergence in (3.4) fails for some P P 1 (1, d, f, f, f min ) with τ Q(x 0 ; τ) Λ ητ c (T ). K N 1/(2ητ ) is sharp!

26 Splines nonparametric model We require an additional condition: (L) For each x X, the vector Z(x) has zeroes in all but at most r consecutive entries, where r is fixed. Furthermore, sup x X Ẽ(Z(x), γ N ) = O(1). which guarantees that the matrix J m (τ) to be a block matrix P L (Z, M, f, f, f min, R) := { all sequences of distributions of (X, Y ) on R d+1 satisfying (A1)-(A3) with M, f, f <, f min > 0, and (L) for some r < R; m 2 (log N) 6 = o(n), c N (γ N ) 2 = o(n 1/2 ) }. (3.5) Definition c N

27 Splines nonparametric model Theorem 3.3 (Oracle rule for β(τ)) Let {(X i, Y i )} N i=1 be distributed P N P L (Z, M, f, f, f min, R), m N ζ for some ζ > 0. S = o((nm 1 (log N) 4 ) 1/2 ) then τ T, x 0 X, NZ(x0 ) (β(τ) γ N (τ)) (Z(x 0 ) J m (τ) 1 E[ZZ ]J m (τ) 1 Z(x 0 )) 1/2 N ( 0, τ(1 τ) ), If S (N/m) 1/2 the weak convergence above fails for some P N P L (Z, M, f, f, f min, R) Sharpness of S = o((n/m) 1/2 (log N) 2 ): nonparametric rate is slower, again miss by a factor of (log N) 2

28 Theorem 3.4 (Oracle rule for β(τ)) Let the assumptions of Theorem 3.3 hold and assume that τ Q(x 0 ; τ) Λ ητ c (T ), r τ η τ, Suppose that c N (γ N ) = o(n 1/2 Z(x 0 ) ) and K G N 1/(2ητ ) Z(x 0 ) 1/ητ, and the limit H(τ 1, τ 2 ) := lim N K N (τ 1, τ 2 ) > 0 K N τ 1, τ 2 T. N ( Z(x0 ) β( ) Q(x0 ; ) ) G( ) in l (T ), Z(x 0 ) where G is a centered Gaussian process with covariance structure E[G(τ)G(τ )] = H(τ, τ ). If G N 1/(2ητ ) Z(x 0 ) 1/ητ the weak convergence fails for some τ Q(x 0 ; τ) Λ ητ c (T ) for all S.

29 Distribution function τu F Y or X ( x 0) := τ L + 1{Z(x) βor (τ) < y}dτ τ L The oracle rule holds for both models. Corollary 3.5 Under the same conditions as Theorem 3.2 (linear model) or Theorem 3.4 (nonparametric spline model), we have for any x 0 X, N ( FY X ( x 0 ) F Y X ( x 0 ) ) f Y X ( x 0 )G ( F Y X ( x 0 ) ), in l ( (Q(x 0 ; τ L ), Q(x 0 ; τ U )) ), where G is a centered Gaussian process defined in respective theorem. The same holds for F or Y X ( x 0).

30 Phase transitions K?? N 1/(2ητ ) ( N m) 1/(2ητ ) ( N 1/2 1/2 N m 1/2 log 2 N m) N 1/2 log 2 N N 1/2 S Figure: Regions (S, K) for the oracle rule of linear model and spline nonparametric model.? region is the discrepancy between the sufficient and necessary conditions.

31 Setting: linear model Model: Y i = β m 1 X i + ε i, m = 4, 16, 32 X i follows a multivariate uniform distribution U([0, 1] m 1 ) with covariance matrix Σ X := E[X i X i ] with Σ jk = j k for j, k = 1,..., m 1 ε N or ε EXP (skewed) x 0 = (1, (m 1) 1/2 l m 1 ) From Theorem 3.1, the coverage probability of the α = 95% confidence interval is P { x 0 β(τ) [ x 0 β(τ) ± N 1/2 fε,τ 1 τ(1 τ)x 0 Σ 1 X x 0Φ 1 (1 α/2) ]} S : the point of S where the coverage starts to drop Additional information

32 β(τ), ε N (0, ) Normal(0, ), τ = 0.1 Normal(0, ), τ = 0.5 Coverage Probability N = 2 14, m = 4 N = 2 22, m = 4 N = 2 14, m = 16 N = 2 22, m = 16 N = 2 14, m = 32 N = 2 22, m = 32 Coverage Probability N = 2 14, m = 4 N = 2 22, m = 4 N = 2 14, m = 16 N = 2 22, m = 16 N = 2 14, m = 32 N = 2 22, m = log N(S) log N(S) Normal(0, ), τ = 0.9 Coverage Probability N = 2 14, m = 4 N = 2 22, m = 4 N = 2 14, m = 16 N = 2 22, m = 16 N = 2 14, m = 32 N = 2 22, m = log N(S)

33 β(τ), ε Exp(0.8) Exp (0.8), τ = 0.1 Exp (0.8), τ = 0.5 Coverage Probability N = 2 14, m = 4 N = 2 22, m = 4 N = 2 14, m = 16 N = 2 22, m = 16 N = 2 14, m = 32 N = 2 22, m = 32 Coverage Probability N = 2 14, m = 4 N = 2 22, m = 4 N = 2 14, m = 16 N = 2 22, m = 16 N = 2 14, m = 32 N = 2 22, m = log N(S) log N(S) Exp (0.8), τ = 0.9 Coverage Probability N = 2 14, m = 4 N = 2 22, m = 4 N = 2 14, m = 16 N = 2 22, m = 16 N = 2 14, m = 32 N = 2 22, m = log N(S)

34 Summary for simulation of β(τ) m increases, S decreases N increases, S gets close to N 1/2 ε N, coverage is symmetric in τ, ε Exp, coverage is asymmetric in τ t distribution behaves similarly to normal distribution

35 Quantile projection setting B: cubic B-spline with q = dim(b) defined on G = 4 + q equidistant knots on [τ L, τ U ]. We require K > q so that β(τ) is computable (see (2.4)) N = 2 14 y 0 = Q(x 0 ; τ) so that F Y X (y 0 x 0 ) = τ From Theorem 3.2, the coverage probability with size α = 0.95 is P { τ [ FY X (Q(x 0 ; τ) x 0 ) ± N 1/2 τ(1 τ)x 0 Σ 1 X x 0Φ 1 (1 α/2) ]}

36 F Y X (y x), ε N (0, ), m = 4 for β(τ) Normal(0, ), m= 4, y 0 = Q( x 0 ; τ = 0.1 ) Normal(0, ), m= 4, y 0 = Q( x 0 ; τ = 0.5 ) Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = 30 Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = log N(S) log N(S) Normal(0, ), m= 4, y 0 = Q( x 0 ; τ = 0.9 ) Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = log N(S)

37 F Y X (y x), ε N (0, ), m = 32 for β(τ) Normal(0, ), m= 32, y 0 = Q( x 0 ; τ = 0.1 ) Normal(0, ), m= 32, y 0 = Q( x 0 ; τ = 0.5 ) Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = 30 Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = log N(S) log N(S) Normal(0, ), m= 32, y 0 = Q( x 0 ; τ = 0.9 ) Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = log N(S)

38 F Y X (y x), ε Exp(0.8), m = 4 for β(τ) Exp (0.8), m= 4, y 0 = Q( x 0 ; τ = 0.1 ) Exp (0.8), m= 4, y 0 = Q( x 0 ; τ = 0.5 ) Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = 30 Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = log N(S) log N(S) Exp (0.8), m= 4, y 0 = Q( x 0 ; τ = 0.9 ) Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = log N(S)

39 F Y X (y x), ε Exp(0.8), m = 32 for β(τ) Exp (0.8), m= 32, y 0 = Q( x 0 ; τ = 0.1 ) Exp (0.8), m= 32, y 0 = Q( x 0 ; τ = 0.5 ) Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = 30 Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = log N(S) log N(S) Exp (0.8), m= 32, y 0 = Q( x 0 ; τ = 0.9 ) Coverage Probability q = 7, K = 20 q = 7, K = 30 q = 10, K = 20 q = 10, K = 30 q = 15, K = 20 q = 15, K = log N(S)

40 Summary for simulation of F Y X (y x) Increase in m lowers both S and q (the critical point in q for the oracle rule) Either S > S or q < q (q = G 4) leads to the collapse of the oracle rule Increase in q and K improves the coverage probability For ε N, coverage is no longer symmetric in τ

41 Thank you for your attention

42 References I Chao, S., Volgushev, S., and Cheng, G. (2016). Quantile process for semi and nonparametric regression models. ArXiv Preprint Arxiv White, T. (2012). Hadoop: The Definitive Guide. O Reilly Media / Yahoo Press.

43 Assumption (A): data (X i, Y i ) i=1,...,n form triangular array and are row-wise i.i.d. with Linear models (A1) Assume that Z i ξ m <, where ξ m = O(N b ) is allowed to increase to infinity, and that 1/M λ min (E[ZZ T ]) λ max (E[ZZ T ]) M holds uniformly in n for some fixed constant M. (A2) The conditional distribution F Y X (y x) is twice differentiable w.r.t. y. Denote the corresponding derivatives by f Y X (y x) and f Y X (y x). Assume that f := sup y,x uniformly in n. (A3) Assume that uniformly in n. f Y X (y x) <, f := sup f Y X (y x) < y,x 0 < f min inf inf f Y X(Q(x; τ) x) τ T x

44 E [ G(τ)G(τ ) ] (5.1) = Z(x 0 ) J m (τ) 1 E [ Z(X)Z(X) ] J m (τ ) 1 Z(x 0 )(τ τ ττ ). Process: linear model K(τ 1, τ 2 ) := Z(x 0 ) 2 Z(x 0 ) J 1 m (τ 1 )E[ZZ ]J 1 m (τ 2 )Z(x 0 )(τ 1 τ 2 τ 1 τ 2 ) Process: local polynomial

45 β 3 = (0.21, 0.89, 0.38) ; β 15 = (β 3, 0.63, 0.11, 1.01, 1.79, 1.39, β 31 = (β 15, 0.21, β 15). β(τ) = ( Φ 1 σ=0.1 (τ), β m 1 ) 0.52, 1.62, 1.26, 0.72, 0.43, 0.41, 0.02) ; (5.2) f ε,τ > 0 is the height of the density of ε i evaluated at ε i s τ quantile. Simulation setting

Quantile Processes for Semi and Nonparametric Regression

Quantile Processes for Semi and Nonparametric Regression Quantile Processes for Semi and Nonparametric Regression Shih-Kang Chao Department of Statistics Purdue University IMS-APRM 2016 A joint work with Stanislav Volgushev and Guang Cheng Quantile Response

More information

Semi-Nonparametric Inferences for Massive Data

Semi-Nonparametric Inferences for Massive Data Semi-Nonparametric Inferences for Massive Data Guang Cheng 1 Department of Statistics Purdue University Statistics Seminar at NCSU October, 2015 1 Acknowledge NSF, Simons Foundation and ONR. A Joint Work

More information

Inference on distributions and quantiles using a finite-sample Dirichlet process

Inference on distributions and quantiles using a finite-sample Dirichlet process Dirichlet IDEAL Theory/methods Simulations Inference on distributions and quantiles using a finite-sample Dirichlet process David M. Kaplan University of Missouri Matt Goldman UC San Diego Midwest Econometrics

More information

Sample Size Requirement For Some Low-Dimensional Estimation Problems

Sample Size Requirement For Some Low-Dimensional Estimation Problems Sample Size Requirement For Some Low-Dimensional Estimation Problems Cun-Hui Zhang, Rutgers University September 10, 2013 SAMSI Thanks for the invitation! Acknowledgements/References Sun, T. and Zhang,

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

Bayesian Aggregation for Extraordinarily Large Dataset

Bayesian Aggregation for Extraordinarily Large Dataset Bayesian Aggregation for Extraordinarily Large Dataset Guang Cheng 1 Department of Statistics Purdue University www.science.purdue.edu/bigdata Department Seminar Statistics@LSE May 19, 2017 1 A Joint Work

More information

Nonparametric Inference via Bootstrapping the Debiased Estimator

Nonparametric Inference via Bootstrapping the Debiased Estimator Nonparametric Inference via Bootstrapping the Debiased Estimator Yen-Chi Chen Department of Statistics, University of Washington ICSA-Canada Chapter Symposium 2017 1 / 21 Problem Setup Let X 1,, X n be

More information

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

Nonparametric Inference In Functional Data

Nonparametric Inference In Functional Data Nonparametric Inference In Functional Data Zuofeng Shang Purdue University Joint work with Guang Cheng from Purdue Univ. An Example Consider the functional linear model: Y = α + where 1 0 X(t)β(t)dt +

More information

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider the problem of

More information

Bootstrapping high dimensional vector: interplay between dependence and dimensionality

Bootstrapping high dimensional vector: interplay between dependence and dimensionality Bootstrapping high dimensional vector: interplay between dependence and dimensionality Xianyang Zhang Joint work with Guang Cheng University of Missouri-Columbia LDHD: Transition Workshop, 2014 Xianyang

More information

Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood

Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood Kuangyu Wen & Ximing Wu Texas A&M University Info-Metrics Institute Conference: Recent Innovations in Info-Metrics October

More information

Can we do statistical inference in a non-asymptotic way? 1

Can we do statistical inference in a non-asymptotic way? 1 Can we do statistical inference in a non-asymptotic way? 1 Guang Cheng 2 Statistics@Purdue www.science.purdue.edu/bigdata/ ONR Review Meeting@Duke Oct 11, 2017 1 Acknowledge NSF, ONR and Simons Foundation.

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University JSM, 2015 E. Christou, M. G. Akritas (PSU) SIQR JSM, 2015

More information

Empirical Bayes Quantile-Prediction aka E-B Prediction under Check-loss;

Empirical Bayes Quantile-Prediction aka E-B Prediction under Check-loss; BFF4, May 2, 2017 Empirical Bayes Quantile-Prediction aka E-B Prediction under Check-loss; Lawrence D. Brown Wharton School, Univ. of Pennsylvania Joint work with Gourab Mukherjee and Paat Rusmevichientong

More information

IV Quantile Regression for Group-level Treatments, with an Application to the Distributional Effects of Trade

IV Quantile Regression for Group-level Treatments, with an Application to the Distributional Effects of Trade IV Quantile Regression for Group-level Treatments, with an Application to the Distributional Effects of Trade Denis Chetverikov Brad Larsen Christopher Palmer UCLA, Stanford and NBER, UC Berkeley September

More information

SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES. Liping Zhu, Mian Huang, & Runze Li. The Pennsylvania State University

SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES. Liping Zhu, Mian Huang, & Runze Li. The Pennsylvania State University SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES Liping Zhu, Mian Huang, & Runze Li The Pennsylvania State University Technical Report Series #10-104 College of Health and Human Development

More information

Estimation of cumulative distribution function with spline functions

Estimation of cumulative distribution function with spline functions INTERNATIONAL JOURNAL OF ECONOMICS AND STATISTICS Volume 5, 017 Estimation of cumulative distribution function with functions Akhlitdin Nizamitdinov, Aladdin Shamilov Abstract The estimation of the cumulative

More information

Estimation of the functional Weibull-tail coefficient

Estimation of the functional Weibull-tail coefficient 1/ 29 Estimation of the functional Weibull-tail coefficient Stéphane Girard Inria Grenoble Rhône-Alpes & LJK, France http://mistis.inrialpes.fr/people/girard/ June 2016 joint work with Laurent Gardes,

More information

Bayesian spatial quantile regression

Bayesian spatial quantile regression Brian J. Reich and Montserrat Fuentes North Carolina State University and David B. Dunson Duke University E-mail:reich@stat.ncsu.edu Tropospheric ozone Tropospheric ozone has been linked with several adverse

More information

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage Lingrui Gan, Naveen N. Narisetty, Feng Liang Department of Statistics University of Illinois at Urbana-Champaign Problem Statement

More information

Asymmetric least squares estimation and testing

Asymmetric least squares estimation and testing Asymmetric least squares estimation and testing Whitney Newey and James Powell Princeton University and University of Wisconsin-Madison January 27, 2012 Outline ALS estimators Large sample properties Asymptotic

More information

Inference For High Dimensional M-estimates: Fixed Design Results

Inference For High Dimensional M-estimates: Fixed Design Results Inference For High Dimensional M-estimates: Fixed Design Results Lihua Lei, Peter Bickel and Noureddine El Karoui Department of Statistics, UC Berkeley Berkeley-Stanford Econometrics Jamboree, 2017 1/49

More information

41903: Introduction to Nonparametrics

41903: Introduction to Nonparametrics 41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific

More information

INFERENCE APPROACHES FOR INSTRUMENTAL VARIABLE QUANTILE REGRESSION. 1. Introduction

INFERENCE APPROACHES FOR INSTRUMENTAL VARIABLE QUANTILE REGRESSION. 1. Introduction INFERENCE APPROACHES FOR INSTRUMENTAL VARIABLE QUANTILE REGRESSION VICTOR CHERNOZHUKOV CHRISTIAN HANSEN MICHAEL JANSSON Abstract. We consider asymptotic and finite-sample confidence bounds in instrumental

More information

Cross-Validation with Confidence

Cross-Validation with Confidence Cross-Validation with Confidence Jing Lei Department of Statistics, Carnegie Mellon University UMN Statistics Seminar, Mar 30, 2017 Overview Parameter est. Model selection Point est. MLE, M-est.,... Cross-validation

More information

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49 State-space Model Eduardo Rossi University of Pavia November 2013 Rossi State-space Model Financial Econometrics - 2013 1 / 49 Outline 1 Introduction 2 The Kalman filter 3 Forecast errors 4 State smoothing

More information

Bayesian Modeling of Conditional Distributions

Bayesian Modeling of Conditional Distributions Bayesian Modeling of Conditional Distributions John Geweke University of Iowa Indiana University Department of Economics February 27, 2007 Outline Motivation Model description Methods of inference Earnings

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

A Resampling Method on Pivotal Estimating Functions

A Resampling Method on Pivotal Estimating Functions A Resampling Method on Pivotal Estimating Functions Kun Nie Biostat 277,Winter 2004 March 17, 2004 Outline Introduction A General Resampling Method Examples - Quantile Regression -Rank Regression -Simulation

More information

Modeling Real Estate Data using Quantile Regression

Modeling Real Estate Data using Quantile Regression Modeling Real Estate Data using Semiparametric Quantile Regression Department of Statistics University of Innsbruck September 9th, 2011 Overview 1 Application: 2 3 4 Hedonic regression data for house prices

More information

Uncertainty Quantification for Inverse Problems. November 7, 2011

Uncertainty Quantification for Inverse Problems. November 7, 2011 Uncertainty Quantification for Inverse Problems November 7, 2011 Outline UQ and inverse problems Review: least-squares Review: Gaussian Bayesian linear model Parametric reductions for IP Bias, variance

More information

A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging

A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging Cheng Li DSAP, National University of Singapore Joint work with Rajarshi Guhaniyogi (UC Santa Cruz), Terrance D. Savitsky (US Bureau of Labor

More information

Nonparametric Heterogeneity Testing for Massive Data

Nonparametric Heterogeneity Testing for Massive Data Nonparametric Heterogeneity Testing for Massive Data Abstract A massive dataset often consists of a growing number of potentially heterogeneous sub-populations. This paper is concerned about testing various

More information

The MNet Estimator. Patrick Breheny. Department of Biostatistics Department of Statistics University of Kentucky. August 2, 2010

The MNet Estimator. Patrick Breheny. Department of Biostatistics Department of Statistics University of Kentucky. August 2, 2010 Department of Biostatistics Department of Statistics University of Kentucky August 2, 2010 Joint work with Jian Huang, Shuangge Ma, and Cun-Hui Zhang Penalized regression methods Penalized methods have

More information

Beyond Mean Regression

Beyond Mean Regression Beyond Mean Regression Thomas Kneib Lehrstuhl für Statistik Georg-August-Universität Göttingen 8.3.2013 Innsbruck Introduction Introduction One of the top ten reasons to become statistician (according

More information

STAT 461/561- Assignments, Year 2015

STAT 461/561- Assignments, Year 2015 STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and

More information

Statistical inference on Lévy processes

Statistical inference on Lévy processes Alberto Coca Cabrero University of Cambridge - CCA Supervisors: Dr. Richard Nickl and Professor L.C.G.Rogers Funded by Fundación Mutua Madrileña and EPSRC MASDOC/CCA student workshop 2013 26th March Outline

More information

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis. 401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis

More information

Gaussian Process Regression

Gaussian Process Regression Gaussian Process Regression 4F1 Pattern Recognition, 21 Carl Edward Rasmussen Department of Engineering, University of Cambridge November 11th - 16th, 21 Rasmussen (Engineering, Cambridge) Gaussian Process

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57

More information

Estimation de mesures de risques à partir des L p -quantiles

Estimation de mesures de risques à partir des L p -quantiles 1/ 42 Estimation de mesures de risques à partir des L p -quantiles extrêmes Stéphane GIRARD (Inria Grenoble Rhône-Alpes) collaboration avec Abdelaati DAOUIA (Toulouse School of Economics), & Gilles STUPFLER

More information

Expecting the Unexpected: Uniform Quantile Regression Bands with an application to Investor Sentiments

Expecting the Unexpected: Uniform Quantile Regression Bands with an application to Investor Sentiments Expecting the Unexpected: Uniform Bands with an application to Investor Sentiments Boston University November 16, 2016 Econometric Analysis of Heterogeneity in Financial Markets Using s Chapter 1: Expecting

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

Comprehensive Examination Quantitative Methods Spring, 2018

Comprehensive Examination Quantitative Methods Spring, 2018 Comprehensive Examination Quantitative Methods Spring, 2018 Instruction: This exam consists of three parts. You are required to answer all the questions in all the parts. 1 Grading policy: 1. Each part

More information

A short introduction to INLA and R-INLA

A short introduction to INLA and R-INLA A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk

More information

A Conditional Approach to Modeling Multivariate Extremes

A Conditional Approach to Modeling Multivariate Extremes A Approach to ing Multivariate Extremes By Heffernan & Tawn Department of Statistics Purdue University s April 30, 2014 Outline s s Multivariate Extremes s A central aim of multivariate extremes is trying

More information

Nonparametric estimation of extreme risks from heavy-tailed distributions

Nonparametric estimation of extreme risks from heavy-tailed distributions Nonparametric estimation of extreme risks from heavy-tailed distributions Laurent GARDES joint work with Jonathan EL METHNI & Stéphane GIRARD December 2013 1 Introduction to risk measures 2 Introduction

More information

Divide-and-combine Strategies in Statistical Modeling for Massive Data

Divide-and-combine Strategies in Statistical Modeling for Massive Data Divide-and-combine Strategies in Statistical Modeling for Massive Data Liqun Yu Washington University in St. Louis March 30, 2017 Liqun Yu (WUSTL) D&C Statistical Modeling for Massive Data March 30, 2017

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

For right censored data with Y i = T i C i and censoring indicator, δ i = I(T i < C i ), arising from such a parametric model we have the likelihood,

For right censored data with Y i = T i C i and censoring indicator, δ i = I(T i < C i ), arising from such a parametric model we have the likelihood, A NOTE ON LAPLACE REGRESSION WITH CENSORED DATA ROGER KOENKER Abstract. The Laplace likelihood method for estimating linear conditional quantile functions with right censored data proposed by Bottai and

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

Independent Component (IC) Models: New Extensions of the Multinormal Model

Independent Component (IC) Models: New Extensions of the Multinormal Model Independent Component (IC) Models: New Extensions of the Multinormal Model Davy Paindaveine (joint with Klaus Nordhausen, Hannu Oja, and Sara Taskinen) School of Public Health, ULB, April 2008 My research

More information

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Abhirup Datta 1 Sudipto Banerjee 1 Andrew O. Finley 2 Alan E. Gelfand 3 1 University of Minnesota, Minneapolis,

More information

Nonparametric regression with martingale increment errors

Nonparametric regression with martingale increment errors S. Gaïffas (LSTA - Paris 6) joint work with S. Delattre (LPMA - Paris 7) work in progress Motivations Some facts: Theoretical study of statistical algorithms requires stationary and ergodicity. Concentration

More information

Additive Isotonic Regression

Additive Isotonic Regression Additive Isotonic Regression Enno Mammen and Kyusang Yu 11. July 2006 INTRODUCTION: We have i.i.d. random vectors (Y 1, X 1 ),..., (Y n, X n ) with X i = (X1 i,..., X d i ) and we consider the additive

More information

Cross-Validation with Confidence

Cross-Validation with Confidence Cross-Validation with Confidence Jing Lei Department of Statistics, Carnegie Mellon University WHOA-PSI Workshop, St Louis, 2017 Quotes from Day 1 and Day 2 Good model or pure model? Occam s razor We really

More information

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun

More information

Stat 710: Mathematical Statistics Lecture 31

Stat 710: Mathematical Statistics Lecture 31 Stat 710: Mathematical Statistics Lecture 31 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 31 April 13, 2009 1 / 13 Lecture 31:

More information

Sliced Inverse Regression

Sliced Inverse Regression Sliced Inverse Regression Ge Zhao gzz13@psu.edu Department of Statistics The Pennsylvania State University Outline Background of Sliced Inverse Regression (SIR) Dimension Reduction Definition of SIR Inversed

More information

arxiv: v1 [cs.it] 21 Feb 2013

arxiv: v1 [cs.it] 21 Feb 2013 q-ary Compressive Sensing arxiv:30.568v [cs.it] Feb 03 Youssef Mroueh,, Lorenzo Rosasco, CBCL, CSAIL, Massachusetts Institute of Technology LCSL, Istituto Italiano di Tecnologia and IIT@MIT lab, Istituto

More information

ESTIMATION AND OUTPUT ANALYSIS (L&K Chapters 9, 10) Review performance measures (means, probabilities and quantiles).

ESTIMATION AND OUTPUT ANALYSIS (L&K Chapters 9, 10) Review performance measures (means, probabilities and quantiles). ESTIMATION AND OUTPUT ANALYSIS (L&K Chapters 9, 10) Set up standard example and notation. Review performance measures (means, probabilities and quantiles). A framework for conducting simulation experiments

More information

On detection of unit roots generalizing the classic Dickey-Fuller approach

On detection of unit roots generalizing the classic Dickey-Fuller approach On detection of unit roots generalizing the classic Dickey-Fuller approach A. Steland Ruhr-Universität Bochum Fakultät für Mathematik Building NA 3/71 D-4478 Bochum, Germany February 18, 25 1 Abstract

More information

sparse and low-rank tensor recovery Cubic-Sketching

sparse and low-rank tensor recovery Cubic-Sketching Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru

More information

Solving Corrupted Quadratic Equations, Provably

Solving Corrupted Quadratic Equations, Provably Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin

More information

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria Weak Leiden University Amsterdam, 13 November 2013 Outline 1 2 3 4 5 6 7 Definition Definition Let µ, µ 1, µ 2,... be probability measures on (R, B). It is said that µ n converges weakly to µ, and we then

More information

Sensitivity to Missing Data Assumptions: Theory and An Evaluation of the U.S. Wage Structure. September 21, 2012

Sensitivity to Missing Data Assumptions: Theory and An Evaluation of the U.S. Wage Structure. September 21, 2012 Sensitivity to Missing Data Assumptions: Theory and An Evaluation of the U.S. Wage Structure Andres Santos UC San Diego September 21, 2012 The Problem Missing data is ubiquitous in modern economic research

More information

Semi and Nonparametric Models in Econometrics

Semi and Nonparametric Models in Econometrics Semi and Nonparametric Models in Econometrics Part 4: partial identification Xavier d Haultfoeuille CREST-INSEE Outline Introduction First examples: missing data Second example: incomplete models Inference

More information

Multivariate Bayesian Linear Regression MLAI Lecture 11

Multivariate Bayesian Linear Regression MLAI Lecture 11 Multivariate Bayesian Linear Regression MLAI Lecture 11 Neil D. Lawrence Department of Computer Science Sheffield University 21st October 2012 Outline Univariate Bayesian Linear Regression Multivariate

More information

Quantile methods. Class Notes Manuel Arellano December 1, Let F (r) =Pr(Y r). Forτ (0, 1), theτth population quantile of Y is defined to be

Quantile methods. Class Notes Manuel Arellano December 1, Let F (r) =Pr(Y r). Forτ (0, 1), theτth population quantile of Y is defined to be Quantile methods Class Notes Manuel Arellano December 1, 2009 1 Unconditional quantiles Let F (r) =Pr(Y r). Forτ (0, 1), theτth population quantile of Y is defined to be Q τ (Y ) q τ F 1 (τ) =inf{r : F

More information

On Model Selection Consistency of Lasso

On Model Selection Consistency of Lasso On Model Selection Consistency of Lasso Peng Zhao Department of Statistics University of Berkeley 367 Evans Hall Berkeley, CA 94720-3860, USA Bin Yu Department of Statistics University of Berkeley 367

More information

Program Evaluation with High-Dimensional Data

Program Evaluation with High-Dimensional Data Program Evaluation with High-Dimensional Data Alexandre Belloni Duke Victor Chernozhukov MIT Iván Fernández-Val BU Christian Hansen Booth ESWC 215 August 17, 215 Introduction Goal is to perform inference

More information

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Vadim Marmer University of British Columbia Artyom Shneyerov CIRANO, CIREQ, and Concordia University August 30, 2010 Abstract

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

19.1 Problem setup: Sparse linear regression

19.1 Problem setup: Sparse linear regression ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 19: Minimax rates for sparse linear regression Lecturer: Yihong Wu Scribe: Subhadeep Paul, April 13/14, 2016 In

More information

Smooth nonparametric estimation of a quantile function under right censoring using beta kernels

Smooth nonparametric estimation of a quantile function under right censoring using beta kernels Smooth nonparametric estimation of a quantile function under right censoring using beta kernels Chanseok Park 1 Department of Mathematical Sciences, Clemson University, Clemson, SC 29634 Short Title: Smooth

More information

Nonparametric Conditional Density Estimation Using Piecewise-Linear Solution Path of Kernel Quantile Regression

Nonparametric Conditional Density Estimation Using Piecewise-Linear Solution Path of Kernel Quantile Regression Nonparametric Conditional Density Estimation Using Piecewise-Linear Solution Path of Kernel Quantile Regression Ichiro Takeuchi Division of Computer Science, Graduate School of Engineering, Mie University

More information

Chapter 4: Interpolation and Approximation. October 28, 2005

Chapter 4: Interpolation and Approximation. October 28, 2005 Chapter 4: Interpolation and Approximation October 28, 2005 Outline 1 2.4 Linear Interpolation 2 4.1 Lagrange Interpolation 3 4.2 Newton Interpolation and Divided Differences 4 4.3 Interpolation Error

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53 State-space Model Eduardo Rossi University of Pavia November 2014 Rossi State-space Model Fin. Econometrics - 2014 1 / 53 Outline 1 Motivation 2 Introduction 3 The Kalman filter 4 Forecast errors 5 State

More information

ECAS Summer Course. Fundamentals of Quantile Regression Roger Koenker University of Illinois at Urbana-Champaign

ECAS Summer Course. Fundamentals of Quantile Regression Roger Koenker University of Illinois at Urbana-Champaign ECAS Summer Course 1 Fundamentals of Quantile Regression Roger Koenker University of Illinois at Urbana-Champaign La Roche-en-Ardennes: September, 2005 An Outline 2 Beyond Average Treatment Effects Computation

More information

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the

More information

The largest eigenvalues of the sample covariance matrix. in the heavy-tail case

The largest eigenvalues of the sample covariance matrix. in the heavy-tail case The largest eigenvalues of the sample covariance matrix 1 in the heavy-tail case Thomas Mikosch University of Copenhagen Joint work with Richard A. Davis (Columbia NY), Johannes Heiny (Aarhus University)

More information

Integrated Likelihood Estimation in Semiparametric Regression Models. Thomas A. Severini Department of Statistics Northwestern University

Integrated Likelihood Estimation in Semiparametric Regression Models. Thomas A. Severini Department of Statistics Northwestern University Integrated Likelihood Estimation in Semiparametric Regression Models Thomas A. Severini Department of Statistics Northwestern University Joint work with Heping He, University of York Introduction Let Y

More information

Convergence Rates of Kernel Quadrature Rules

Convergence Rates of Kernel Quadrature Rules Convergence Rates of Kernel Quadrature Rules Francis Bach INRIA - Ecole Normale Supérieure, Paris, France ÉCOLE NORMALE SUPÉRIEURE NIPS workshop on probabilistic integration - Dec. 2015 Outline Introduction

More information

Concentration behavior of the penalized least squares estimator

Concentration behavior of the penalized least squares estimator Concentration behavior of the penalized least squares estimator Penalized least squares behavior arxiv:1511.08698v2 [math.st] 19 Oct 2016 Alan Muro and Sara van de Geer {muro,geer}@stat.math.ethz.ch Seminar

More information

Sparse Legendre expansions via l 1 minimization

Sparse Legendre expansions via l 1 minimization Sparse Legendre expansions via l 1 minimization Rachel Ward, Courant Institute, NYU Joint work with Holger Rauhut, Hausdorff Center for Mathematics, Bonn, Germany. June 8, 2010 Outline Sparse recovery

More information

Negative Association, Ordering and Convergence of Resampling Methods

Negative Association, Ordering and Convergence of Resampling Methods Negative Association, Ordering and Convergence of Resampling Methods Nicolas Chopin ENSAE, Paristech (Joint work with Mathieu Gerber and Nick Whiteley, University of Bristol) Resampling schemes: Informal

More information

Multivariate quantiles and conditional depth

Multivariate quantiles and conditional depth M. Hallin a,b, Z. Lu c, D. Paindaveine a, and M. Šiman d a Université libre de Bruxelles, Belgium b Princenton University, USA c University of Adelaide, Australia d Institute of Information Theory and

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

Linear Models for Regression CS534

Linear Models for Regression CS534 Linear Models for Regression CS534 Example Regression Problems Predict housing price based on House size, lot size, Location, # of rooms Predict stock price based on Price history of the past month Predict

More information

Regression: Lecture 2

Regression: Lecture 2 Regression: Lecture 2 Niels Richard Hansen April 26, 2012 Contents 1 Linear regression and least squares estimation 1 1.1 Distributional results................................ 3 2 Non-linear effects and

More information

Why experimenters should not randomize, and what they should do instead

Why experimenters should not randomize, and what they should do instead Why experimenters should not randomize, and what they should do instead Maximilian Kasy Department of Economics, Harvard University Maximilian Kasy (Harvard) Experimental design 1 / 42 project STAR Introduction

More information

Ultra High Dimensional Variable Selection with Endogenous Variables

Ultra High Dimensional Variable Selection with Endogenous Variables 1 / 39 Ultra High Dimensional Variable Selection with Endogenous Variables Yuan Liao Princeton University Joint work with Jianqing Fan Job Market Talk January, 2012 2 / 39 Outline 1 Examples of Ultra High

More information

Chapter 2 Interpolation

Chapter 2 Interpolation Chapter 2 Interpolation Experiments usually produce a discrete set of data points (x i, f i ) which represent the value of a function f (x) for a finite set of arguments {x 0...x n }. If additional data

More information

Distributed Inexact Newton-type Pursuit for Non-convex Sparse Learning

Distributed Inexact Newton-type Pursuit for Non-convex Sparse Learning Distributed Inexact Newton-type Pursuit for Non-convex Sparse Learning Bo Liu Department of Computer Science, Rutgers Univeristy Xiao-Tong Yuan BDAT Lab, Nanjing University of Information Science and Technology

More information