Constrained Gaussian processes: methodology, theory and applications

Size: px
Start display at page:

Download "Constrained Gaussian processes: methodology, theory and applications"

Transcription

1 Constrained Gaussian processes: methodology, theory and applications Hassan Maatouk Workshop on Gaussian Processes, November 6-7, 2017, St-Etienne (France) Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 1 / 47

2 Plan 1 General introduction of GP regression and motivating eample 2 Gaussian processes with inequality constraints Finite-dimensional approimation of GPs Simulation of truncated Gaussian vectors 3 Generalization of the Kimeldorf-Wahba correspondence 4 Real application in Insurance and Finance : estimation of discount factors and default probabilities Discount factors Default probabilities Credit Default Swaps (CDS) 5 Noisy case Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 2 / 47

3 Plan 1 General introduction of GP regression and motivating eample 2 Gaussian processes with inequality constraints Finite-dimensional approimation of GPs Simulation of truncated Gaussian vectors 3 Generalization of the Kimeldorf-Wahba correspondence 4 Real application in Insurance and Finance : estimation of discount factors and default probabilities Discount factors Default probabilities Credit Default Swaps (CDS) 5 Noisy case Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 3 / 47

4 Gaussian Process Regression (GPR) or Kriging The following nonparametric function estimation is considered y = f(), R d. Observations : f( (i) ) = y i, i = 1,..., n. In statistical framework, y is viewed as a realization of a GP Y : Y () := η() + Z(), where η is the mean and Z is a zero-mean GP with covariance function K. The conditional process remains a GP { } Y () Y ( (1) ) = y 1,..., Y ( (n) ) = y n N ( ζ(), τ 2 () ), with { ζ() = η() + k() K 1 (y µ) τ 2 () = K(, ) k() K 1 k() µ i = η( (i) ), K i,j = K( (i), (j) ), i, j = 1,..., n and (k()) i = ( K(, (i) ) ). Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 4 / 47

5 Gaussian Process Regression Let f() = sin(5π), [0, 1] be a real function vraie fonction moyenne de krigeage Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 5 / 47

6 Motivating eample y() GP sample paths true function posterior mean Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 6 / 47

7 Motivating eample y() GP sample paths true function posterior mean y() true function posterior mean Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 6 / 47

8 Motivating eample y() GP sample paths true function posterior mean y() true function posterior mean monotone GP true function inequality mode Hassan Maatouk / Rennes 2 Gaussian processes withinequality constraints November 7, 2017, St-Etienne 6 / 47

9 Plan 1 General introduction of GP regression and motivating eample 2 Gaussian processes with inequality constraints Finite-dimensional approimation of GPs Simulation of truncated Gaussian vectors 3 Generalization of the Kimeldorf-Wahba correspondence 4 Real application in Insurance and Finance : estimation of discount factors and default probabilities Discount factors Default probabilities Credit Default Swaps (CDS) 5 Noisy case Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 7 / 47

10 Gaussian processes with inequality constraints I is the space of interpolation conditions : f( (i) ) = y i, i = 1,..., n. C is the space of conveity constraints (such as boundedness, monotonicity and conveity). (Y ()) [0,1] d is a zero-mean GP with covariance function : K (, ) = Cov (Y (), Y ( )) = E (Y ()Y ( )). Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 8 / 47

11 Gaussian processes with inequality constraints I is the space of interpolation conditions : f( (i) ) = y i, i = 1,..., n. C is the space of conveity constraints (such as boundedness, monotonicity and conveity). (Y ()) [0,1] d is a zero-mean GP with covariance function : K (, ) = Cov (Y (), Y ( )) = E (Y ()Y ( )). Formulation of the problem : simulate the Gaussian process Y conditionally to : Y ( (i) ) = y i, i = 1,..., n, Y C. (I C) Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 8 / 47

12 Gaussian processes with inequality constraints I is the space of interpolation conditions : f( (i) ) = y i, i = 1,..., n. C is the space of conveity constraints (such as boundedness, monotonicity and conveity). (Y ()) [0,1] d is a zero-mean GP with covariance function : K (, ) = Cov (Y (), Y ( )) = E (Y ()Y ( )). Formulation of the problem : simulate the Gaussian process Y conditionally to : Y ( (i) ) = y i, i = 1,..., n, Y C. (I C) Remark This is a difficult problem because the conditional process Y Y I C is not a GP in general. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 8 / 47

13 Finite-dimensional approimation of GPs Methodology : we develop a finite-dimensional approimation of GPs of the form N Y N () := ξ j φ j (), with ξ = (ξ 0 ξ N ) N (0, Γ N ) and {φ j } the deterministic basis functions. j=0 Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 9 / 47

14 Finite-dimensional approimation of GPs Methodology : we develop a finite-dimensional approimation of GPs of the form N Y N () := ξ j φ j (), with ξ = (ξ 0 ξ N ) N (0, Γ N ) and {φ j } the deterministic basis functions. j=0 Fundamental property of the basis functions (φ j ) j=0,,n we choose the basis functions {φ j } such that Y N (.) is positive ξ j 0 ; 0 j N. Y N () [a, b] a ξ j b ; 0 j N. Y N (.) is non-decreasing ξ j 0 ; 0 j N. Y N (.) is conve ξ j 0 ; 0 j N. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 9 / 47

15 Boundedness constraints 1 C = {f : [0, 1] R : a f() b}. 2 The basis functions (h j ) j are the hat functions associated to the knots (u j ) j=0,...,n such that : h j (u k ) = δ j,k. hi Y N () = N j=0 ξ jh j () and Y N (u k ) = N j=0 ξ jh j (u k ) = N j=0 ξ jδ j,k = ξ k. Y N N () = j=0 Y N (u j )h j () is a piecewise linear function. N Y N () = Y N (u j )h j () [a, b], ξ j = Y N (u j ) [a, b], j = 0,..., N. j=0 Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 10 / 47

16 Covariance matri of the random coefficients ξ j By the special choose of the basis functions, we have N N Y N () := ξ j h j () = Y N (u j )h j (). j=0 j=0 To ensure the almost surly uniform convergence of Y N to Y we suppose : Y N (u j ) = Y (u j ). Thus, N Y N () = Y (u j )h j (). j=0 Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 11 / 47

17 Covariance matri of the random coefficients ξ j By the special choose of the basis functions, we have N N Y N () := ξ j h j () = Y N (u j )h j (). j=0 j=0 To ensure the almost surly uniform convergence of Y N to Y we suppose : Y N (u j ) = Y (u j ). Thus, N Y N () = Y (u j )h j (). j=0 Γ N i,j = Cov(ξ i, ξ j ) = Cov(Y (u i ), Y (u j )) = K(u i, u j ), with K the covariance function of the original Gaussian process Y. Cov ( Y N (), Y N ( ) ) = h() Γ N h(). Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 11 / 47

18 Monotonicity constraints The basis functions are choose as the primitive of the hat functions : φ j () := 0 h j(t)dt. φ i In that case, the finite-dimensional approimation of GPs can be reformulated as N N Y N () := ζ + ξ j φ j () = Y (0) + Y (u j )φ j (). j=0 Thus, Γ N j,k = Cov(ξ j, ξ k ) = Cov (Y (u j ), Y (u k )) = 2 K(u j,u k ). Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 12 / 47 j=0

19 New formulation of the problem - case of monotonicity constraints The simulation of Y N conditionally to interpolation and inequality constraints is equivalent to the simulation of the Gaussian vector ξ such that N ( ) (Aξ) i := ζ + ξ j φ j (i) = y i, i = 1,..., n (n interpolation conditions) I ξ j=0 ξ j 0, j = 0,..., N (N+1 inequality constraints) C coef with A i,j := φ j ( (i) ) and A i,1 = 1, i = 1,..., n. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 13 / 47

20 New formulation of the problem - case of monotonicity constraints The simulation of Y N conditionally to interpolation and inequality constraints is equivalent to the simulation of the Gaussian vector ξ such that N ( ) (Aξ) i := ζ + ξ j φ j (i) = y i, i = 1,..., n (n interpolation conditions) I ξ j=0 ξ j 0, j = 0,..., N (N+1 inequality constraints) C coef with A i,j := φ j ( (i) ) and A i,1 = 1, i = 1,..., n. The problem is reduced to simulate a truncated Gaussian vector restricted to conve sets. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 13 / 47

21 New formulation of the problem - case of monotonicity constraints The simulation of Y N conditionally to interpolation and inequality constraints is equivalent to the simulation of the Gaussian vector ξ such that N ( ) (Aξ) i := ζ + ξ j φ j (i) = y i, i = 1,..., n (n interpolation conditions) I ξ j=0 ξ j 0, j = 0,..., N (N+1 inequality constraints) C coef with A i,j := φ j ( (i) ) and A i,1 = 1, i = 1,..., n. The problem is reduced to simulate a truncated Gaussian vector restricted to conve sets. truncated normal distribution ξ I = µ ξ C ξ I µ ξ C truncated normal distribution Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 13 / 47

22 Mean and maimum of the posterior distribution Definition (mean a posteriori estimate) The mean of the posterior distribution ( is defined as ) m N pos () := E Y N () Y N ( (i) ) = y i, ξ C coef = φ() ξ pos, ( ) where ξ pos = E ξ ξ Iξ C coef. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 14 / 47

23 Mean and maimum of the posterior distribution Definition (mean a posteriori estimate) The mean of the posterior distribution is defined as m N pos () := E ( Y N () Y N ( (i) ) = y i, ξ C coef ) = φ() ξ pos, where ξ pos = E ( ξ ξ Iξ C coef ). { } Let µ be the mode of the truncated Gaussian vector ξ ξ Iξ C coef. Then ( ) 1 µ = arg min ξ I ξ C coef 2 ξ (Γ N ) 1 ξ, (1) with Γ N the covariance matri of the Gaussian vector ξ. Definition (maimum a posteriori estimate) The maimum a posterior (MAP) estimate is defined as N Mpos() N := µ j φ j (), µ = (µ 0,..., µ N ) is computed from (1). j=0 Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 14 / 47

24 Truncated normal variables Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 15 / 47

25 Truncated normal variables Proposition Let f and g be two pseudo-density functions defined as f() := f( 0, Σ)1 C and g() := g( µ, Σ)1 C. The optimal constant k such that f() k g(), C is equal to k = ep ( 12 ) (µ ) Σ 1 µ. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 15 / 47

26 Simulation of the truncated multivariate Gaussian random variables Accept/Rejection algorithm 1 Generate X with density g while X / C. 2 Generate U uniformly on [0, 1] and accept X if U e (µ ) T Σ 1 µ X T Σ 1 µ. Otherwise, repeat from step 1. y[,2] [,2] y[,1] [,1] Figure: Crude rejection sampling with 2% acceptance rate (left) and the so-called rejection sampling from the mode (RSM) with 20% acceptance rate (right) Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 16 / 47

27 Illustrative eample - monotonicity constraints cases PG monotone Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 17 / 47

28 Illustrative eample - monotonicity constraints cases PG monotone moyenne de krigeage usuelle Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 17 / 47

29 Illustrative eample - monotonicity constraints cases PG monotone moyenne de krigeage usuelle moyenne de krigeage monotone Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 17 / 47

30 Illustrative eample - monotonicity constraints cases PG monotone moyenne de krigeage usuelle moyenne de krigeage monotone 95% intervalle de prédiction Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 17 / 47

31 Illustrative eample - monotonicity constraints cases PG monotone moyenne de krigeage usuelle moyenne de krigeage monotone 95% intervalle de prédiction PG monotone Figure: 100 sample paths taken from the GP approimation conditionally to interpolation condition and monotonicity constraints. The unconstrained kriging mean coincides with the MAP estimate in the left figure but not in the right one Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 17 / 47

32 Illustrative eample - monotonicity constraints cases PG monotone moyenne de krigeage usuelle moyenne de krigeage monotone 95% intervalle de prédiction PG monotone moyenne de krigeage usuelle Figure: 100 sample paths taken from the GP approimation conditionally to interpolation condition and monotonicity constraints. The unconstrained kriging mean coincides with the MAP estimate in the left figure but not in the right one Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 17 / 47

33 Illustrative eample - monotonicity constraints cases PG monotone moyenne de krigeage usuelle moyenne de krigeage monotone 95% intervalle de prédiction PG monotone moyenne de krigeage usuelle moyenne de krigeage monotone Figure: 100 sample paths taken from the GP approimation conditionally to interpolation condition and monotonicity constraints. The unconstrained kriging mean coincides with the MAP estimate in the left figure but not in the right one Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 17 / 47

34 Illustrative eample - monotonicity constraints cases PG monotone moyenne de krigeage usuelle moyenne de krigeage monotone 95% intervalle de prédiction PG monotone moyenne de krigeage usuelle moyenne de krigeage monotone mode Figure: 100 sample paths taken from the GP approimation conditionally to interpolation condition and monotonicity constraints. The unconstrained kriging mean coincides with the MAP estimate in the left figure but not in the right one Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 17 / 47

35 Illustrative eample - monotonicity constraints cases PG monotone moyenne de krigeage usuelle moyenne de krigeage monotone 95% intervalle de prédiction PG monotone moyenne de krigeage usuelle moyenne de krigeage monotone mode 95% intervalle de prédiction Figure: 100 sample paths taken from the GP approimation conditionally to interpolation condition and monotonicity constraints. The unconstrained kriging mean coincides with the MAP estimate in the left figure but not in the right one Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 17 / 47

36 Illustrative eample - boundedness constraints case Gaussian process approimation : N N Y N () := ξ j h j () = Y (u j )h j (). (2) j=0 j=0 Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 18 / 47

37 Illustrative eample - boundedness constraints case Gaussian process approimation : N N Y N () := ξ j h j () = Y (u j )h j (). (2) j=0 j=0 PG borné moyenne de krigeage usuelle moyenne de krigeage bornée mode 95% intervalle de prédiction observation Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 18 / 47

38 Illustrative eample - boundedness constraints case Gaussian process approimation : N N Y N () := ξ j h j () = Y (u j )h j (). (2) j=0 The basis functions h j, j = 0,..., N are the hat functions. j=0 Y N () [a, b] if and only if Y (u j ) [a, b]. The sample paths verify boundedness constraints in the entire domain. The mean and the maimum of the posterior distribution respect boundedness constraints contrarily to the unconstrained kriging mean. PG borné moyenne de krigeage usuelle moyenne de krigeage bornée mode 95% intervalle de prédiction observation Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 18 / 47

39 Illustrative eample - conveity constraints case The Gaussian process approimation is equal to N Y N () := Y (0) + Y (0) + Y (u j )ϕ j (). (3) j=0 ϕ j := ( ) t 0 0 h j(u)du dt. Y N is conve if and only if Y (u j ) 0. The simulated paths respect conveity constraints in the entire domain. The mean and the maimum of the posterior distribution respect conveity constraints contrarily to the unconstrained kriging mean. PG convee moyenne de krigeage usuelle moyenne de krigeage convee mode Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 19 / 47

40 Isotonicity in two dimensions The input = ( 1, 2 ) is supposed to be in [0, 1] 2. The monotonicity constraints with respect to the two inputs is defined as 1 [0, 1], 2 f( 1, 2 ) is monotone and 2 [0, 1], 1 f( 1, 2 ) is monotone. 2 1 Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 20 / 47

41 Isotonicity in two dimensions The finite-dimensional approimation of GPs is defined as N Y N ( 1, 2 ) := Y (u i, u j )h i ( 1 )h j ( 2 ), (4) i,j=0 where h j, j = 0,..., N are the hat functions. Y N is monotone with respect to the two inputs if and only if the random coefficients Y (u i, u j ) verify Y (u i 1, u j ) Y (u i, u j ) Y (u i, u j 1 ) Y (u i, u j ), i, j = 1,..., N. Y (u i 1, u 0 ) Y (u i, u 0 ), i = 1,..., N. Y (u 0, u j 1 ) Y (u 0, u j ), j = 1,..., N. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 21 / 47

42 Isotonicity with respect to only one variable The finite-dimensional approimation of GPs Y N ( 1, 2 ) = N Y (u i, u j )h i ( 1 )h j ( 2 ) (5) i,j=0 respects monotonicity constraints for only the first variable if and only if : Y (u i 1, u j ) Y (u i, u j ), i = 1,..., N and j = 0,..., N. X2 X1 Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 22 / 47

43 Plan 1 General introduction of GP regression and motivating eample 2 Gaussian processes with inequality constraints Finite-dimensional approimation of GPs Simulation of truncated Gaussian vectors 3 Generalization of the Kimeldorf-Wahba correspondence 4 Real application in Insurance and Finance : estimation of discount factors and default probabilities Discount factors Default probabilities Credit Default Swaps (CDS) 5 Noisy case Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 23 / 47

44 Kimeldorf-Wahba correspondence We consider the following optimization problem : min h H I h 2 H, (Q) H the RKHS with r.k. K which is the covariance kernel of Y, I the space of functions which verify the interpolation condition. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 24 / 47

45 Kimeldorf-Wahba correspondence We consider the following optimization problem : min h H I h 2 H, (Q) H the RKHS with r.k. K which is the covariance kernel of Y, I the space of functions which verify the interpolation condition. Theorem (Kimeldorf and Wahba 1970) The problem (Q) has the kriging mean as an unique solution : h opt () = k() K 1 y, (6) where y = (y 1,..., y n ), k() = ( K( (i), ) ) i and K i,j = ( K( (i), (j) ) ). Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 24 / 47

46 Generalization of the Kimeldorf-Wahba correspondence We consider the following conve optimization problem : min h H I C h 2 H, (P ) H the RKHS with r.k. K which is the covariance kernel of Y, I the space of functions which verify the interpolation condition, C the space of functions which verify inequality constraints (such as monotonicity, conveity,...). Theorem (Bay, Grammont, Maatouk, Electron. J. Statist.) The maimum a posteriori estimate (MAP) converges to the constrained spline M N pos() := N µ j φ j () h opt := arg min N + h H I C h 2 H. j=0 Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 25 / 47

47 Numerical illustration of the new correspondence boundedness constraints unconstrained mean mean a posteriori maimum a posteriori Figure: The unconstrained kriging mean coincides with the MAP in the right figure but not in the left one Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 26 / 47

48 Numerical illustration of the new correspondence boundedness constraints unconstrained mean mean a posteriori maimum a posteriori boundedness constraints unconstrained mean mean a posteriori maimum a posteriori Figure: The unconstrained kriging mean coincides with the MAP in the right figure but not in the left one Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 26 / 47

49 boundedness constraints 0.2 unconstrained mean mean a posteriori maimum a posteriori unconstrained mean mean a posteriori maimum a posteriori 30 boundedness constraints Numerical illustration of the new correspondence Figure: The unconstrained kriging mean coincides with the MAP in the right figure but not in the left one Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 26 / 47

50 Constrained cubic spline interpolation The cubic spline is the function minimises the linear enegry criterion (LE) (see, e.g., [Wolberg and Alfy, 2002]) : Subject to h I C : { 1 min 0 E L = 1 0 (h (t)) 2 dt. (7) } (h (t)) 2 dt, h H 2 I C, where H 2 = { h L 2 ([0, 1]) tel que h, h L 2 ([0, 1]) }. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 27 / 47

51 Constrained cubic spline interpolation The cubic spline is the function minimises the linear enegry criterion (LE) (see, e.g., [Wolberg and Alfy, 2002]) : Subject to h I C : { 1 min 0 E L = 1 0 (h (t)) 2 dt. (7) } (h (t)) 2 dt, h H 2 I C, where H 2 = { h L 2 ([0, 1]) tel que h, h L 2 ([0, 1]) }. This is equivalent to min h H α+β (i) +h( (i) )=y i α+β+h() C 1 0 (h (t)) 2 dt = h 2 H. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 27 / 47

52 Monotone cubic spline interpolation [Wolberg and Alfy, 2002] Table: Wolberg s data used to compare different methods f() approimation hyman Figure: Monotone cubic spline interpolation using Wolberg s data : Hyman approach (blue curve) and the proposed approach M N pos() with N = 1000 (red curve) Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 28 / 47

53 Monotone cubic spline interpolation [Wolberg and Alfy, 2002] Table: Linear Energy (LE) criterion using Wolberg s data Method E L approimation Hyman CSE FE LE SDDE MDE FB N LE approimation LE = Figure: LE using our approach and Wolberg s data Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 29 / 47

54 Plan 1 General introduction of GP regression and motivating eample 2 Gaussian processes with inequality constraints Finite-dimensional approimation of GPs Simulation of truncated Gaussian vectors 3 Generalization of the Kimeldorf-Wahba correspondence 4 Real application in Insurance and Finance : estimation of discount factors and default probabilities Discount factors Default probabilities Credit Default Swaps (CDS) 5 Noisy case Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 30 / 47

55 Discount factors Discount factor is known to be a non-increasing function with respect to time-to-maturities. It verifies the following linear equality constraints : ( A f(x) = b, f(x) = f( (1) ),..., f( )) (n), (8) where A is a m n matri and b R m. 1 Bounds for OIS discounting curves 0.9 OIS discount factors time to maturity Figure: Cousin and Niang 2014 Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 31 / 47

56 Discount factors using Swap vs Euribor 6M market quotes as of 02/06/2010 discount factor monotone GP 95% confidence interval mode Nelson Siegel Nelson Siegel Svensson discount factor monotone GP 95% confidence interval mode Nelson Siegel Nelson Siegel Svensson Figure: Simulated paths (gray lines) taken from the conditional GP with non-increasing constraints and market-fit constraints using the Gaussian covariance function with nugget equal to 10 5 (left) and the Matérn 5/2 covariance function without nugget (right). Swap vs. Euribor 6M market quotes as of 02/06/2010 Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 32 / 47

57 Forward rates using Swap vs Euribor 6M market quotes as 02/06/2010 Forward rate TF sample paths 95% confidence interval TF mode Nelson Siegel Nelson Siegel Svensson Forward rate TF sample paths 95% confidence interval TF mode Nelson Siegel Nelson Siegel Svensson Figure: Forward rates obtained from sample paths of previous figures with Gaussian covariance function (left) and Matérn 5/2 covariance function (right) Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 33 / 47

58 Several quotation dates Swap vs Euribor discount factors time to maturity dates Figure: Swap-vs.-Euribor discount factors as a function of time-to-maturities and quotation dates Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 34 / 47

59 Default probabilities Credit Default Swaps (CDS) on 06/01/2005 survival probability survival probability Figure: CDS implied survival curves (gray lines) given as simulated paths of a conditional GP with non-increasing constraints using a Gaussian covariance function (left) and a Matérn 5/2 covariance function (right) Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 35 / 47

60 Several quotation dates Credit Default Swaps (CDS) protection maturities dates Figure: CDS implied survival probabilities as a function of time-to-maturities and quotation dates Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 36 / 47

61 Plan 1 General introduction of GP regression and motivating eample 2 Gaussian processes with inequality constraints Finite-dimensional approimation of GPs Simulation of truncated Gaussian vectors 3 Generalization of the Kimeldorf-Wahba correspondence 4 Real application in Insurance and Finance : estimation of discount factors and default probabilities Discount factors Default probabilities Credit Default Swaps (CDS) 5 Noisy case Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 37 / 47

62 Noisy In many application situations, an approimate response is available f( (i) ) = y i + ɛ i = ỹ i, i = 1,..., n, where ɛ N (0, σnoise 2 I), with σ2 noise the noise variance and I the identity matri. Conditionally to noisy ỹ = (ỹ 1,..., ỹ n ), the process remains a GP Y () Y (X) = ỹ GP ( ζ(), τ 2 () ), where ζ() = η() + k() (K + σ 2 noisei) 1 (ỹ µ) ; τ 2 () = K(, ) k() (K + σ 2 noisei) 1 k(), and µ = η(x) is the vector of trend values at the design of eperiments, K i,j = K( (i), (j) ), i, j = 1,..., n is the covariance matri of Y (X) and k() = (K(, (i) )) i is the vector of covariance between Y () and Y (X). Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 38 / 47

63 Illustrative eample - boundedness constraints Bounded GP unconstrained mean mean a posteriori maimum a posteriori Positive GP unconstrained mean mean a posteriori maimum a posteriori Figure: The GP approimation with positivity constraints (right) and boundedness constraints (left). The unconstrained mean coincides with the maimum a posteriori in the right figure but not in the left one Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 39 / 47

64 Illustrative eample - monotonicity constraints Monotone GP unconstrained mean mean a posteriori maimum a posteriori Figure: The GP approimation with monotonicity constraints for sinusoidal function f() = 0.32( + sin()). The unconstrained mean does not coincide with the maimum a posteriori Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 40 / 47

65 Illustrative eample - isotonicity in two dimensions f(1,2) 2 1 X2 X1 Figure: The maimum a posteriori estimate respecting monotonicity (non-decreasing) constraints for the two inputs, and the associated contour levels Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 41 / 47

66 Simulation study The real non-decreasing functions proposed by [Holmes and Heard, 2003, Neelon and Dunson, 2004] and used in a comparative study by [Shively et al., 2009, Lin and Dunson, 2014] are considered 1 flat function f 1() = 3, (0, 10] ; 2 sinusoidal function f 2() = 0.32{ + sin()}, (0, 10] ; 3 step function f 3() = 3 if (0, 8] and f 3() = 8 if (8, 10] ; 4 linear function f 4() = 0.3, (0, 10] ; 5 eponential function f 5() = 0.15 ep(0.6 3), (0, 10] ; 6 logistic function f 6() = 3/{1 + ep( )}, (0, 10]. The root-mean-square error (RMSE) of the estimates is computed at the one hundred values taken uniformly (equidistant) in the interval (0, 10] : RMSE = 1 n ( 2, f( i ) n ˆf( i )) i=1 where ˆf() is the estimate of f() and i are the n equally-spaced -values. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 42 / 47

67 Simulation study Table: Root-mean-square error ( 100) for data of size n = 100. The results are obtained by repeating the simulation 5000 times Flat Step Linear Eponential Logistic Sinusoidal GP GP projection Regression spline GP approimation GP projection : Lin, L. and Dunson, D. B. (2014). Bayesian monotone regression using Gaussian process projection. Biometrika, 101(2) : Regression spline : Shively, T. S., Sager, T. W., and Walker, S. G. (2009). A Bayesian approach to non-parametric monotone function estimation. J. R. Stat. Soc. B, 71(1) : GP approimation : Maatouk, H. and Bay, X. (2017). Gaussian process emulators for computer eperiments with inequality constraints. Math.Geosci., 49 : Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 43 / 47

68 Simulation study f() % credible intervals true sinusoidal function maimum a posteriori Figure: The 95% credible intervals of the GP approimation together with the sinusoidal function, the (grey crosses) and the maimum a posteriori estimate Table: Empirical coverage (%) for 95% credible intervals at different values. The simulations are repeated 1000 times GP GP projection GP approimation Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 44 / 47

69 Figure: The root-mean-square error at different sample sizes together with the optimal values obtained in [Riihimäki and Vehtari, 2010]. The results are based on 1000 simulation replicates using the logistic function : 2/{1 + ep( 8 + 4)}, [0, 1] Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 45 / 47 Simulation study (methodology based on the knowledge of the derivatives of the GP at some input locations) root mean square error root mean square error RMSE=0.062 RMSE= sample size

70 References Holmes, C. and Heard, N. (2003). Generalized monotonic regression using random change points. Stat. Med., 22(4) : Lin, L. and Dunson, D. B. (2014). Bayesian monotone regression using Gaussian process projection. Biometrika, 101(2) : Maatouk, H. and Bay, X. (2016). A new rejection sampling method for truncated multivariate Gaussian random variables restricted to conve sets. In Cools, R. and Nuyens, D., editors, Monte Carlo and Quasi-Monte Carlo Methods, pages Springer International Publishing, Cham. Neelon, B. and Dunson, D. B. (2004). Bayesian isotonic regression and trend analysis. Biometrics, 60(2) : Riihimäki, J. and Vehtari, A. (2010). Gaussian processes with monotonicity information. J. Mach. Learn. Res., 9 : Shively, T. S., Sager, T. W., and Walker, S. G. (2009). A Bayesian approach to non-parametric monotone function estimation. J. R. Stat. Soc. B, 71(1) : Wolberg, G. and Alfy, I. (2002). An energy-minimization framework for monotonic cubic spline interpolation. Journal of Computational and Applied Mathematics, 143(2) : Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 46 / 47

71 Thank you for your attention. Hassan Maatouk / Rennes 2 Gaussian processes with inequality constraints November 7, 2017, St-Etienne 47 / 47

A New Method For Interpolating In A Convex Subset Of A Hilbert Space

A New Method For Interpolating In A Convex Subset Of A Hilbert Space A New Method For Interpolating In A Convex Subset Of A Hilbert Space Xavier Bay, Laurence Grammont, Hassan Maatouk To cite this version: Xavier Bay, Laurence Grammont, Hassan Maatouk. A New Method For

More information

Bayesian Estimation and Inference for the Generalized Partial Linear Model

Bayesian Estimation and Inference for the Generalized Partial Linear Model Bayesian Estimation Inference for the Generalized Partial Linear Model Haitham M. Yousof 1, Ahmed M. Gad 2 1 Department of Statistics, Mathematics Insurance, Benha University, Egypt. 2 Department of Statistics,

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

Uncertainty Quantification for Inverse Problems. November 7, 2011

Uncertainty Quantification for Inverse Problems. November 7, 2011 Uncertainty Quantification for Inverse Problems November 7, 2011 Outline UQ and inverse problems Review: least-squares Review: Gaussian Bayesian linear model Parametric reductions for IP Bias, variance

More information

Spatial smoothing using Gaussian processes

Spatial smoothing using Gaussian processes Spatial smoothing using Gaussian processes Chris Paciorek paciorek@hsph.harvard.edu August 5, 2004 1 OUTLINE Spatial smoothing and Gaussian processes Covariance modelling Nonstationary covariance modelling

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Spatially Adaptive Smoothing Splines

Spatially Adaptive Smoothing Splines Spatially Adaptive Smoothing Splines Paul Speckman University of Missouri-Columbia speckman@statmissouriedu September 11, 23 Banff 9/7/3 Ordinary Simple Spline Smoothing Observe y i = f(t i ) + ε i, =

More information

Effective Computation for Odds Ratio Estimation in Nonparametric Logistic Regression

Effective Computation for Odds Ratio Estimation in Nonparametric Logistic Regression Communications of the Korean Statistical Society 2009, Vol. 16, No. 4, 713 722 Effective Computation for Odds Ratio Estimation in Nonparametric Logistic Regression Young-Ju Kim 1,a a Department of Information

More information

Monitoring Wafer Geometric Quality using Additive Gaussian Process

Monitoring Wafer Geometric Quality using Additive Gaussian Process Monitoring Wafer Geometric Quality using Additive Gaussian Process Linmiao Zhang 1 Kaibo Wang 2 Nan Chen 1 1 Department of Industrial and Systems Engineering, National University of Singapore 2 Department

More information

Anomaly Detection and Removal Using Non-Stationary Gaussian Processes

Anomaly Detection and Removal Using Non-Stationary Gaussian Processes Anomaly Detection and Removal Using Non-Stationary Gaussian ocesses Steven Reece Roman Garnett, Michael Osborne and Stephen Roberts Robotics Research Group Dept Engineering Science Oford University, UK

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Spatial omain Hierarchical Modelling for Univariate Spatial ata Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A.

More information

Lecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu

Lecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu Lecture: Gaussian Process Regression STAT 6474 Instructor: Hongxiao Zhu Motivation Reference: Marc Deisenroth s tutorial on Robot Learning. 2 Fast Learning for Autonomous Robots with Gaussian Processes

More information

Gaussian Processes (10/16/13)

Gaussian Processes (10/16/13) STA561: Probabilistic machine learning Gaussian Processes (10/16/13) Lecturer: Barbara Engelhardt Scribes: Changwei Hu, Di Jin, Mengdi Wang 1 Introduction In supervised learning, we observe some inputs

More information

Bayesian Dynamic Linear Modelling for. Complex Computer Models

Bayesian Dynamic Linear Modelling for. Complex Computer Models Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

Computer Emulation With Density Estimation

Computer Emulation With Density Estimation Computer Emulation With Density Estimation Jake Coleman, Robert Wolpert May 8, 2017 Jake Coleman, Robert Wolpert Emulation and Density Estimation May 8, 2017 1 / 17 Computer Emulation Motivation Expensive

More information

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling François Bachoc former PhD advisor: Josselin Garnier former CEA advisor: Jean-Marc Martinez Department

More information

STA414/2104 Statistical Methods for Machine Learning II

STA414/2104 Statistical Methods for Machine Learning II STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements

More information

Gaussian Process Optimization with Mutual Information

Gaussian Process Optimization with Mutual Information Gaussian Process Optimization with Mutual Information Emile Contal 1 Vianney Perchet 2 Nicolas Vayatis 1 1 CMLA Ecole Normale Suprieure de Cachan & CNRS, France 2 LPMA Université Paris Diderot & CNRS,

More information

E(x i ) = µ i. 2 d. + sin 1 d θ 2. for d < θ 2 0 for d θ 2

E(x i ) = µ i. 2 d. + sin 1 d θ 2. for d < θ 2 0 for d θ 2 1 Gaussian Processes Definition 1.1 A Gaussian process { i } over sites i is defined by its mean function and its covariance function E( i ) = µ i c ij = Cov( i, j ) plus joint normality of the finite

More information

Where now? Machine Learning and Bayesian Inference

Where now? Machine Learning and Bayesian Inference Machine Learning and Bayesian Inference Dr Sean Holden Computer Laboratory, Room FC6 Telephone etension 67 Email: sbh@clcamacuk wwwclcamacuk/ sbh/ Where now? There are some simple take-home messages from

More information

Likelihood Bounds for Constrained Estimation with Uncertainty

Likelihood Bounds for Constrained Estimation with Uncertainty Proceedings of the 44th IEEE Conference on Decision and Control, and the European Control Conference 5 Seville, Spain, December -5, 5 WeC4. Likelihood Bounds for Constrained Estimation with Uncertainty

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Practical Bayesian Optimization of Machine Learning. Learning Algorithms

Practical Bayesian Optimization of Machine Learning. Learning Algorithms Practical Bayesian Optimization of Machine Learning Algorithms CS 294 University of California, Berkeley Tuesday, April 20, 2016 Motivation Machine Learning Algorithms (MLA s) have hyperparameters that

More information

Type II variational methods in Bayesian estimation

Type II variational methods in Bayesian estimation Type II variational methods in Bayesian estimation J. A. Palmer, D. P. Wipf, and K. Kreutz-Delgado Department of Electrical and Computer Engineering University of California San Diego, La Jolla, CA 9093

More information

Gaussian Process priors with Uncertain Inputs: Multiple-Step-Ahead Prediction

Gaussian Process priors with Uncertain Inputs: Multiple-Step-Ahead Prediction Gaussian Process priors with Uncertain Inputs: Multiple-Step-Ahead Prediction Agathe Girard Dept. of Computing Science University of Glasgow Glasgow, UK agathe@dcs.gla.ac.uk Carl Edward Rasmussen Gatsby

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

Soft and hard models. Jiří Militky Computer assisted statistical modeling in the

Soft and hard models. Jiří Militky Computer assisted statistical modeling in the Soft and hard models Jiří Militky Computer assisted statistical s modeling in the applied research Monsters Giants Moore s law: processing capacity doubles every 18 months : CPU, cache, memory It s more

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

Basic concepts in estimation

Basic concepts in estimation Basic concepts in estimation Random and nonrandom parameters Definitions of estimates ML Maimum Lielihood MAP Maimum A Posteriori LS Least Squares MMS Minimum Mean square rror Measures of quality of estimates

More information

Spatial Process Estimates as Smoothers: A Review

Spatial Process Estimates as Smoothers: A Review Spatial Process Estimates as Smoothers: A Review Soutir Bandyopadhyay 1 Basic Model The observational model considered here has the form Y i = f(x i ) + ɛ i, for 1 i n. (1.1) where Y i is the observed

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 7 Unsupervised Learning Statistical Perspective Probability Models Discrete & Continuous: Gaussian, Bernoulli, Multinomial Maimum Likelihood Logistic

More information

DISCUSSION: COVERAGE OF BAYESIAN CREDIBLE SETS. By Subhashis Ghosal North Carolina State University

DISCUSSION: COVERAGE OF BAYESIAN CREDIBLE SETS. By Subhashis Ghosal North Carolina State University Submitted to the Annals of Statistics DISCUSSION: COVERAGE OF BAYESIAN CREDIBLE SETS By Subhashis Ghosal North Carolina State University First I like to congratulate the authors Botond Szabó, Aad van der

More information

Advanced Introduction to Machine Learning CMU-10715

Advanced Introduction to Machine Learning CMU-10715 Advanced Introduction to Machine Learning CMU-10715 Gaussian Processes Barnabás Póczos http://www.gaussianprocess.org/ 2 Some of these slides in the intro are taken from D. Lizotte, R. Parr, C. Guesterin

More information

A short introduction to INLA and R-INLA

A short introduction to INLA and R-INLA A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk

More information

Random Fields in Bayesian Inference: Effects of the Random Field Discretization

Random Fields in Bayesian Inference: Effects of the Random Field Discretization Random Fields in Bayesian Inference: Effects of the Random Field Discretization Felipe Uribe a, Iason Papaioannou a, Wolfgang Betz a, Elisabeth Ullmann b, Daniel Straub a a Engineering Risk Analysis Group,

More information

Regression, Ridge Regression, Lasso

Regression, Ridge Regression, Lasso Regression, Ridge Regression, Lasso Fabio G. Cozman - fgcozman@usp.br October 2, 2018 A general definition Regression studies the relationship between a response variable Y and covariates X 1,..., X n.

More information

arxiv: v2 [stat.me] 24 Sep 2018

arxiv: v2 [stat.me] 24 Sep 2018 ANOTHER LOOK AT STATISTICAL CALIBRATION: A NON-ASYMPTOTIC THEORY AND PREDICTION-ORIENTED OPTIMALITY arxiv:1802.00021v2 [stat.me] 24 Sep 2018 By Xiaowu Dai, and Peter Chien University of Wisconsin-Madison

More information

Gaussian Process Vine Copulas for Multivariate Dependence

Gaussian Process Vine Copulas for Multivariate Dependence Gaussian Process Vine Copulas for Multivariate Dependence José Miguel Hernández-Lobato 1,2 joint work with David López-Paz 2,3 and Zoubin Ghahramani 1 1 Department of Engineering, Cambridge University,

More information

Regression with Input-Dependent Noise: A Bayesian Treatment

Regression with Input-Dependent Noise: A Bayesian Treatment Regression with Input-Dependent oise: A Bayesian Treatment Christopher M. Bishop C.M.BishopGaston.ac.uk Cazhaow S. Qazaz qazazcsgaston.ac.uk eural Computing Research Group Aston University, Birmingham,

More information

Introduction to emulators - the what, the when, the why

Introduction to emulators - the what, the when, the why School of Earth and Environment INSTITUTE FOR CLIMATE & ATMOSPHERIC SCIENCE Introduction to emulators - the what, the when, the why Dr Lindsay Lee 1 What is a simulator? A simulator is a computer code

More information

Bayesian Quadrature: Model-based Approximate Integration. David Duvenaud University of Cambridge

Bayesian Quadrature: Model-based Approximate Integration. David Duvenaud University of Cambridge Bayesian Quadrature: Model-based Approimate Integration David Duvenaud University of Cambridge The Quadrature Problem ˆ We want to estimate an integral Z = f ()p()d ˆ Most computational problems in inference

More information

Regularization in Reproducing Kernel Banach Spaces

Regularization in Reproducing Kernel Banach Spaces .... Regularization in Reproducing Kernel Banach Spaces Guohui Song School of Mathematical and Statistical Sciences Arizona State University Comp Math Seminar, September 16, 2010 Joint work with Dr. Fred

More information

Kriging by Example: Regression of oceanographic data. Paris Perdikaris. Brown University, Division of Applied Mathematics

Kriging by Example: Regression of oceanographic data. Paris Perdikaris. Brown University, Division of Applied Mathematics Kriging by Example: Regression of oceanographic data Paris Perdikaris Brown University, Division of Applied Mathematics! January, 0 Sea Grant College Program Massachusetts Institute of Technology Cambridge,

More information

Analytic Long-Term Forecasting with Periodic Gaussian Processes

Analytic Long-Term Forecasting with Periodic Gaussian Processes Nooshin Haji Ghassemi School of Computing Blekinge Institute of Technology Sweden Marc Peter Deisenroth Department of Computing Imperial College London United Kingdom Department of Computer Science TU

More information

Data Mining Stat 588

Data Mining Stat 588 Data Mining Stat 588 Lecture 9: Basis Expansions Department of Statistics & Biostatistics Rutgers University Nov 01, 2011 Regression and Classification Linear Regression. E(Y X) = f(x) We want to learn

More information

Hybrid Dirichlet processes for functional data

Hybrid Dirichlet processes for functional data Hybrid Dirichlet processes for functional data Sonia Petrone Università Bocconi, Milano Joint work with Michele Guindani - U.T. MD Anderson Cancer Center, Houston and Alan Gelfand - Duke University, USA

More information

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes COMP 55 Applied Machine Learning Lecture 2: Gaussian processes Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~hvanho2/comp55

More information

Color Scheme. swright/pcmi/ M. Figueiredo and S. Wright () Inference and Optimization PCMI, July / 14

Color Scheme.   swright/pcmi/ M. Figueiredo and S. Wright () Inference and Optimization PCMI, July / 14 Color Scheme www.cs.wisc.edu/ swright/pcmi/ M. Figueiredo and S. Wright () Inference and Optimization PCMI, July 2016 1 / 14 Statistical Inference via Optimization Many problems in statistical inference

More information

A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging

A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging Cheng Li DSAP, National University of Singapore Joint work with Rajarshi Guhaniyogi (UC Santa Cruz), Terrance D. Savitsky (US Bureau of Labor

More information

Can we do statistical inference in a non-asymptotic way? 1

Can we do statistical inference in a non-asymptotic way? 1 Can we do statistical inference in a non-asymptotic way? 1 Guang Cheng 2 Statistics@Purdue www.science.purdue.edu/bigdata/ ONR Review Meeting@Duke Oct 11, 2017 1 Acknowledge NSF, ONR and Simons Foundation.

More information

Abel Inversion using the Maximum Entropy Method

Abel Inversion using the Maximum Entropy Method Abel Inversion using the Maximum Entropy Method Danilo R. Neuber Wolfgang von der Linden 3rd October 2003 Inst. für Theoretische Physik, Tel.: +43/3 16/8 73-81 88; neuber@itp.tu-graz.ac.at Inst. für Theoretische

More information

Étude de classes de noyaux adaptées à la simplification et à l interprétation des modèles d approximation.

Étude de classes de noyaux adaptées à la simplification et à l interprétation des modèles d approximation. Étude de classes de noyaux adaptées à la simplification et à l interprétation des modèles d approximation. Soutenance de thèse de N. Durrande Ecole des Mines de St-Etienne, 9 november 2011 Directeurs Laurent

More information

Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach"

Kneib, Fahrmeir: Supplement to Structured additive regression for categorical space-time data: A mixed model approach Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach" Sonderforschungsbereich 386, Paper 43 (25) Online unter: http://epub.ub.uni-muenchen.de/

More information

Bayesian Methods / G.D. Hager S. Leonard

Bayesian Methods / G.D. Hager S. Leonard Bayesian Methods Recall Robot Localization Given Sensor readings z 1, z 2,, z t = z 1:t Known control inputs u 0, u 1, u t = u 0:t Known model t+1 t, u t ) with initial 1 u 0 ) Known map z t t ) Compute

More information

Computation fundamentals of discrete GMRF representations of continuous domain spatial models

Computation fundamentals of discrete GMRF representations of continuous domain spatial models Computation fundamentals of discrete GMRF representations of continuous domain spatial models Finn Lindgren September 23 2015 v0.2.2 Abstract The fundamental formulas and algorithms for Bayesian spatial

More information

Machine Learning Basics III

Machine Learning Basics III Machine Learning Basics III Benjamin Roth CIS LMU München Benjamin Roth (CIS LMU München) Machine Learning Basics III 1 / 62 Outline 1 Classification Logistic Regression 2 Gradient Based Optimization Gradient

More information

Slice Sampling with Adaptive Multivariate Steps: The Shrinking-Rank Method

Slice Sampling with Adaptive Multivariate Steps: The Shrinking-Rank Method Slice Sampling with Adaptive Multivariate Steps: The Shrinking-Rank Method Madeleine B. Thompson Radford M. Neal Abstract The shrinking rank method is a variation of slice sampling that is efficient at

More information

Doubly Penalized Likelihood Estimator in Heteroscedastic Regression 1

Doubly Penalized Likelihood Estimator in Heteroscedastic Regression 1 DEPARTMENT OF STATISTICS University of Wisconsin 1210 West Dayton St. Madison, WI 53706 TECHNICAL REPORT NO. 1084rr February 17, 2004 Doubly Penalized Likelihood Estimator in Heteroscedastic Regression

More information

L20: MLPs, RBFs and SPR Bayes discriminants and MLPs The role of MLP hidden units Bayes discriminants and RBFs Comparison between MLPs and RBFs

L20: MLPs, RBFs and SPR Bayes discriminants and MLPs The role of MLP hidden units Bayes discriminants and RBFs Comparison between MLPs and RBFs L0: MLPs, RBFs and SPR Bayes discriminants and MLPs The role of MLP hidden units Bayes discriminants and RBFs Comparison between MLPs and RBFs CSCE 666 Pattern Analysis Ricardo Gutierrez-Osuna CSE@TAMU

More information

Bayesian Isotonic Regression and Trend Analysis

Bayesian Isotonic Regression and Trend Analysis Bayesian Isotonic Regression and Trend Analysis Brian Neelon 1, and David B. Dunson 2 June 9, 2003 1 Department of Biostatistics, University of North Carolina, Chapel Hill, NC 2 Biostatistics Branch, MD

More information

Lecture 1c: Gaussian Processes for Regression

Lecture 1c: Gaussian Processes for Regression Lecture c: Gaussian Processes for Regression Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London c.archambeau@cs.ucl.ac.uk

More information

Nonparmeteric Bayes & Gaussian Processes. Baback Moghaddam Machine Learning Group

Nonparmeteric Bayes & Gaussian Processes. Baback Moghaddam Machine Learning Group Nonparmeteric Bayes & Gaussian Processes Baback Moghaddam baback@jpl.nasa.gov Machine Learning Group Outline Bayesian Inference Hierarchical Models Model Selection Parametric vs. Nonparametric Gaussian

More information

Mohsen Pourahmadi. 1. A sampling theorem for multivariate stationary processes. J. of Multivariate Analysis, Vol. 13, No. 1 (1983),

Mohsen Pourahmadi. 1. A sampling theorem for multivariate stationary processes. J. of Multivariate Analysis, Vol. 13, No. 1 (1983), Mohsen Pourahmadi PUBLICATIONS Books and Editorial Activities: 1. Foundations of Time Series Analysis and Prediction Theory, John Wiley, 2001. 2. Computing Science and Statistics, 31, 2000, the Proceedings

More information

Learning gradients: prescriptive models

Learning gradients: prescriptive models Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan

More information

Simulation optimization via bootstrapped Kriging: Survey

Simulation optimization via bootstrapped Kriging: Survey Simulation optimization via bootstrapped Kriging: Survey Jack P.C. Kleijnen Department of Information Management / Center for Economic Research (CentER) Tilburg School of Economics & Management (TiSEM)

More information

Reduction of Model Complexity and the Treatment of Discrete Inputs in Computer Model Emulation

Reduction of Model Complexity and the Treatment of Discrete Inputs in Computer Model Emulation Reduction of Model Complexity and the Treatment of Discrete Inputs in Computer Model Emulation Curtis B. Storlie a a Los Alamos National Laboratory E-mail:storlie@lanl.gov Outline Reduction of Emulator

More information

A STATISTICAL TECHNIQUE FOR MODELLING NON-STATIONARY SPATIAL PROCESSES

A STATISTICAL TECHNIQUE FOR MODELLING NON-STATIONARY SPATIAL PROCESSES A STATISTICAL TECHNIQUE FOR MODELLING NON-STATIONARY SPATIAL PROCESSES JOHN STEPHENSON 1, CHRIS HOLMES, KERRY GALLAGHER 1 and ALEXANDRE PINTORE 1 Dept. Earth Science and Engineering, Imperial College,

More information

Bayesian Inference of Noise Levels in Regression

Bayesian Inference of Noise Levels in Regression Bayesian Inference of Noise Levels in Regression Christopher M. Bishop Microsoft Research, 7 J. J. Thomson Avenue, Cambridge, CB FB, U.K. cmbishop@microsoft.com http://research.microsoft.com/ cmbishop

More information

Hierarchical Modeling for Univariate Spatial Data

Hierarchical Modeling for Univariate Spatial Data Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This

More information

Deconvolution. Parameter Estimation in Linear Inverse Problems

Deconvolution. Parameter Estimation in Linear Inverse Problems Image Parameter Estimation in Linear Inverse Problems Chair for Computer Aided Medical Procedures & Augmented Reality Department of Computer Science, TUM November 10, 2006 Contents A naive approach......with

More information

Nonparametric Bayesian Methods - Lecture I

Nonparametric Bayesian Methods - Lecture I Nonparametric Bayesian Methods - Lecture I Harry van Zanten Korteweg-de Vries Institute for Mathematics CRiSM Masterclass, April 4-6, 2016 Overview of the lectures I Intro to nonparametric Bayesian statistics

More information

Inversion Base Height. Daggot Pressure Gradient Visibility (miles)

Inversion Base Height. Daggot Pressure Gradient Visibility (miles) Stanford University June 2, 1998 Bayesian Backtting: 1 Bayesian Backtting Trevor Hastie Stanford University Rob Tibshirani University of Toronto Email: trevor@stat.stanford.edu Ftp: stat.stanford.edu:

More information

On the relation between the Smith-Wilson method and integrated Ornstein-Uhlenbeck processes

On the relation between the Smith-Wilson method and integrated Ornstein-Uhlenbeck processes Mathematical Statistics Stockholm University On the relation between the Smith-Wilson method and integrated Ornstein-Uhlenbeck processes Håkan Andersson Mathias Lindholm Research Report 213:1 ISSN 165-377

More information

Modeling and Interpolation of Non-Gaussian Spatial Data: A Comparative Study

Modeling and Interpolation of Non-Gaussian Spatial Data: A Comparative Study Modeling and Interpolation of Non-Gaussian Spatial Data: A Comparative Study Gunter Spöck, Hannes Kazianka, Jürgen Pilz Department of Statistics, University of Klagenfurt, Austria hannes.kazianka@uni-klu.ac.at

More information

An Introduction to GAMs based on penalized regression splines. Simon Wood Mathematical Sciences, University of Bath, U.K.

An Introduction to GAMs based on penalized regression splines. Simon Wood Mathematical Sciences, University of Bath, U.K. An Introduction to GAMs based on penalied regression splines Simon Wood Mathematical Sciences, University of Bath, U.K. Generalied Additive Models (GAM) A GAM has a form something like: g{e(y i )} = η

More information

Application of modified Leja sequences to polynomial interpolation

Application of modified Leja sequences to polynomial interpolation Special Issue for the years of the Padua points, Volume 8 5 Pages 66 74 Application of modified Leja sequences to polynomial interpolation L. P. Bos a M. Caliari a Abstract We discuss several applications

More information

Perturbation Theory for Variational Inference

Perturbation Theory for Variational Inference Perturbation heor for Variational Inference Manfred Opper U Berlin Marco Fraccaro echnical Universit of Denmark Ulrich Paquet Apple Ale Susemihl U Berlin Ole Winther echnical Universit of Denmark Abstract

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

A Nonparametric Monotone Regression Method for Bernoulli Responses with Applications to Wafer Acceptance Tests

A Nonparametric Monotone Regression Method for Bernoulli Responses with Applications to Wafer Acceptance Tests A Nonparametric Monotone Regression Method for Bernoulli Responses with Applications to Wafer Acceptance Tests Jyh-Jen Horng Shiau (Joint work with Shuo-Huei Lin and Cheng-Chih Wen) Institute of Statistics

More information

Bayesian reconstruction of free energy profiles from umbrella samples

Bayesian reconstruction of free energy profiles from umbrella samples Bayesian reconstruction of free energy profiles from umbrella samples Thomas Stecher with Gábor Csányi and Noam Bernstein 21st November 212 Motivation Motivation and Goals Many conventional methods implicitly

More information

A Practitioner s Guide to Generalized Linear Models

A Practitioner s Guide to Generalized Linear Models A Practitioners Guide to Generalized Linear Models Background The classical linear models and most of the minimum bias procedures are special cases of generalized linear models (GLMs). GLMs are more technically

More information

Iterative Gaussian Process Regression for Potential Energy Surfaces. Matthew Shelley University of York ISNET-5 Workshop 6th November 2017

Iterative Gaussian Process Regression for Potential Energy Surfaces. Matthew Shelley University of York ISNET-5 Workshop 6th November 2017 Iterative Gaussian Process Regression for Potential Energy Surfaces Matthew Shelley University of York ISNET-5 Workshop 6th November 2017 Outline Motivation: Calculation of potential energy surfaces (PES)

More information

Approximately Optimal Experimental Design for Heteroscedastic Gaussian Process Models

Approximately Optimal Experimental Design for Heteroscedastic Gaussian Process Models Neural Computing Research Group Aston University Birmingham B4 7ET United Kingdom Tel: +44 (0)121 333 4631 Fax: +44 (0)121 333 4586 http://www.ncrg.aston.ac.uk/ Approximately Optimal Experimental Design

More information

Jyh-Jen Horng Shiau 1 and Lin-An Chen 1

Jyh-Jen Horng Shiau 1 and Lin-An Chen 1 Aust. N. Z. J. Stat. 45(3), 2003, 343 352 A MULTIVARIATE PARALLELOGRAM AND ITS APPLICATION TO MULTIVARIATE TRIMMED MEANS Jyh-Jen Horng Shiau 1 and Lin-An Chen 1 National Chiao Tung University Summary This

More information

CS-E4830 Kernel Methods in Machine Learning

CS-E4830 Kernel Methods in Machine Learning CS-E483 Kernel Methods in Machine Learning Lecture : Gaussian processes Markus Heinonen 3. November, 26 Today s topic Gaussian processes: Kernel as a covariance function C. E. Rasmussen & C. K. I. Williams,

More information

Gaussian Process Regression

Gaussian Process Regression Gaussian Process Regression 4F1 Pattern Recognition, 21 Carl Edward Rasmussen Department of Engineering, University of Cambridge November 11th - 16th, 21 Rasmussen (Engineering, Cambridge) Gaussian Process

More information

12 - Nonparametric Density Estimation

12 - Nonparametric Density Estimation ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department

More information

An Adaptive Test of Independence with Analytic Kernel Embeddings

An Adaptive Test of Independence with Analytic Kernel Embeddings An Adaptive Test of Independence with Analytic Kernel Embeddings Wittawat Jitkrittum 1 Zoltán Szabó 2 Arthur Gretton 1 1 Gatsby Unit, University College London 2 CMAP, École Polytechnique ICML 2017, Sydney

More information

Independent and conditionally independent counterfactual distributions

Independent and conditionally independent counterfactual distributions Independent and conditionally independent counterfactual distributions Marcin Wolski European Investment Bank M.Wolski@eib.org Society for Nonlinear Dynamics and Econometrics Tokyo March 19, 2018 Views

More information

CSci 8980: Advanced Topics in Graphical Models Gaussian Processes

CSci 8980: Advanced Topics in Graphical Models Gaussian Processes CSci 8980: Advanced Topics in Graphical Models Gaussian Processes Instructor: Arindam Banerjee November 15, 2007 Gaussian Processes Outline Gaussian Processes Outline Parametric Bayesian Regression Gaussian

More information

Model Selection for Gaussian Processes

Model Selection for Gaussian Processes Institute for Adaptive and Neural Computation School of Informatics,, UK December 26 Outline GP basics Model selection: covariance functions and parameterizations Criteria for model selection Marginal

More information

Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones

Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones http://www.mpia.de/homes/calj/mlpr_mpia2008.html 1 1 Last week... supervised and unsupervised methods need adaptive

More information

Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes

Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes TTU, October 26, 2012 p. 1/3 Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes Hao Zhang Department of Statistics Department of Forestry and Natural Resources Purdue University

More information

Overview. DS GA 1002 Probability and Statistics for Data Science. Carlos Fernandez-Granda

Overview. DS GA 1002 Probability and Statistics for Data Science.   Carlos Fernandez-Granda Overview DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing

More information

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the

More information

On a Data Assimilation Method coupling Kalman Filtering, MCRE Concept and PGD Model Reduction for Real-Time Updating of Structural Mechanics Model

On a Data Assimilation Method coupling Kalman Filtering, MCRE Concept and PGD Model Reduction for Real-Time Updating of Structural Mechanics Model On a Data Assimilation Method coupling, MCRE Concept and PGD Model Reduction for Real-Time Updating of Structural Mechanics Model 2016 SIAM Conference on Uncertainty Quantification Basile Marchand 1, Ludovic

More information

PART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics

PART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics Table of Preface page xi PART I INTRODUCTION 1 1 The meaning of probability 3 1.1 Classical definition of probability 3 1.2 Statistical definition of probability 9 1.3 Bayesian understanding of probability

More information