Bayesian inference & process convolution models Dave Higdon, Statistical Sciences Group, LANL

Size: px
Start display at page:

Download "Bayesian inference & process convolution models Dave Higdon, Statistical Sciences Group, LANL"

Transcription

1 1 Bayesian inference & process convolution models Dave Higdon, Statistical Sciences Group, LANL

2 2 MOVING AVERAGE SPATIAL MODELS

3 Kernel basis representation for spatial processes z(s) Define m basis functions k 1 (s),..., k m (s). basis ! Here k j (s) is normal density cetered at spatial location ω j : k j (s) = 1 2π exp{ 1 2 (s ω j) 2 } s set z(s) = m k j(s)x j where x N(0,I m ). j=1 Can represent z =(z(s 1 ),..., z(s n )) T as z = Kx where K ij = k j (s i )

4 x and k(s) determine spatial processes z(s) k j (s)x j z(s) basis! z(s)! ! s! s Continuous representation: z(s) = m j=1 k j(s)x j where x N(0,I m ). Discrete representation: For z =(z(s 1 ),..., z(s n )) T, z = Kx where K ij = k j (s i )

5 Estimate z(s) by specifying k j (s) and estimating x k j (s)x j z(s) fitted bases!1.0! y(s)!2.0!1.5!1.0! ! s! s Discrete representation: For z =(z(s 1 ),..., z(s n )) T, z = Kx where K ij = k j (s i ) standard regression model: y = Kx + ɛ

6 Formulation for the 1-d example Data y = (y(s 1 ),..., y(s n )) T observed at locations s 1,..., s n. Once knot locations ω j,j=1,..., m and kernel choice k(s) are specified, the remaining model formulation is trivial: Likelihood: L(y x, λ y ) λ n 2 y exp 1 2 λ y(y Kx) T (y Kx) 6 where K ij = k(ω j s i ). Priors: π(x λ x ) λ m 2 x exp 1 2 λ xx T x Posterior: π(λ x ) λ a x 1 x exp{ b x λ x } π(λ y ) λ a y 1 y exp{ b y λ y } π(x, λ x, λ y y) λ a y+ n 2 1 y exp { λ y [b y +.5(y Kx) T (y Kx)] } λ a x+ m 2 1 x exp { λ x [b x +.5x T x] }

7 Posterior and full conditionals Posterior: π(x, λ x, λ y y) λ a y+ n 2 1 y exp { λ y [b y +.5(y Kx) T (y Kx)] } Full conditionals: λ a x+ m 2 1 x exp { λ x [b x +.5x T x] } π(x ) exp{ 1 2 [λ yx T K T Kx 2λ y x T K T y + λ x x T x]} π(λ x ) λ a x+ m 2 1 x exp { λ x [b x +.5x T x] } 7 π(λ y ) λ a y+ n 2 1 y exp { λ y [b y +.5(y Kx) T (y Kx)] } Gibbs sampler implementation x N((λ y K T K + λ x I m ) 1 λ y K T y, (λ y K T K + λ x I m ) 1 ) λ x Γ(a x + m 2,b x +.5x T x) λ y Γ(a y + n 2,b y +.5(y Kx) T (y Kx))

8 Gibbs sampler: intuition Gibbs sampler for a bivariate normal density π(z) =π(z 1,z 2 ) 1 ρ ρ exp 1 2 ( z 1 z 2 ) 1 ρ ρ 1 1 z 1 z 2 Full conditionals of π(z): 9 z 1 z 2 N(ρz 2, 1 ρ 2 ) z 2 z 1 N(ρz 1, 1 ρ 2 ) initialize chain with z 0 N 0 0, 1 ρ ρ 1 z 2 (z 1 1,z 1 2)=z 1 draw z 1 1 N(ρz 0 2, 1 ρ 2 ) now (z 1 1,z 0 2) T π(z) z 0 =(z 0 1,z 0 2) (z 1 1,z 0 2) z 1

9 Gibbs sampler: intuition Gibbs sampler gives z 0,z 2,...,z T which can be treated as dependent draws from π(z). 10 If z 0 is not a draw from π(z), then the initial realizations will not have the correct distribution. In practice, the first 100?, 1000? realizations are discarded. The draws can be used to make inference about π(z): Posterior mean of z is estimated by: ˆµ 1 = 1 ˆµ 2 T Posterior probabilities: T k=1 z1 k z2 k z 2 P (z 1 > 1) = 1 T P (z 1 >z 2 )= 1 T T k=1 T k=1 90% interval: (z [5%] 1,z [95%] 1 ). I[z k 1 > 1] I[z k 1 >z k 2] z 1

10 1-d example m =6knots evenly spaced between.3 and 1.2. n =5data points at s =.05,.25,.52,.65,.91. k(s) is N(0, sd =.3) a y = 10, b y = 10 (.25 2 ) strong prior at λ y =1/.25 2 ; a x =1, b x = basis functions 1 1 y(s) 0 k j (s) 0!1!1!2!2 20 x j! s !2! iteration! x,! y! s iteration

11 1-d example From posterior realizations of knot weights x, one can construct posterior realizations of the smooth fitted function z(s) = m j=1 k j (s)x j. Note strong prior on λ y required since n is small. 3 2 posterior realizations of z(s) 3 2 mean & pointwise 90% bounds z(s), y(s) 1 0 z(s), y(s) !1!1!2!2! s! s

12 1-d example m = 20 knots evenly spaced between 2 and 12. n = 18 data points evenly spaced between 0 and 10. k(s) is N(0, sd = 2) data posterior mean and 80% CI for z(s) 8 y y s s posterior dist for lam_y posterior dist for lam_x lambda_y lambda_x

13 1-d example m = 20 knots evenly spaced between 2 and 12. n = 18 data points evenly spaced between 0 and 10. k(s) is N(0, sd = 1) data posterior mean and 80% CI for z(s) 9 y y s s posterior dist for lam_y posterior dist for lam_x lambda_y lambda_x

14 Basis representations for spatial processes z(s) Represent z(s) at spatial locations s 1,..., s n. z =(z(s 1 ),..., z(s n )) T N(0, Σ z ). Recall z = Kx, where KK T = Σ z and x N(0,I n ). Gives a discrete representation of z(s) at locations s 1,..., s n. discrete representation continuous representation 10 basis!1.0! basis!1.0! s s z(s i )= n j=1 K ijx j z(s) = n k j(s)x j j=1 Columns of K give basis functions. Can use a subset of these basis functions to reduce dimensionality.

15 decomposition basis covariance Cholesky (w/piv) basis!1.0! location s s location s 11 SVD basis!1.0! location s s location s kernels basis!1.0! location s s location s

16 How many basis kernels? Define m basis functions k 1 (s),..., k m (s). m = 20? basis!1.0! location s s location s 12 m = 10? basis!1.0! location s s location s m =6? basis!1.0! location s s location s

17 Moving average specifications for spatial models z(s) smoothing kernel k(s) 13 white noise process x =(x 1,..., x m ) at spatial locations ω 1,..., ω m. x N(0, 1 I m ) λ x spatial process z(s) constructed by convolving x with smoothing kernel k(s) z(s) = m x jk(ω j s) j=1 z(s) is a Gaussian process with mean 0 and covariance given by Cov(z(s),z(s )) = 1 λ x m j=1 k(ω j s)k(ω j s )

18 14

19 Example: constructing 1-d models for z(s) z(s), x(s) s m = 20 knot locations ω 1,..., ω m equally spaced between 2 and 12. x =(x 1,..., x m ) T N(0,I m ) z(s), x(s) z(s) = m k=1 k(ω k s)x k k(s) is a normal density with sd = 1 4, 1 2, and 1 4th frame uses k(s) = 1.4 exp{ 1.4 s }. s 16 z(s), x(s) z(s), x(s) s General points: smooth kernels required spacing depends on kernel width knot spacing 1 sd for normal k(s) kernel width is equivalent to scale parameter in GP models s

20 17

21 MRF formulation for the 1-d example Data y = (y(s 1 ),..., y(s n )) T observed at locations s 1,..., s n. Once knot locations ω j,j=1,..., m and kernel choice k(s) are specified, the remaining model formulation is trivial: Likelihood: L(y x, λ y ) λ n 2 y exp 1 2 λ y(y Kx) T (y Kx) 18 where K ij = k(ω j s i )x j. Priors: π(x λ x ) λ m 2 x exp 1 2 λ xx T Wx Posterior: π(λ x ) λ a x 1 x exp{ b x λ x } π(λ y ) λ a y 1 y exp{ b y λ y } π(x, λ x, λ y y) λ a y+ n 2 1 y exp { λ y [b y +.5(y Kx) T (y Kx)] } λ a x+ m 2 1 x exp { λ x [b x +.5x T Wx] }

22 Posterior: Posterior and full conditionals (MRF formulation) π(x, λ x, λ y y) λ a y+ n 2 1 y exp { λ y [b y +.5(y Kx) T (y Kx)] } Full conditionals: λ a x+ m 2 1 x exp { λ x [b x +.5x T Wx] } π(x ) exp{ 1 2 [λ yx T K T Kx 2λ y x T K T y + λ x x T Wx]} π(λ x ) λ a x+ m 2 1 x exp { λ x [b x +.5x T Wx] } 19 π(λ y ) λ a y+ n 2 1 y exp { λ y [b y +.5(y Kx) T (y Kx)] } Gibbs sampler implementation x N((λ y K T K + λ x W ) 1 λ y K T y, (λ y K T K + λ x W ) 1 ) λ x Γ(a x + m 2,b x +.5x T x) λ y Γ(a y + n 2,b y +.5(y Kx) T (y Kx))

23 1-d example - using MRF prior for x post mean & 80% ci for z(s) post mean for x s (d) gp model (c) iid x s; k(s) fat (b) rw x s; k(s) skinny (a) iid x s; k(s) skinny estimated process z(s) estimated x & Kx spatial location s

24 8 hour max for ozone on a summer day in the Eastern US 21 n = 510 ozone measurements ω k s laid out on a hexagonal lattice as shown. k(s) is circular normal, with sd = lattice spacing. Choice of width for k(s): could look at empirical variogram could estimate using ML or cross-validation could treat as additional parameter in posterior

25 A multiresolution spatial model formulation 22 z(s) =z coarse (s)+z fine (s) Coarse process: m c = 27 locations ω1, c..., ωm c c grid. x c =(x c1,..., x cmc ) T N(0, 1 λ c I mc ) on a hexagonal coarse smoothing kernel k c (s) is normal with sd = coarse grid spacing. Fine process: m f = 87 locations ω1 f,..., ωm f f grid. on a hexagonal x f =(x f1,..., x fmf ) T N(0, 1 λ I mf ) f fine smoothing kernel k f (s) is normal with sd = fine grid spacing. note: coarse kernel width is twice the fine kernel width.

26 23 Model: where Multiresolution formulation and full conditionals y = K c x c + K f x f + ɛ y = Kx + ɛ K =(K c K f ) and x = Define λ W x = c I mc 0 0 λ f I mf Gibbs sampler implementation then becomes x c x f x N((λ y K T K + W x ) 1 λ y K T y, (λ y K T K + W x ) 1 ) λ c Γ(a x + m c 2,b x +.5x T c x c ) λ f Γ(a x + m f 2,b x +.5x T f x f ) λ y Γ(a y + n 2,b y +.5(y Kx) T (y Kx))

27 Multiresolution model for 8 hour max ozone coarse formulation 24 coarse + fine formulation

28 Basic binary classification example data true field binary spatial process z (s) spatial area partitioned into two regions: z (s) =1and z (s) =0. n = 10 measurements y =(y 1,..., y n ) T taken at spatial locations s 1,..., s n. y i = z (s i )+ɛ i, i =1,..., n; ɛ i iid N(0, 1), i =1,..., n

29 x Z Constructing a binary spatial process z (s) knot values x smoothing kernel z(s) k(s) Z ! s s s s Y X x convolved with k(s) gives z(s) x N(0,I m ) where each x j is located at ω j ; z(s) = m j=1 x jk(s ω j ) 26 z(s) binary z*(s) Z Y X Y X Now define the binary field z (s) by clipping z(s): z (s) =I[z(s) > 0].

30 Model Formulation Define n = 10 observations y =(y(s 1 ),..., y(s n )) T. Define z =(z (s 1 ),..., z (s n )) T. Define x =(x(ω 1 ),..., x(ω m )) T to be the m = 25-vector of white noise knot values at spatial grid sites ω 1,..., ω m. Recall z (s) and the vector z are determined by the knot values x. 27 Likelihood Independent normal prior for x L(y z ) exp{ 1 2 (y z ) T (y z )} π(x) exp{ 1 2 xt x} Posterior distribution π(x y) exp{ 1 2 (y z ) T (y z ) 1 2 xt x}

31 Sampling the posterior via Metropolis Full conditional distributions π(x j x j,y) exp{ 1 2 (y z ) T (y z ) 1 2 x2 j}, j=1,..., m Metropolis implementation for sampling from π(x y): Initialize x at x =0. Cycle thru full conditionals updating each x j according to Metropolis rules. 28 generate proposal x j U[x j r, x j + r]. compute acceptance probability α = min 1, π(x j x j,y) π(x j x j,y) update x j to new value: x new j = x j x j with probability α with probability 1 α Here we ran for T = 1000 scans, giving realizations x 1,..., x T from the posterior. Discarded the first 100 for burn in. Note: proposal width r tuned so that x j is accepted about half the time.

32 Sampling from non-standard multivariate distributions Nick Metropolis Computing pioneer at Los Alamos National Laboratory inventor of the Monte Carlo method inventor of Markov chain Monte Carlo: 29 Equation of State Calculations by Fast Computing Machines (1953) by N. Metropolis, A. Rosenbluth, M. Rosenbluth, A. Teller and E. Teller, Journal of Chemical Physics. Originally implemented on the MANIAC1 computer at LANL Algorithm constructs a Markov chain whose realizations are draws from the target (posterior) distribution. Constructs steps that maintain detailed balance.

33 Gibbs Sampling and Metropolis for a bivariate normal density π(z 1,z 2 ) 1 ρ ρ exp 1 2 ( z 1 z 2 ) 1 ρ ρ 1 1 z 1 z 2 30 sampling from the full conditionals also called heat bath z 1 z 2 N(ρz 2, 1 ρ 2 ) z 2 z 1 N(ρz 1, 1 ρ 2 ) Metropolis updating: generate z1 U[z 1 r, z 1 + r] calculate α = min{1, π(z 1,z 2) set z new 1 = π(z 1,z 2 ) = π(z 1 z 2) π(z 1 z 2 ) } z 1 with probability α z 1 with probability 1 α

34 31 Posterior realizations of z (s) =I[z(s) > 0]

35 data Posterior mean of z (s) =I[z(s) > 0] true field posterior mean

36 Application locating archeological sites ?? ? ?? ? ? ?? Data at locations over 10 m 2 area Assume y(s i ) z (s i ) N(µ(z (s i )), 1) µ(0) = 1, µ(1) = 2. Recall z(s) = m j=1 x j k(s ω j ) and z (s) =I[z(s) > 0]. m =8 8 lattice of knot locations ω 1,..., ω m ; sd of k(s) equal to knot spacings.

37 Posterior mean and realizations for z (s) =I[z(s) > 0] 34

38 Bayesian formulation Well.NW Well.N Well.NE Time Time Time Well.W Well.E Time Time 42 Well.SW Well.S Well.SE Time L(y η(z)) Σ 1 2 exp 1 2 (y η(z))t Σ 1 (y η(z)) π(z λ z ) λ m 2 z exp Time 1 2 zt W z z π(λ z ) λ a z 1 z exp {b z λ z } π(z,λ z y) L(y η(z)) π(z λ z ) π(λ z ) Time

39 Posterior realizations of z under MRF and moving average priors Well Data MRF Realization MRF Realization MRF Posterior Mean I I I I P P P GP Realization GP Realization GP Posterior Mean

40 44 References D. Higdon (1998) A process-convolution approach to modeling temperatures in the North Atlantic Ocean (with discussion), Environmental and Ecological Statistics, 5: D. Higdon (2002) Space and space-time modeling using process convolutions, in Quantitative Methods for Current Environmental Issues (C. Anderson and V. Barnett and P. C. Chatwin and A. H. El-Shaarawi, eds), D. Higdon, H. Lee and C. Holloman (2003) Markov chain Monte Carlo approaches for inference in computationally intensive inverse problems, in Bayesian Statistics 7, Proceedings of the Seventh Valencia International Meeting (Bernardo, Bayarri, Berger, Dawid, Heckerman, Smith and West, eds).

41 45 NON-STATIONARY SPATIAL CONVOLUTION MODELS

42 A convolution-based approach for building non-stationary spatial models * = white noise smoothing kernel spatial field 46 z(s) = m x jk(ω j s) = m x jk s (ω j ) j=1 j=1 x N(0,I m ) where each x j is located at ω j over a regular 2-d grid.

43 Convolutions for constructing non-stationary spatial models 47 m z(s) = x jk s (ω j ) j=1 x N(0,λ 1 x I m ) at regular 2-d lattice locations ω 1,..., ω m. z(s) GP (0,C(s 1,s 2 ) where C(s 1,s 2 )=λ 1 x k s 1 (ω j )k s2 (ω j ) j=1 smoothing kernel k s () changes smoothly over spatial location m

44 Defining smoothly varying kernels via basis kernels k 1 () k 2 () k 3 () 48 Define k s () =w 1 (s)k 1 ( s)+w 2 (s)k 2 ( s)+w 3 (s)k 3 ( s) k s () is a weighted combination of kernels centered at s. Define weights that change smoothly over space.

45 Defining smoothly varying kernels via basis kernels 49 Define iid, mean 0 Gaussian Processes φ 1 (s), φ 2 (s) and φ 3 (s) Set w i (s) = exp{φ i (s)} exp{φ 1 (s)} + exp{φ 2 (s)} + exp{φ 3 (s)} Estimate φ i (s) s like any other parameter in the analysis.

46 50 An application to Piazza Road Superfund site

47 51 MOVING AVERAGE/BASIS SPACE-TIME MODELS

48 A convolution-based approach for building space-time models marked point process smoothing kernel space-time field time time 52 space space Define space-time domain S T Define discrete knot process x(s, t) on {(ω 1, τ 1 ),..., (ω m, τ m )} within S T Define smoothing kernel k(s, t) Construct space-time process z(s, t) z(s, t) = z(s, t) = m j=1 k((s, t) (ω j, τ j ))x(ω j, τ j ) or with varying kernel m j=1 k st(ω j, τ j )x j

49 A Space-time model for ocean temperatures ocean temperature at constant potential density 50N 45N latitude 40N 35N 30N 25N 53 Data: y =(y 1,..., y n ) T at space-time locations: (s 1,t 1 ),..., (s n,t n ) Times: assume data are centered (ȳ =0) degrees C 20N W 25W 20W 15W 10W 5W longitude

50 Knot locations and kernels smoothing kernels and latent grid latitude 50N 45N 40N 35N 30N 25N 20N 54 m knot locations over a space-time grid (ω 1, τ 1 ),..., (ω m, τ m ) Spatial knot locations shown; Temporal knot spacing 7 years; ; Kernels vary with spatial location k st (ω, τ) =k s (ω, τ) use spatial mixture of normal basis kernels. 30W 25W 20W 15W 10W 5W longitude

51 Likelihood: Formulation for the ocean example L(y x, λ y ) λ n 2 y exp 1 2 λ y(y Kx) T (y Kx) where K ij = k si t i (ω j, τ j )x j. Priors: π(x λ x ) λ m 2 x exp 1 2 λ xx T x 55 Posterior: π(λ x ) λ a x 1 x exp{ b x λ x } π(λ y ) λ a y 1 y exp{ b y λ y } π(x, λ x, λ y y) λ a y+ n 2 1 y exp { λ y [b y +.5(y Kx) T (y Kx)] } λ a x+ m 2 1 x exp { λ x [b x +.5x T x] }

52 Full conditionals: Full conditionals for ocean formulation π(x ) exp{ 1 2 [λ yx T K T Kx 2λ y x T K T y + λ x x T x]} π(λ x ) λ a x+ m 2 1 x exp { λ x [b x +.5x T x] } π(λ y ) λ a y+ n 2 1 y exp { λ y [b y +.5(y Kx) T (y Kx)] } 56 Gibbs sampler implementation x N((λ y K T K + λ x I m ) 1 λ y K T y, (λ y K T K + λ x I m ) 1 ) x j N λ y rj T k j 1, λ y kj T k j + λ x λ y kj T k j + λ x λ x Γ(a x + m 2,b x +.5x T x) λ y Γ(a y + n 2,b y +.5(y Kx) T (y Kx)) where k j is jth column of K and r j = y j j k j x j.

53 Posterior mean of the space-time temperature field 50N N N N 45N 45N 40N 40N 40N 35N 35N 35N 30N 30N 30N 25N 25N 25N 20N N N W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W 57 50N N N N 45N 45N 40N 40N 40N 35N 35N 35N 30N 30N 30N 25N 25N 25N 20N N N W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W

54 Deviations from time-averaged mean temperature field 50N N N N 45N 45N 40N 40N 40N 35N 35N 35N 30N 30N 30N 25N 25N 25N 20N N N W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W 58 50N N N N 45N 45N 40N 40N 40N 35N 35N 35N 30N 30N 30N 25N 25N 25N 20N N N W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W

55 Posterior probabilities of differing from time-averaged mean field 50N N N N 45N 45N 40N 40N 40N 35N 35N 35N 30N 30N 30N 25N 25N 25N 20N N N W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W 59 50N N N N 45N 45N 40N 40N 40N 35N 35N 35N 30N 30N 30N 25N 25N 25N 20N N N W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W 30W 25W 20W 15W 10W 5W

56 Alternative approaches for building space-time models latent process smoothing kernel space-time field time time space 60 space Define space-time domain S T with T = {1,..., n t } space Discrete knot process x(s, t) on {ω 1,..., ω ns } {1,..., n t }, n t n s = m Define spatial smoothing kernel k(s) Construct space-time process z(s, t) z(s, t) = n s z(s, t) = n s j=1 k(s ω j)x(ω j,t) or with varying kernel j=1 k st(ω j )x jt

57 8 hour max for ozone over summer days in the Eastern US n = 510 ozone measurements each of 30 days ω k s laid out on same hexagonal lattice

58 62 Temporally evolving latent x(s, t) process Times: t T = {1,..., n t }, n t = 30 spatial knot locations: W = {ω 1,..., ω ns }, n s = 27 x(s, t) defined on W T Space-time process z(s, t) obtained by convolving x(s, t) with k(s): z(s, t) = n s j=1 k(ω j s)x(ω j,t) = n s j=1 k s(ω j )x jt Specify locally linear MRF priors for each x j =(x j1,..., x jnt ) T π(x j λ x ) λ n t 2 exp 1 2 λ xx T j Wx j where 1 if i j =1 1 if i = j =1or i = j = n W ij = t 2 if 1 <i= j < n t 0 otherwise So for x =(x 11,x 21..., x ns 1,x 12,..., x ns 2,..., x 1nt,..., x ns n t ) T π(x λ x ) λ n t n s 2 exp 1 2 λ xx T (W I ns )x

59 Formulation for temporally evolving z(s, t) Data: at each time t, observe n-vector y t =(y 1t,..., y nt ) T at sites s 1,..., s n. Likelihood for data observed at time t: L(y t x t, λ y ) λ n 2 y exp 1 2 λ y(y t K t x t ) T (y K t x t ) where K t ij = k(ω j s i ) 63 Define n n t -vector y =(y T 1,..., y T n t ) T Likelihood for entire data y: L(y x, λ y ) λ nn t 2 y where K = diag(k 1,..., K n t). Priors: exp 1 2 λ y(y Kx) T (y Kx) π(x λ x ) λ m 2 x exp 1 2 λ xx T Wx π(λ x ) λ a x 1 x exp{ b x λ x } π(λ y ) λ a y 1 y exp{ b y λ y }

60 Posterior and full conditionals π(x, λ x, λ y y) λ a y+ nn t 2 1 y exp { λ y [b y +.5(y Kx) T (y Kx)] } Full conditionals: λ a x+ n sn t 2 1 x exp { λ x [b x +.5x T Wx] } π(x ) exp{ 1 2 [λ yx T K T Kx 2λ y x T K T y + λ x x T Wx]} π(λ x ) λ a x+ m 2 1 x exp { λ x [b x +.5x T Wx] } π(λ y ) λ a y+ n 2 1 y exp { λ y [b y +.5(y Kx) T (y Kx)] } 64 Gibbs sampler implementation x N((λ y K T K + λ x W ) 1 λ y K T y, (λ y K T K + λ x W ) 1 ) x jt N λ y r T tjk tj + n j x j λ y k T tjk tj + n j λ x, 1 λ y k T j k j + n j λ x λ x Γ(a x + m 2,b x +.5x T x) λ y Γ(a y + n 2,b y +.5(y Kx) T (y Kx)) where k tj is jth column of K t, r tj = y t j j k tj x tj, n j = number of neighbors of x jt, and x jt =mean of neighbors of x jt

61 DLM setup for ozone example Given latent process x t =(x 1,t,..., x 27,t ) T, t =1,..., 30 y t =(y 1t,..., y ny t) T at sites s 1t,..., s ny t y t = K t x t + ɛ t x t = x t 1 + ν t 65 where K t is the n y 27 matrix given by: Kij t = k(s it ω j ),t=1,..., 30, ɛ t ν t i.i.d. N(0, σɛ 2 ), t=1,..., 30, i.i.d. N(0, σν), 2 t=1,..., 30, and x 1 N(0, σ 2 xi 27 ). can use dynamic linear model (DLM)/Kalman filter machinery single site MCMC works too See Stroud et.al. (1999) for alternative model.

62 Posterior mean for first 9 days

63 Posterior mean of selected x j s latent proces values x(s,t) independent over time t (days) latent proces values x(s,t) random walk over time t (days)

64 68 References D. Higdon (1998) A process-convolution approach to modeling temperatures in the North Atlantic Ocean (with discussion), Environmental and Ecological Statistics, 5: J. Stroud, P. Müler and B. Sanso (2001) Dynamic Models For Spatio- Temporal Data, Journal of the Royal Statistical Society, Series B, 63, D. Higdon (2002) Space and space-time modeling using process convolutions, in Quantitative Methods for Current Environmental Issues (C. Anderson and V. Barnett and P. C. Chatwin and A. H. El-Shaarawi, eds), M. West and J. Harrison (1997) Bayesian Forecasting and Dynamic Models (Second Edition), Springer-Verlag.

Efficient Posterior Inference and Prediction of Space-Time Processes Using Dynamic Process Convolutions

Efficient Posterior Inference and Prediction of Space-Time Processes Using Dynamic Process Convolutions Efficient Posterior Inference and Prediction of Space-Time Processes Using Dynamic Process Convolutions Catherine A. Calder Department of Statistics The Ohio State University 1958 Neil Avenue Columbus,

More information

Bayesian Dynamic Linear Modelling for. Complex Computer Models

Bayesian Dynamic Linear Modelling for. Complex Computer Models Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

Flexible Gaussian Processes via Convolution

Flexible Gaussian Processes via Convolution Flexible Gaussian Processes via Convolution Herbert K. H. Lee, Christopher H. Holloman, Catherine A. Calder, and Dave M. Higdon Duke University (and LANL) June 21, 2002 Abstract Spatial and spatio-temporal

More information

Hierarchical Modelling for Multivariate Spatial Data

Hierarchical Modelling for Multivariate Spatial Data Hierarchical Modelling for Multivariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Point-referenced spatial data often come as

More information

Hierarchical Modeling for Multivariate Spatial Data

Hierarchical Modeling for Multivariate Spatial Data Hierarchical Modeling for Multivariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Introduction to Bayesian methods in inverse problems

Introduction to Bayesian methods in inverse problems Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction

More information

Hierarchical Modeling for Univariate Spatial Data

Hierarchical Modeling for Univariate Spatial Data Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

A Framework for Daily Spatio-Temporal Stochastic Weather Simulation

A Framework for Daily Spatio-Temporal Stochastic Weather Simulation A Framework for Daily Spatio-Temporal Stochastic Weather Simulation, Rick Katz, Balaji Rajagopalan Geophysical Statistics Project Institute for Mathematics Applied to Geosciences National Center for Atmospheric

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

An introduction to Bayesian statistics and model calibration and a host of related topics

An introduction to Bayesian statistics and model calibration and a host of related topics An introduction to Bayesian statistics and model calibration and a host of related topics Derek Bingham Statistics and Actuarial Science Simon Fraser University Cast of thousands have participated in the

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

Markov Random Fields

Markov Random Fields Markov Random Fields 1. Markov property The Markov property of a stochastic sequence {X n } n 0 implies that for all n 1, X n is independent of (X k : k / {n 1, n, n + 1}), given (X n 1, X n+1 ). Another

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

Bayesian data analysis in practice: Three simple examples

Bayesian data analysis in practice: Three simple examples Bayesian data analysis in practice: Three simple examples Martin P. Tingley Introduction These notes cover three examples I presented at Climatea on 5 October 0. Matlab code is available by request to

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

Nearest Neighbor Gaussian Processes for Large Spatial Data

Nearest Neighbor Gaussian Processes for Large Spatial Data Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns

More information

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Abhirup Datta 1 Sudipto Banerjee 1 Andrew O. Finley 2 Alan E. Gelfand 3 1 University of Minnesota, Minneapolis,

More information

Nonstationary spatial process modeling Part II Paul D. Sampson --- Catherine Calder Univ of Washington --- Ohio State University

Nonstationary spatial process modeling Part II Paul D. Sampson --- Catherine Calder Univ of Washington --- Ohio State University Nonstationary spatial process modeling Part II Paul D. Sampson --- Catherine Calder Univ of Washington --- Ohio State University this presentation derived from that presented at the Pan-American Advanced

More information

INTRODUCTION TO BAYESIAN STATISTICS

INTRODUCTION TO BAYESIAN STATISTICS INTRODUCTION TO BAYESIAN STATISTICS Sarat C. Dass Department of Statistics & Probability Department of Computer Science & Engineering Michigan State University TOPICS The Bayesian Framework Different Types

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Te Whare Wananga o te Upoko o te Ika a Maui. Computer Science

Te Whare Wananga o te Upoko o te Ika a Maui. Computer Science VICTORIA UNIVERSITY OF WELLINGTON Te Whare Wananga o te Upoko o te Ika a Maui School of Mathematical and Computing Sciences Computer Science Multiple Output Gaussian Process Regression Phillip Boyle and

More information

Likelihood NIPS July 30, Gaussian Process Regression with Student-t. Likelihood. Jarno Vanhatalo, Pasi Jylanki and Aki Vehtari NIPS-2009

Likelihood NIPS July 30, Gaussian Process Regression with Student-t. Likelihood. Jarno Vanhatalo, Pasi Jylanki and Aki Vehtari NIPS-2009 with with July 30, 2010 with 1 2 3 Representation Representation for Distribution Inference for the Augmented Model 4 Approximate Laplacian Approximation Introduction to Laplacian Approximation Laplacian

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

Groundwater permeability

Groundwater permeability Groundwater permeability Easy to solve the forward problem: flow of groundwater given permeability of aquifer Inverse problem: determine permeability from flow (usually of tracers With some models enough

More information

Spatial Dynamic Factor Analysis

Spatial Dynamic Factor Analysis Spatial Dynamic Factor Analysis Esther Salazar Federal University of Rio de Janeiro Department of Statistical Methods Sixth Workshop on BAYESIAN INFERENCE IN STOCHASTIC PROCESSES Bressanone/Brixen, Italy

More information

CS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling

CS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling CS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling Professor Erik Sudderth Brown University Computer Science October 27, 2016 Some figures and materials courtesy

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Surveying the Characteristics of Population Monte Carlo

Surveying the Characteristics of Population Monte Carlo International Research Journal of Applied and Basic Sciences 2013 Available online at www.irjabs.com ISSN 2251-838X / Vol, 7 (9): 522-527 Science Explorer Publications Surveying the Characteristics of

More information

Wrapped Gaussian processes: a short review and some new results

Wrapped Gaussian processes: a short review and some new results Wrapped Gaussian processes: a short review and some new results Giovanna Jona Lasinio 1, Gianluca Mastrantonio 2 and Alan Gelfand 3 1-Università Sapienza di Roma 2- Università RomaTRE 3- Duke University

More information

Physician Performance Assessment / Spatial Inference of Pollutant Concentrations

Physician Performance Assessment / Spatial Inference of Pollutant Concentrations Physician Performance Assessment / Spatial Inference of Pollutant Concentrations Dawn Woodard Operations Research & Information Engineering Cornell University Johns Hopkins Dept. of Biostatistics, April

More information

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Alan Gelfand 1 and Andrew O. Finley 2 1 Department of Statistical Science, Duke University, Durham, North

More information

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018 Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Chapter 5 Markov Chain Monte Carlo MCMC is a kind of improvement of the Monte Carlo method By sampling from a Markov chain whose stationary distribution is the desired sampling distributuion, it is possible

More information

Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models

Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models Kohta Aoki 1 and Hiroshi Nagahashi 2 1 Interdisciplinary Graduate School of Science and Engineering, Tokyo Institute of Technology

More information

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota,

More information

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous

More information

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Andrew O. Finley 1 and Sudipto Banerjee 2 1 Department of Forestry & Department of Geography, Michigan

More information

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

On Bayesian Computation

On Bayesian Computation On Bayesian Computation Michael I. Jordan with Elaine Angelino, Maxim Rabinovich, Martin Wainwright and Yun Yang Previous Work: Information Constraints on Inference Minimize the minimax risk under constraints

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

ST 740: Markov Chain Monte Carlo

ST 740: Markov Chain Monte Carlo ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:

More information

Spatio-temporal precipitation modeling based on time-varying regressions

Spatio-temporal precipitation modeling based on time-varying regressions Spatio-temporal precipitation modeling based on time-varying regressions Oleg Makhnin Department of Mathematics New Mexico Tech Socorro, NM 87801 January 19, 2007 1 Abstract: A time-varying regression

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

On Gaussian Process Models for High-Dimensional Geostatistical Datasets

On Gaussian Process Models for High-Dimensional Geostatistical Datasets On Gaussian Process Models for High-Dimensional Geostatistical Datasets Sudipto Banerjee Joint work with Abhirup Datta, Andrew O. Finley and Alan E. Gelfand University of California, Los Angeles, USA May

More information

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute

More information

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus

More information

An introduction to Sequential Monte Carlo

An introduction to Sequential Monte Carlo An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods

More information

MARKOV CHAIN MONTE CARLO

MARKOV CHAIN MONTE CARLO MARKOV CHAIN MONTE CARLO RYAN WANG Abstract. This paper gives a brief introduction to Markov Chain Monte Carlo methods, which offer a general framework for calculating difficult integrals. We start with

More information

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Andrew O. Finley Department of Forestry & Department of Geography, Michigan State University, Lansing

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

Spatial Inference of Nitrate Concentrations in Groundwater

Spatial Inference of Nitrate Concentrations in Groundwater Spatial Inference of Nitrate Concentrations in Groundwater Dawn Woodard Operations Research & Information Engineering Cornell University joint work with Robert Wolpert, Duke Univ. Dept. of Statistical

More information

Adaptive Monte Carlo methods

Adaptive Monte Carlo methods Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert

More information

Markov random fields. The Markov property

Markov random fields. The Markov property Markov random fields The Markov property Discrete time: (X k X k!1,x k!2,... = (X k X k!1 A time symmetric version: (X k! X!k = (X k X k!1,x k+1 A more general version: Let A be a set of indices >k, B

More information

Markov chain Monte Carlo methods in atmospheric remote sensing

Markov chain Monte Carlo methods in atmospheric remote sensing 1 / 45 Markov chain Monte Carlo methods in atmospheric remote sensing Johanna Tamminen johanna.tamminen@fmi.fi ESA Summer School on Earth System Monitoring and Modeling July 3 Aug 11, 212, Frascati July,

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee September 03 05, 2017 Department of Biostatistics, Fielding School of Public Health, University of California, Los Angeles Linear Regression Linear regression is,

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

A note on Reversible Jump Markov Chain Monte Carlo

A note on Reversible Jump Markov Chain Monte Carlo A note on Reversible Jump Markov Chain Monte Carlo Hedibert Freitas Lopes Graduate School of Business The University of Chicago 5807 South Woodlawn Avenue Chicago, Illinois 60637 February, 1st 2006 1 Introduction

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Parameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets

Parameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets Parameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets Matthias Katzfuß Advisor: Dr. Noel Cressie Department of Statistics The Ohio State University

More information

Gaussian predictive process models for large spatial data sets.

Gaussian predictive process models for large spatial data sets. Gaussian predictive process models for large spatial data sets. Sudipto Banerjee, Alan E. Gelfand, Andrew O. Finley, and Huiyan Sang Presenters: Halley Brantley and Chris Krut September 28, 2015 Overview

More information

Advances and Applications in Perfect Sampling

Advances and Applications in Perfect Sampling and Applications in Perfect Sampling Ph.D. Dissertation Defense Ulrike Schneider advisor: Jem Corcoran May 8, 2003 Department of Applied Mathematics University of Colorado Outline Introduction (1) MCMC

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Contents in latter part Linear Dynamical Systems What is different from HMM? Kalman filter Its strength and limitation Particle Filter

More information

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J. Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

VCMC: Variational Consensus Monte Carlo

VCMC: Variational Consensus Monte Carlo VCMC: Variational Consensus Monte Carlo Maxim Rabinovich, Elaine Angelino, Michael I. Jordan Berkeley Vision and Learning Center September 22, 2015 probabilistic models! sky fog bridge water grass object

More information

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model

More information

Efficient MCMC Sampling for Hierarchical Bayesian Inverse Problems

Efficient MCMC Sampling for Hierarchical Bayesian Inverse Problems Efficient MCMC Sampling for Hierarchical Bayesian Inverse Problems Andrew Brown 1,2, Arvind Saibaba 3, Sarah Vallélian 2,3 CCNS Transition Workshop SAMSI May 5, 2016 Supported by SAMSI Visiting Research

More information

Bayesian Gaussian Process Regression

Bayesian Gaussian Process Regression Bayesian Gaussian Process Regression STAT8810, Fall 2017 M.T. Pratola October 7, 2017 Today Bayesian Gaussian Process Regression Bayesian GP Regression Recall we had observations from our expensive simulator,

More information

16 : Approximate Inference: Markov Chain Monte Carlo

16 : Approximate Inference: Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School

More information

Sub-kilometer-scale space-time stochastic rainfall simulation

Sub-kilometer-scale space-time stochastic rainfall simulation Picture: Huw Alexander Ogilvie Sub-kilometer-scale space-time stochastic rainfall simulation Lionel Benoit (University of Lausanne) Gregoire Mariethoz (University of Lausanne) Denis Allard (INRA Avignon)

More information

Hierarchical Modelling for Univariate and Multivariate Spatial Data

Hierarchical Modelling for Univariate and Multivariate Spatial Data Hierarchical Modelling for Univariate and Multivariate Spatial Data p. 1/4 Hierarchical Modelling for Univariate and Multivariate Spatial Data Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota

More information

Multivariate spatial modeling

Multivariate spatial modeling Multivariate spatial modeling Point-referenced spatial data often come as multivariate measurements at each location Chapter 7: Multivariate Spatial Modeling p. 1/21 Multivariate spatial modeling Point-referenced

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Spatial omain Hierarchical Modelling for Univariate Spatial ata Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A.

More information

Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies. Abstract

Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies. Abstract Bayesian Estimation of A Distance Functional Weight Matrix Model Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies Abstract This paper considers the distance functional weight

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Lecture 8: The Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm 30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:

More information

Monte Carlo Methods. Leon Gu CSD, CMU

Monte Carlo Methods. Leon Gu CSD, CMU Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

Markov Chain Monte Carlo Algorithms for Gaussian Processes

Markov Chain Monte Carlo Algorithms for Gaussian Processes Markov Chain Monte Carlo Algorithms for Gaussian Processes Michalis K. Titsias, Neil Lawrence and Magnus Rattray School of Computer Science University of Manchester June 8 Outline Gaussian Processes Sampling

More information

Brief introduction to Markov Chain Monte Carlo

Brief introduction to Markov Chain Monte Carlo Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical

More information

DAG models and Markov Chain Monte Carlo methods a short overview

DAG models and Markov Chain Monte Carlo methods a short overview DAG models and Markov Chain Monte Carlo methods a short overview Søren Højsgaard Institute of Genetics and Biotechnology University of Aarhus August 18, 2008 Printed: August 18, 2008 File: DAGMC-Lecture.tex

More information

Advanced Sampling Algorithms

Advanced Sampling Algorithms + Advanced Sampling Algorithms + Mobashir Mohammad Hirak Sarkar Parvathy Sudhir Yamilet Serrano Llerena Advanced Sampling Algorithms Aditya Kulkarni Tobias Bertelsen Nirandika Wanigasekara Malay Singh

More information

Kobe University Repository : Kernel

Kobe University Repository : Kernel Kobe University Repository : Kernel タイトル Title 著者 Author(s) 掲載誌 巻号 ページ Citation 刊行日 Issue date 資源タイプ Resource Type 版区分 Resource Version 権利 Rights DOI URL Note on the Sampling Distribution for the Metropolis-

More information

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that

More information

What s for today. Continue to discuss about nonstationary models Moving windows Convolution model Weighted stationary model

What s for today. Continue to discuss about nonstationary models Moving windows Convolution model Weighted stationary model What s for today Continue to discuss about nonstationary models Moving windows Convolution model Weighted stationary model c Mikyoung Jun (Texas A&M) Stat647 Lecture 11 October 2, 2012 1 / 23 Nonstationary

More information

Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood

Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood Kuangyu Wen & Ximing Wu Texas A&M University Info-Metrics Institute Conference: Recent Innovations in Info-Metrics October

More information

Hilbert Space Methods for Reduced-Rank Gaussian Process Regression

Hilbert Space Methods for Reduced-Rank Gaussian Process Regression Hilbert Space Methods for Reduced-Rank Gaussian Process Regression Arno Solin and Simo Särkkä Aalto University, Finland Workshop on Gaussian Process Approximation Copenhagen, Denmark, May 2015 Solin &

More information