Hierarchical matrix techniques for maximum likelihood covariance estimation

Size: px
Start display at page:

Download "Hierarchical matrix techniques for maximum likelihood covariance estimation"

Transcription

1 Hierarchical matrix techniques for maximum likelihood covariance estimation tifica Alexander Litvinenko, Extreme Computing Research Center and Uncertainty Center, KAUST (joint work with M. Genton, Y. Sun and D. Keyes)

2 The structure of the talk 1. Motivation 2. Hierarchical matrices [Hackbusch 1]: 3. Matérn covariance function 4. Uncertain parameters of the covariance function: 4.1 Uncertain covariance length 4.2 Uncertain smoothness parameter 5. Identification of these parameters via maximizing the log-likelihood. 2 / 37

3 Motivation, problem 1 Task: to predict temperature, velocity, salinity, estimate parameters of covariance Grid: 50Mi locations on 50 levels, (X*Y*Z) + X*Y= 500*500* *500 = 50Mi. High-resolution time-dependent data about Red Sea: zonal velocity and temperature 3 / 37

4 Motivation, problem 2 Task: to predict moisture, compute covariance, estimate its parameters 2D-Grid: 2.5Mi locations with 2.1Mi observations and 278K missing values. Soil moisture latitude longitude High-resolution daily soil moisture data at the top layer of the Mississippi basin, U.S.A., (Chaney et al., in review). Important for agriculture, defense. Moisture is very heterogeneous. 4 / 37

5 Motivation, estimation of uncertain parameters cov. length H-matrix rank tion Logo Box-plots Lock-up for l = (domain [0, 1] 2 ) vs different H-matrix ranks k = {3, 7, }. Which H-matrix rank is sufficient for identification of parameters of a particular type of cov. matrix? 5 / 37

6 Motivation for H-matrices General dense matrix requires O(n 3 ) storage and time. Could be expensive! If covariance matrix is structured (diagonal, Toeplitz, circulant) then we can apply e.g. FFT with O(nlogn), but if not? 6 / 37

7 Hierarchical (H)-matrices Introduction into Hierarchical (H)-matrix technique 7 / 37

8 Examples of H-matrix approximations Figure : Three examples of H-matrix approximations, (left-middle) R n n, n = 2 10, of the discretised covariance function cov(x, y) = e r, l 1 = 0.15, l 2 = 0.2, x, y [0, 1] 2 ; (right) Soil moisture from example above with dofs. The biggest dense (dark) blocks R 32 32, max. rank k = 4 on the left, k = in the middle, and k = on the right. 8 / 37

9 Matérn covariance functions Matérn covariance functions C θ = 2σ2 ( r ν ( r ) Kν, θ = (σ Γ(ν) 2l) 2, ν, l). l / 37

10 Examples of Matérn covariance matrices ( ) ( ) 3r 3r C ν=3/2 (r) = 1 + exp l l ( ) ( 5r C ν=5/2 (r) = r ) 2 5r l 3l 2 exp l (1) (2) ν = 1/2 exponential covariance function C ν=1/2 (r) = exp( r), ν Gaussian covariance function C ν= (r) = exp( r 2 ). 10 / 37

11 Identifying uncertain parameters Identifying uncertain parameters 11 / 37

12 Identifying uncertain parameters Given: a vector of measurements z = (z 1,..., z n ) T with a covariance matrix C(θ ) = C(σ 2, ν, l). C θ = 2σ2 ( r ν ( r ) Kν, θ = (σ Γ(ν) 2l) 2, ν, l). l To identify: uncertain parameters (σ 2, ν, l). Plan: Maximize the log-likelihood function L(θ) = 1 ( ) Nlog2π + log det{c(θ)} + z T C(θ) 1 z, 2 tion Logo On each Lock-up iteration i we have a new matrix C(θ i ). 12 / 37

13 Other works 1. S. AMBIKASARAN, et al., Fast direct methods for gaussian processes and the analysis of NASA Kepler mission, arxiv: , (2014). 2. S. AMBIKASARAN, J. Y. LI, P. K. KITANIDIS, AND E. DARVE, Large-scale stochastic linear inversion using hierarchical matrices, Computational Geosciences, (20) 3. J. BALLANI AND D. KRESSNER, Sparse inverse covariance estimation with hierarchical matrices, (2015). 4. M. BEBENDORF, Why approximate LU decompositions of finite element discretizations of elliptic operators can be computed with almost linear complexity, (2007). 5. S. BOERM AND J. GARCKE, Approximating gaussian processes with H2-matrices, J. E. CASTRILLON, M. G. GENTON, AND R. YOKOTA, Multi-Level Restricted Maximum Likelihood Covariance Estimation and Kriging for Large Non-Gridded Spatial Datasets, (2015). 7. J. DOELZ, H. HARBRECHT, AND C. SCHWAB, Covariance regularity and H-matrix approximation for rough random fields, ETH-Zuerich, H. HARBRECHT et al, Efficient approximation of random fields for numerical applications, Numerical Linear Algebra with Applications, (2015).. C.-J. HSIEH, et al, Big QUIC: Sparse inverse covariance estimation for a million variables, J. QUINONERO-CANDELA, et al, A unifying view of sparse approximate gaussian process regression, (2005). 11. A. SAIBABA, S. AMBIKASARAN, J. YUE LI, P. KITANIDIS, AND E. DARVE, Application of hierarchical matrices to linear inverse problems in geostatistics, Oil & Gas Science (2012). / 37

14 Convergence of the optimization method 14 / 37

15 Details of the identification To maximize the log-likelihood function we use the Brent s method [Brent 73] (combining bisection method, secant method and inverse quadratic interpolation). 1. C(θ) C H (θ, k). 2. H-Cholesky: C H (θ, k) = LL T 3. z T C 1 z = z T (LL T ) 1 z = v T v, where v is a solution of L(θ, k)v(θ) := z(θ ). 4. Let λ i be diagonal elements of H-Cholesky factor L, then log det{c} = log det{ll T } = log det{ n λ 2 i } = 2 i=1 n logλ i, L(θ, k) = N N 2 log(2π) log{ L ii (θ, k)} 1 2 (v(θ)t v(θ)). (3) i=1 i=1 15 / 37

16 Log likelihood(θ ) Shape of Log likelihood(θ) log(det(c)) z T C 1 z Log likelihood parameter θ, truth θ*=12 Figure : Minimum of negative log-likelihood (black) is at θ = (,, l) 12 (σ 2 and ν are fixed) 16 / 37

17 Convergence of H-matrix approximations 0 5 Spectral norm, L=0.1, nu=1 Frob. norm, L=0.1 Spectral norm, L=0.2 Frob. norm, L=0.2 Spectral norm, L=0.5 Frob. norm, L= Spectral norm, L=0.1, nu=0.5 Frob. norm, L=0.1 Spectral norm, L=0.2 Frob. norm, L=0.2 Spectral norm, L=0.5 Frob. norm, L=0.5 log(rel. error) log(rel. error) rank k rank k ν = 1(left) and ν = 0.5 (right) for different cov. lengths l = {0.1, 02, 0.5} 17 / 37

18 Convergence of H-matrix approximations 0 2 Spectral norm, nu=1.5, L=0.1 Spectral norm, nu=1 Spectral norm, nu= Spectral norm, nu=1.5, L=0.5 Spectral norm, nu=1 Spectral norm, nu= log(rel. error) log(rel. error) rank k rank k ν = {1.5, 1, 0.5}, l = 0.1 (left) and ν = {1.5, 1, 0.5}, l = 0.5 (right) 18 / 37

19 What will change? Approximate C by C H 1. How the eigenvalues of C and C H differ? 2. How det(c) differs from det(c H )? [Below] 3. How L differs from L H? [Mario Bebendorf et al] 4. How C 1 differs from (C H ) 1? [Mario Bebendorf et al] 5. How L(θ, k) differs from L(θ)? [Below] 6. What is optimal H-matrix rank? [Below] 7. How θ H differs from θ? [Below] For theory, estimates for the rank and accuracy see works of Bebendorf, Grasedyck, Le Borne, Hackbusch,... 1 / 37

20 Remark For a small H-matrix rank k the H-matrix Cholesky of C H crashes when eigenvalues of C come very close to zero. A remedy is to increase the rank k. In our example for n = 65 2 we increased k from 7 to. To avoid this instability, we can modify C H m = C H + δ 2 I. Assume λ i are eigenvalues of C H. Then eigenvalues of C H m will be λ i + δ 2. n log det(cm H ) = log (λ i + δ 2 ) = i=1 n log(λ i + δ 2 ). (4) i=1 20 / 37

21 Error analysis Theorem (Existence of H-matrix inverse in [Bebendorf 11, Ballani, Kressner 14) Under certain conditions an H-matrix inverse exist C 1 H C 1 ε C 1, (5) theoretical estimations for rank k inv of C 1 H Theorem (Error in log det) are given. Let E := C C H, (C H ) 1 E := (C H ) 1 C I and for the spectral radius ρ((c H ) 1 E) = ρ((c H ) 1 C I) ε. (6) Then log det(c) log det(c H ) plog(1 ε). Proof: See [Ballani, Kressner 14], [Ipsen 05]. 21 / 37

22 How sensitive is Log-Likelihood to the H-matrix rank? It is not at all sensible. H-matrix approximation changes function L(θ, k) and estimation of θ very-very small. θ L(exact) L(7) L(20) Comparison of three likelihood functions, computed with different tion Logo H-matrix Lock-up ranks: exact, H-rank 7, H-rank 20. Exponential covariance function, with covariance length l = 0., domain G = [0, 1] / 37

23 How sensitive is Log-Likelihood to the H-matrix rank? loglikelihood θ Figure : Three negative log-likelihood functions: exact, commuted with H-matrix rank 7 and 17. One can see that even with rank 7 one can achieve very accurate results. 23 / 37

24 Do we need all measurements? Boxplots vs n cov. length number of measurements Moisture data. Boxplots with increasing of number of measurements, n = {1000,..., 32000}. The mean and median are obtained after averaging 100 simulations. 24 / 37

25 Decreasing of error bars with increasing number of measurements Error bars (mean +/- st. dev.) computed for different n. Decreasing of error bars with increasing of number of measurements/dimension, n = {17 2, 33 2, 65 2 }. The mean and median are obtained after averaging 200 simulations. 25 / 37

26 H-matrix approximation is robust w.r.t. parameter ν Figure : Dependence of H-matrix approximation error on parameter ν. Relative error C C H 2 C H 2 via smoothness parameter ν. H-matrix rank k = 8, n = 16641, Matern covariance matrix. 26 / 37

27 H-matrix approximation is robust w.r.t. cov. length l Figure : Dependence of H-matrix approximation error on cov. length l. Relative error C C H 2 C H 2 via covariance length l. H-matrix rank k = 8, n = 16641, Matern covariance matrix. 27 / 37

28 log-determinant of C rank5 ranks=3,4,5 -log-likelihood cov. length cov. length H-matrix approximation of log det(c) (left) and of the tion Logo log-likelihood Lock-up L (right); σ 2 = 1, ν = 0.5, k = 5. The red line - k = 5 and the blue - k = 3 on [0.01, 0.3], k = 4 on [0.3, 0.6] and k = 5 on [0.6, 1]. The rank k = 3 is sufficient to approximate C, but insufficient to approximate C 1 on the whole interval [0.01, 1]. The first numerical instability appears at point l i1 0.3, to avoid it the rank k is increased by 1, until the second instability appears at point l i / 37

29 10 8 log-likelihood 10 7 z T C -1 z parameter \nu H-matrix approx. of z T C 1 z and of log-likelihood L; σ 2 = 1, ν = 0.5, k increases from 5 until 12 after each jump. The red line - rank k = 5 and blue - k = 5 on [0.1, 1.42]; k = 6 on [1.42, 1.57]; k = 7 on [1.57, 1.3]; k = 8 on [1.3, 2.24] and k = {, 10, 12} between [2.24, 3.14]. k has to be increased to approximate C 1. The first numerical instability appears at point ν i1 1.42, to avoid it the rank k is increased by 1, until the second instability appears at point l i etc. 2 / 37

30 Time profiling, C-language 30 / 37

31 Parallel implementation with HLIBpro (R. Kriemann) We used to setup the exponential covariance matrix (cov. length=1), to compute Cholesky and the inverse. We used adaptive rank arithmetics with ε = 1e 4 for each block of C H and ε = 1e 8 for each block of C H. Number of processing cores is 40. We took moisture data (see above) with N points. N compute C H LL T inverse Compr. time size time size ε 1 time size ε 2 rate % sec. MB sec. MB sec. MB % e e % e e-1 Table : Here ε 1 := I (LL T ) 1 C 2, where L and C are H-matrices, I -identity matrix; ε 2 := I BC 2, where B is an H-matrix approximation of C / 37

32 Take into account the gradient L(θ i ) θ i = 1 ( ) 2 tr 1 C C 1 θ i 2 zt 1 C C C 1 z. (7) θ i For an exponential random field, have ( C(θ i ) = θ i l exp ) ( x y 2 = x y 2 ) x y l l 2 exp 2 l (8) C(θ i ) =: C 2 θ i L(θ i ) = 1 θ i 2 tr ( C 1 ) 1 C 2 2 zt C 1 C 2 C 1 z. 32 / 37

33 Conclusion Covariance matrices can be approximated in H-matrix format. Hypotes: H-matrix approximation is robust w.r.t. ν and l. Influence of H-matrix approximation error on the estimated parameters is small. With application of H-matrices we extend the class of covariance functions to work with, allows non-regular discretization of the covariance function on large spatial grids. With the maximizing algorithm we are able to identify both tion Logo parameters: Lock-up covariance lengths l and the smoothness ν 33 / 37

34 Future plans Parallel H-Cholesky for very large covariance matrices on non-regular grids Preconditioning for log-likelihood to decrease cond(c) Domain decomposition for large domains + H-matrix in each sub-domain Apply H-matrices for 1. Kriging estimate ŝ := C sy Cyy 1 y 2. Estimation of variance ˆσ, is the diagonal of conditional cov. matrix C ss y = diag ( ) C ss C sy Cyy 1 C ys, 3. Gestatistical optimal design ϕ A := n 1 tracec ss y, ϕ C := c ( ) T C ss C sy Cyy 1 C ys c, tion Logo To Lock-up implement gradient-based version Compare with the Bayesian Update (H. Matthies, H. Najm, K. Law, A. Stuart et al) 34 / 37

35 Literature 1. Application of hierarchical matrices for computing the Karhunen-Loéve expansion, B.N. Khoromskij, A. Litvinenko, H.G. Matthies, Computing 84 (1-2), 4-67, 31, Parameter identification in a probabilistic setting, B.V. Rosić, A. Kucerová, J Sýkora, O. Pajonk, A. Litvinenko, H.G. Matthies, Engineering Structures 50, 17-16, Methods for statistical data analysis with decision trees, V. Berikov, A. Litvinenko, Novosibirsk, Sobolev Institute of Mathematics, Parametric and uncertainty computations with tensor product representations, H.G. Matthies, A. Litvinenko, O. Pajonk, B.V. Rosić, E. Zander, Uncertainty in Scientific Computing, -150, Data sparse computation of the Karhunen-Loève expansion, B.N. Khoromskij, A. Litvinenko, AIP Conference Proceedings 1048 (1), 311, Kriging and spatial design accelerated by orders of magnitude: Combining low-rank covariance approximations with FFT-techniques W. Nowak, A. Litvinenko, Mathematical Geosciences 45 (4), , / 37

36 Acknowledgement 1. Lars Grasedyck (RWTH Aachen) and Steffen Boerm (Uni Kiel) for HLIB ( 2. Ronald Kriemann (MPI Leipzig) for 3. KAUST Research Computing group, KAUST Supercomputing Lab (KSL) 36 / 37

37 Matérn Fields (Whittle, 63) Taken from D. Simpson (see also Finn Lindgren, Havard Rue, David Bolin,...) Theorem The covariance function of a Matérn field c(x, y) = 1 Γ(ν + d/2)(4π) d/2 κ 2ν 2 ν 1 (κ x y )ν K ν (κ x y ) is the Green s function of the differential operator () L 2 ν = ( κ 2 ) ν+d/2. (10) 37 / 37

Uncertainty Quantification and related areas - An overview

Uncertainty Quantification and related areas - An overview Uncertainty Quantification and related areas - An overview Alexander Litvinenko, talk given at Department of Mathematical Sciences, Durham University Bayesian Computational Statistics & Modeling, KAUST

More information

Lecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion:

Lecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion: tifica Lecture 1: Computation of Karhunen-Loeve Expansion: Alexander Litvinenko http://sri-uq.kaust.edu.sa/ Stochastic PDEs We consider div(κ(x, ω) u) = f (x, ω) in G, u = 0 on G, with stochastic coefficients

More information

NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE

NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE tifica NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE Alexander Litvinenko 1, Hermann G. Matthies 2, Elmar Zander 2 http://sri-uq.kaust.edu.sa/ 1 Extreme Computing Research Center, KAUST, 2 Institute of Scientific

More information

Sampling and low-rank tensor approximation of the response surface

Sampling and low-rank tensor approximation of the response surface Sampling and low-rank tensor approximation of the response surface tifica Alexander Litvinenko 1,2 (joint work with Hermann G. Matthies 3 ) 1 Group of Raul Tempone, SRI UQ, and 2 Group of David Keyes,

More information

Lecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko

Lecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko tifica Lecture 1: Introduction to low-rank tensor representation/approximation Alexander Litvinenko http://sri-uq.kaust.edu.sa/ KAUST Figure : KAUST campus, 5 years old, approx. 7000 people (include 1400

More information

Lecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko

Lecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko tifica Lecture 1: Introduction to low-rank tensor representation/approximation Alexander Litvinenko http://sri-uq.kaust.edu.sa/ KAUST Figure : KAUST campus, 5 years old, approx. 7000 people (include 1400

More information

Overview of Spatial Statistics with Applications to fmri

Overview of Spatial Statistics with Applications to fmri with Applications to fmri School of Mathematics & Statistics Newcastle University April 8 th, 2016 Outline Why spatial statistics? Basic results Nonstationary models Inference for large data sets An example

More information

Multivariate Gaussian Random Fields with SPDEs

Multivariate Gaussian Random Fields with SPDEs Multivariate Gaussian Random Fields with SPDEs Xiangping Hu Daniel Simpson, Finn Lindgren and Håvard Rue Department of Mathematics, University of Oslo PASI, 214 Outline The Matérn covariance function and

More information

A Spectral Approach to Linear Bayesian Updating

A Spectral Approach to Linear Bayesian Updating A Spectral Approach to Linear Bayesian Updating Oliver Pajonk 1,2, Bojana V. Rosic 1, Alexander Litvinenko 1, and Hermann G. Matthies 1 1 Institute of Scientific Computing, TU Braunschweig, Germany 2 SPT

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

Hierarchical Low Rank Approximation of Likelihoods for Large Spatial Datasets

Hierarchical Low Rank Approximation of Likelihoods for Large Spatial Datasets Hierarchical Low Rank Approximation of Likelihoods for Large Spatial Datasets Huang Huang and Ying Sun CEMSE Division, King Abdullah University of Science and Technology, Thuwal, 23955, Saudi Arabia. July

More information

LA Support for Scalable Kernel Methods. David Bindel 29 Sep 2018

LA Support for Scalable Kernel Methods. David Bindel 29 Sep 2018 LA Support for Scalable Kernel Methods David Bindel 29 Sep 2018 Collaborators Kun Dong (Cornell CAM) David Eriksson (Cornell CAM) Jake Gardner (Cornell CS) Eric Lee (Cornell CS) Hannes Nickisch (Phillips

More information

Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques

Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Institut für Numerische Mathematik und Optimierung Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Oliver Ernst Computational Methods with Applications Harrachov, CR,

More information

The Bayesian approach to inverse problems

The Bayesian approach to inverse problems The Bayesian approach to inverse problems Youssef Marzouk Department of Aeronautics and Astronautics Center for Computational Engineering Massachusetts Institute of Technology ymarz@mit.edu, http://uqgroup.mit.edu

More information

Sampling and Low-Rank Tensor Approximations

Sampling and Low-Rank Tensor Approximations Sampling and Low-Rank Tensor Approximations Hermann G. Matthies Alexander Litvinenko, Tarek A. El-Moshely +, Brunswick, Germany + MIT, Cambridge, MA, USA wire@tu-bs.de http://www.wire.tu-bs.de $Id: 2_Sydney-MCQMC.tex,v.3

More information

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Abhirup Datta 1 Sudipto Banerjee 1 Andrew O. Finley 2 Alan E. Gelfand 3 1 University of Minnesota, Minneapolis,

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Data are collected along transect lines, with dense data along. Spatial modelling using GMRFs. 200m. Today? Spatial modelling using GMRFs

Data are collected along transect lines, with dense data along. Spatial modelling using GMRFs. 200m. Today? Spatial modelling using GMRFs Centre for Mathematical Sciences Lund University Engineering geology Lund University Results A non-stationary extension The model Estimation Gaussian Markov random fields Basics Approximating Mate rn covariances

More information

Fast Direct Methods for Gaussian Processes

Fast Direct Methods for Gaussian Processes Fast Direct Methods for Gaussian Processes Mike O Neil Departments of Mathematics New York University oneil@cims.nyu.edu December 12, 2015 1 Collaborators This is joint work with: Siva Ambikasaran Dan

More information

Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes

Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes TTU, October 26, 2012 p. 1/3 Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes Hao Zhang Department of Statistics Department of Forestry and Natural Resources Purdue University

More information

Hierarchical Modeling for Univariate Spatial Data

Hierarchical Modeling for Univariate Spatial Data Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This

More information

Parametric Problems, Stochastics, and Identification

Parametric Problems, Stochastics, and Identification Parametric Problems, Stochastics, and Identification Hermann G. Matthies a B. Rosić ab, O. Pajonk ac, A. Litvinenko a a, b University of Kragujevac c SPT Group, Hamburg wire@tu-bs.de http://www.wire.tu-bs.de

More information

Solving Corrupted Quadratic Equations, Provably

Solving Corrupted Quadratic Equations, Provably Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin

More information

Bayesian Statistics and Data Assimilation. Jonathan Stroud. Department of Statistics The George Washington University

Bayesian Statistics and Data Assimilation. Jonathan Stroud. Department of Statistics The George Washington University Bayesian Statistics and Data Assimilation Jonathan Stroud Department of Statistics The George Washington University 1 Outline Motivation Bayesian Statistics Parameter Estimation in Data Assimilation Combined

More information

Quasi-optimal and adaptive sparse grids with control variates for PDEs with random diffusion coefficient

Quasi-optimal and adaptive sparse grids with control variates for PDEs with random diffusion coefficient Quasi-optimal and adaptive sparse grids with control variates for PDEs with random diffusion coefficient F. Nobile, L. Tamellini, R. Tempone, F. Tesei CSQI - MATHICSE, EPFL, Switzerland Dipartimento di

More information

Hilbert Space Methods for Reduced-Rank Gaussian Process Regression

Hilbert Space Methods for Reduced-Rank Gaussian Process Regression Hilbert Space Methods for Reduced-Rank Gaussian Process Regression Arno Solin and Simo Särkkä Aalto University, Finland Workshop on Gaussian Process Approximation Copenhagen, Denmark, May 2015 Solin &

More information

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Vahid Dehdari and Clayton V. Deutsch Geostatistical modeling involves many variables and many locations.

More information

Theory and Computation for Gaussian Processes

Theory and Computation for Gaussian Processes University of Chicago IPAM, February 2015 Funders & Collaborators US Department of Energy, US National Science Foundation (STATMOS) Mihai Anitescu, Jie Chen, Ying Sun Gaussian processes A process Z on

More information

Kernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.

Kernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt. SINGAPORE SHANGHAI Vol TAIPEI - Interdisciplinary Mathematical Sciences 19 Kernel-based Approximation Methods using MATLAB Gregory Fasshauer Illinois Institute of Technology, USA Michael McCourt University

More information

Scalable kernel methods and their use in black-box optimization

Scalable kernel methods and their use in black-box optimization with derivatives Scalable kernel methods and their use in black-box optimization David Eriksson Center for Applied Mathematics Cornell University dme65@cornell.edu November 9, 2018 1 2 3 4 1/37 with derivatives

More information

Karhunen-Loève decomposition of Gaussian measures on Banach spaces

Karhunen-Loève decomposition of Gaussian measures on Banach spaces Karhunen-Loève decomposition of Gaussian measures on Banach spaces Jean-Charles Croix jean-charles.croix@emse.fr Génie Mathématique et Industriel (GMI) First workshop on Gaussian processes at Saint-Etienne

More information

Gradient Enhanced Universal Kriging Model for Uncertainty Propagation in Nuclear Engineering

Gradient Enhanced Universal Kriging Model for Uncertainty Propagation in Nuclear Engineering Gradient Enhanced Universal Kriging Model for Uncertainty Propagation in Nuclear Engineering Brian A. Lockwood 1 and Mihai Anitescu 2 1 Department of Mechanical Engineering University of Wyoming 2 Mathematics

More information

Contents. Preface to the Third Edition (2007) Preface to the Second Edition (1992) Preface to the First Edition (1985) License and Legal Information

Contents. Preface to the Third Edition (2007) Preface to the Second Edition (1992) Preface to the First Edition (1985) License and Legal Information Contents Preface to the Third Edition (2007) Preface to the Second Edition (1992) Preface to the First Edition (1985) License and Legal Information xi xiv xvii xix 1 Preliminaries 1 1.0 Introduction.............................

More information

Computation fundamentals of discrete GMRF representations of continuous domain spatial models

Computation fundamentals of discrete GMRF representations of continuous domain spatial models Computation fundamentals of discrete GMRF representations of continuous domain spatial models Finn Lindgren September 23 2015 v0.2.2 Abstract The fundamental formulas and algorithms for Bayesian spatial

More information

Monte Carlo Methods for Uncertainty Quantification

Monte Carlo Methods for Uncertainty Quantification Monte Carlo Methods for Uncertainty Quantification Mike Giles Mathematical Institute, University of Oxford Contemporary Numerical Techniques Mike Giles (Oxford) Monte Carlo methods 1 / 23 Lecture outline

More information

On Gaussian Process Models for High-Dimensional Geostatistical Datasets

On Gaussian Process Models for High-Dimensional Geostatistical Datasets On Gaussian Process Models for High-Dimensional Geostatistical Datasets Sudipto Banerjee Joint work with Abhirup Datta, Andrew O. Finley and Alan E. Gelfand University of California, Los Angeles, USA May

More information

Non-Intrusive Solution of Stochastic and Parametric Equations

Non-Intrusive Solution of Stochastic and Parametric Equations Non-Intrusive Solution of Stochastic and Parametric Equations Hermann G. Matthies a Loïc Giraldi b, Alexander Litvinenko c, Dishi Liu d, and Anthony Nouy b a,, Brunswick, Germany b École Centrale de Nantes,

More information

Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields

Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields 1 Introduction Jo Eidsvik Department of Mathematical Sciences, NTNU, Norway. (joeid@math.ntnu.no) February

More information

Gaussian Processes 1. Schedule

Gaussian Processes 1. Schedule 1 Schedule 17 Jan: Gaussian processes (Jo Eidsvik) 24 Jan: Hands-on project on Gaussian processes (Team effort, work in groups) 31 Jan: Latent Gaussian models and INLA (Jo Eidsvik) 7 Feb: Hands-on project

More information

Fixed-domain Asymptotics of Covariance Matrices and Preconditioning

Fixed-domain Asymptotics of Covariance Matrices and Preconditioning Fixed-domain Asymptotics of Covariance Matrices and Preconditioning Jie Chen IBM Thomas J. Watson Research Center Presented at Preconditioning Conference, August 1, 2017 Jie Chen (IBM Research) Covariance

More information

1. Structured representation of high-order tensors revisited. 2. Multi-linear algebra (MLA) with Kronecker-product data.

1. Structured representation of high-order tensors revisited. 2. Multi-linear algebra (MLA) with Kronecker-product data. Lect. 4. Toward MLA in tensor-product formats B. Khoromskij, Leipzig 2007(L4) 1 Contents of Lecture 4 1. Structured representation of high-order tensors revisited. - Tucker model. - Canonical (PARAFAC)

More information

A Framework for Daily Spatio-Temporal Stochastic Weather Simulation

A Framework for Daily Spatio-Temporal Stochastic Weather Simulation A Framework for Daily Spatio-Temporal Stochastic Weather Simulation, Rick Katz, Balaji Rajagopalan Geophysical Statistics Project Institute for Mathematics Applied to Geosciences National Center for Atmospheric

More information

ECE521 lecture 4: 19 January Optimization, MLE, regularization

ECE521 lecture 4: 19 January Optimization, MLE, regularization ECE521 lecture 4: 19 January 2017 Optimization, MLE, regularization First four lectures Lectures 1 and 2: Intro to ML Probability review Types of loss functions and algorithms Lecture 3: KNN Convexity

More information

An Inversion-Free Estimating Equations Approach for. Gaussian Process Models

An Inversion-Free Estimating Equations Approach for. Gaussian Process Models An Inversion-Free Estimating Equations Approach for Gaussian Process Models Mihai Anitescu Jie Chen Michael L. Stein November 29, 2015 Abstract One of the scalability bottlenecks for the large-scale usage

More information

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling François Bachoc former PhD advisor: Josselin Garnier former CEA advisor: Jean-Marc Martinez Department

More information

Curriculum Vitae. Alexander Litvenenko

Curriculum Vitae. Alexander Litvenenko Curriculum Vitae Alexander Litvinenko https://ecrc.kaust.edu.sa/pages/litvinenko.aspx http://stochastic_numerics.kaust.edu.sa/pages/litvinenkoalexander.aspx Name: Nationality: Languages: Alexander Litvenenko

More information

Adaptive Sampling of Clouds with a Fleet of UAVs: Improving Gaussian Process Regression by Including Prior Knowledge

Adaptive Sampling of Clouds with a Fleet of UAVs: Improving Gaussian Process Regression by Including Prior Knowledge Master s Thesis Presentation Adaptive Sampling of Clouds with a Fleet of UAVs: Improving Gaussian Process Regression by Including Prior Knowledge Diego Selle (RIS @ LAAS-CNRS, RT-TUM) Master s Thesis Presentation

More information

Hierarchical Parallel Solution of Stochastic Systems

Hierarchical Parallel Solution of Stochastic Systems Hierarchical Parallel Solution of Stochastic Systems Second M.I.T. Conference on Computational Fluid and Solid Mechanics Contents: Simple Model of Stochastic Flow Stochastic Galerkin Scheme Resulting Equations

More information

Latent Gaussian Processes and Stochastic Partial Differential Equations

Latent Gaussian Processes and Stochastic Partial Differential Equations Latent Gaussian Processes and Stochastic Partial Differential Equations Johan Lindström 1 1 Centre for Mathematical Sciences Lund University Lund 2016-02-04 Johan Lindström - johanl@maths.lth.se Gaussian

More information

Fast matrix algebra for dense matrices with rank-deficient off-diagonal blocks

Fast matrix algebra for dense matrices with rank-deficient off-diagonal blocks CHAPTER 2 Fast matrix algebra for dense matrices with rank-deficient off-diagonal blocks Chapter summary: The chapter describes techniques for rapidly performing algebraic operations on dense matrices

More information

State Space Representation of Gaussian Processes

State Space Representation of Gaussian Processes State Space Representation of Gaussian Processes Simo Särkkä Department of Biomedical Engineering and Computational Science (BECS) Aalto University, Espoo, Finland June 12th, 2013 Simo Särkkä (Aalto University)

More information

Some Areas of Recent Research

Some Areas of Recent Research University of Chicago Department Retreat, October 2012 Funders & Collaborators NSF (STATMOS), US Department of Energy Faculty: Mihai Anitescu, Liz Moyer Postdocs: Jie Chen, Bill Leeds, Ying Sun Grad students:

More information

Hierarchical Matrices. Jon Cockayne April 18, 2017

Hierarchical Matrices. Jon Cockayne April 18, 2017 Hierarchical Matrices Jon Cockayne April 18, 2017 1 Sources Introduction to Hierarchical Matrices with Applications [Börm et al., 2003] 2 Sources Introduction to Hierarchical Matrices with Applications

More information

K-Means and Gaussian Mixture Models

K-Means and Gaussian Mixture Models K-Means and Gaussian Mixture Models David Rosenberg New York University October 29, 2016 David Rosenberg (New York University) DS-GA 1003 October 29, 2016 1 / 42 K-Means Clustering K-Means Clustering David

More information

H 2 -matrices with adaptive bases

H 2 -matrices with adaptive bases 1 H 2 -matrices with adaptive bases Steffen Börm MPI für Mathematik in den Naturwissenschaften Inselstraße 22 26, 04103 Leipzig http://www.mis.mpg.de/ Problem 2 Goal: Treat certain large dense matrices

More information

Sub-kilometer-scale space-time stochastic rainfall simulation

Sub-kilometer-scale space-time stochastic rainfall simulation Picture: Huw Alexander Ogilvie Sub-kilometer-scale space-time stochastic rainfall simulation Lionel Benoit (University of Lausanne) Gregoire Mariethoz (University of Lausanne) Denis Allard (INRA Avignon)

More information

Covariance Matrix Simplification For Efficient Uncertainty Management

Covariance Matrix Simplification For Efficient Uncertainty Management PASEO MaxEnt 2007 Covariance Matrix Simplification For Efficient Uncertainty Management André Jalobeanu, Jorge A. Gutiérrez PASEO Research Group LSIIT (CNRS/ Univ. Strasbourg) - Illkirch, France *part

More information

Nearest Neighbor Gaussian Processes for Large Spatial Data

Nearest Neighbor Gaussian Processes for Large Spatial Data Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns

More information

Gaussian with mean ( µ ) and standard deviation ( σ)

Gaussian with mean ( µ ) and standard deviation ( σ) Slide from Pieter Abbeel Gaussian with mean ( µ ) and standard deviation ( σ) 10/6/16 CSE-571: Robotics X ~ N( µ, σ ) Y ~ N( aµ + b, a σ ) Y = ax + b + + + + 1 1 1 1 1 1 1 1 1 1, ~ ) ( ) ( ), ( ~ ), (

More information

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means

More information

Multi-resolution models for large data sets

Multi-resolution models for large data sets Multi-resolution models for large data sets Douglas Nychka, National Center for Atmospheric Research National Science Foundation NORDSTAT, Umeå, June, 2012 Credits Steve Sain, NCAR Tia LeRud, UC Davis

More information

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means

More information

Learning the Linear Dynamical System with ASOS ( Approximated Second-Order Statistics )

Learning the Linear Dynamical System with ASOS ( Approximated Second-Order Statistics ) Learning the Linear Dynamical System with ASOS ( Approximated Second-Order Statistics ) James Martens University of Toronto June 24, 2010 Computer Science UNIVERSITY OF TORONTO James Martens (U of T) Learning

More information

Fundamentals of Data Assimila1on

Fundamentals of Data Assimila1on 014 GSI Community Tutorial NCAR Foothills Campus, Boulder, CO July 14-16, 014 Fundamentals of Data Assimila1on Milija Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University

More information

Preface to Second Edition... vii. Preface to First Edition...

Preface to Second Edition... vii. Preface to First Edition... Contents Preface to Second Edition..................................... vii Preface to First Edition....................................... ix Part I Linear Algebra 1 Basic Vector/Matrix Structure and

More information

Optimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering

Optimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering Optimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering Hermann G. Matthies Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de

More information

Stochastic Spectral Approaches to Bayesian Inference

Stochastic Spectral Approaches to Bayesian Inference Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to

More information

Multivariate modelling and efficient estimation of Gaussian random fields with application to roller data

Multivariate modelling and efficient estimation of Gaussian random fields with application to roller data Multivariate modelling and efficient estimation of Gaussian random fields with application to roller data Reinhard Furrer, UZH PASI, Búzios, 14-06-25 NZZ.ch Motivation Microarray data: construct alternative

More information

Principal Component Analysis CS498

Principal Component Analysis CS498 Principal Component Analysis CS498 Today s lecture Adaptive Feature Extraction Principal Component Analysis How, why, when, which A dual goal Find a good representation The features part Reduce redundancy

More information

Hands-on Matrix Algebra Using R

Hands-on Matrix Algebra Using R Preface vii 1. R Preliminaries 1 1.1 Matrix Defined, Deeper Understanding Using Software.. 1 1.2 Introduction, Why R?.................... 2 1.3 Obtaining R.......................... 4 1.4 Reference Manuals

More information

Platzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen

Platzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen Platzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen PARAMetric UNCertainties, Budapest STOCHASTIC PROCESSES AND FIELDS Noémi Friedman Institut für Wissenschaftliches Rechnen, wire@tu-bs.de

More information

Multilevel accelerated quadrature for elliptic PDEs with random diffusion. Helmut Harbrecht Mathematisches Institut Universität Basel Switzerland

Multilevel accelerated quadrature for elliptic PDEs with random diffusion. Helmut Harbrecht Mathematisches Institut Universität Basel Switzerland Multilevel accelerated quadrature for elliptic PDEs with random diffusion Mathematisches Institut Universität Basel Switzerland Overview Computation of the Karhunen-Loéve expansion Elliptic PDE with uniformly

More information

Frequentist-Bayesian Model Comparisons: A Simple Example

Frequentist-Bayesian Model Comparisons: A Simple Example Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Spatial omain Hierarchical Modelling for Univariate Spatial ata Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A.

More information

Approximating the Covariance Matrix with Low-rank Perturbations

Approximating the Covariance Matrix with Low-rank Perturbations Approximating the Covariance Matrix with Low-rank Perturbations Malik Magdon-Ismail and Jonathan T. Purnell Department of Computer Science Rensselaer Polytechnic Institute Troy, NY 12180 {magdon,purnej}@cs.rpi.edu

More information

Statistics & Data Sciences: First Year Prelim Exam May 2018

Statistics & Data Sciences: First Year Prelim Exam May 2018 Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book

More information

Non-linear least squares

Non-linear least squares Non-linear least squares Concept of non-linear least squares We have extensively studied linear least squares or linear regression. We see that there is a unique regression line that can be determined

More information

Circulant embedding of approximate covariances for inference from Gaussian data on large lattices

Circulant embedding of approximate covariances for inference from Gaussian data on large lattices Circulant embedding of approximate covariances for inference from Gaussian data on large lattices Joseph Guinness and Montserrat Fuentes Abstract Recently proposed computationally efficient Markov chain

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Douglas Nychka, Soutir Bandyopadhyay, Dorit Hammerling, Finn Lindgren, and Stephan Sain. October 10, 2012

Douglas Nychka, Soutir Bandyopadhyay, Dorit Hammerling, Finn Lindgren, and Stephan Sain. October 10, 2012 A multi-resolution Gaussian process model for the analysis of large spatial data sets. Douglas Nychka, Soutir Bandyopadhyay, Dorit Hammerling, Finn Lindgren, and Stephan Sain October 10, 2012 Abstract

More information

Cross-covariance Functions for Tangent Vector Fields on the Sphere

Cross-covariance Functions for Tangent Vector Fields on the Sphere Cross-covariance Functions for Tangent Vector Fields on the Sphere Minjie Fan 1 Tomoko Matsuo 2 1 Department of Statistics University of California, Davis 2 Cooperative Institute for Research in Environmental

More information

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,

More information

Lecture 5. Gaussian Models - Part 1. Luigi Freda. ALCOR Lab DIAG University of Rome La Sapienza. November 29, 2016

Lecture 5. Gaussian Models - Part 1. Luigi Freda. ALCOR Lab DIAG University of Rome La Sapienza. November 29, 2016 Lecture 5 Gaussian Models - Part 1 Luigi Freda ALCOR Lab DIAG University of Rome La Sapienza November 29, 2016 Luigi Freda ( La Sapienza University) Lecture 5 November 29, 2016 1 / 42 Outline 1 Basics

More information

Gaussian processes for spatial modelling in environmental health: parameterizing for flexibility vs. computational efficiency

Gaussian processes for spatial modelling in environmental health: parameterizing for flexibility vs. computational efficiency Gaussian processes for spatial modelling in environmental health: parameterizing for flexibility vs. computational efficiency Chris Paciorek March 11, 2005 Department of Biostatistics Harvard School of

More information

Uncertainty Quantification in Discrete Fracture Network Models

Uncertainty Quantification in Discrete Fracture Network Models Uncertainty Quantification in Discrete Fracture Network Models Claudio Canuto Dipartimento di Scienze Matematiche, Politecnico di Torino claudio.canuto@polito.it Joint work with Stefano Berrone, Sandra

More information

Gaussian Process Regression

Gaussian Process Regression Gaussian Process Regression 4F1 Pattern Recognition, 21 Carl Edward Rasmussen Department of Engineering, University of Cambridge November 11th - 16th, 21 Rasmussen (Engineering, Cambridge) Gaussian Process

More information

Bayesian and Maximum Likelihood Estimation for Gaussian Processes on an Incomplete Lattice

Bayesian and Maximum Likelihood Estimation for Gaussian Processes on an Incomplete Lattice Bayesian and Maximum Likelihood Estimation for Gaussian Processes on an Incomplete Lattice Jonathan R. Stroud, Michael L. Stein and Shaun Lysen Georgetown University, University of Chicago, and Google,

More information

One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017

One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017 One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017 1 One Picture and a Thousand Words Using Matrix Approximations Dianne P. O Leary

More information

Nonstationary spatial process modeling Part II Paul D. Sampson --- Catherine Calder Univ of Washington --- Ohio State University

Nonstationary spatial process modeling Part II Paul D. Sampson --- Catherine Calder Univ of Washington --- Ohio State University Nonstationary spatial process modeling Part II Paul D. Sampson --- Catherine Calder Univ of Washington --- Ohio State University this presentation derived from that presented at the Pan-American Advanced

More information

Model Selection for Geostatistical Models

Model Selection for Geostatistical Models Model Selection for Geostatistical Models Richard A. Davis Colorado State University http://www.stat.colostate.edu/~rdavis/lectures Joint work with: Jennifer A. Hoeting, Colorado State University Andrew

More information

Multiple-step Time Series Forecasting with Sparse Gaussian Processes

Multiple-step Time Series Forecasting with Sparse Gaussian Processes Multiple-step Time Series Forecasting with Sparse Gaussian Processes Perry Groot ab Peter Lucas a Paul van den Bosch b a Radboud University, Model-Based Systems Development, Heyendaalseweg 135, 6525 AJ

More information

Tests for separability in nonparametric covariance operators of random surfaces

Tests for separability in nonparametric covariance operators of random surfaces Tests for separability in nonparametric covariance operators of random surfaces Shahin Tavakoli (joint with John Aston and Davide Pigoli) April 19, 2016 Analysis of Multidimensional Functional Data Shahin

More information

SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning

SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning Mark Schmidt University of British Columbia, May 2016 www.cs.ubc.ca/~schmidtm/svan16 Some images from this lecture are

More information

Mini-project in scientific computing

Mini-project in scientific computing Mini-project in scientific computing Eran Treister Computer Science Department, Ben-Gurion University of the Negev, Israel. March 7, 2018 1 / 30 Scientific computing Involves the solution of large computational

More information

A short introduction to INLA and R-INLA

A short introduction to INLA and R-INLA A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk

More information

Chapter 17: Undirected Graphical Models

Chapter 17: Undirected Graphical Models Chapter 17: Undirected Graphical Models The Elements of Statistical Learning Biaobin Jiang Department of Biological Sciences Purdue University bjiang@purdue.edu October 30, 2014 Biaobin Jiang (Purdue)

More information

Schwarz Preconditioner for the Stochastic Finite Element Method

Schwarz Preconditioner for the Stochastic Finite Element Method Schwarz Preconditioner for the Stochastic Finite Element Method Waad Subber 1 and Sébastien Loisel 2 Preprint submitted to DD22 conference 1 Introduction The intrusive polynomial chaos approach for uncertainty

More information

Exponential Family and Maximum Likelihood, Gaussian Mixture Models and the EM Algorithm. by Korbinian Schwinger

Exponential Family and Maximum Likelihood, Gaussian Mixture Models and the EM Algorithm. by Korbinian Schwinger Exponential Family and Maximum Likelihood, Gaussian Mixture Models and the EM Algorithm by Korbinian Schwinger Overview Exponential Family Maximum Likelihood The EM Algorithm Gaussian Mixture Models Exponential

More information

Association studies and regression

Association studies and regression Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration

More information