Hierarchical matrix techniques for maximum likelihood covariance estimation
|
|
- Madison Neal
- 5 years ago
- Views:
Transcription
1 Hierarchical matrix techniques for maximum likelihood covariance estimation tifica Alexander Litvinenko, Extreme Computing Research Center and Uncertainty Center, KAUST (joint work with M. Genton, Y. Sun and D. Keyes)
2 The structure of the talk 1. Motivation 2. Hierarchical matrices [Hackbusch 1]: 3. Matérn covariance function 4. Uncertain parameters of the covariance function: 4.1 Uncertain covariance length 4.2 Uncertain smoothness parameter 5. Identification of these parameters via maximizing the log-likelihood. 2 / 37
3 Motivation, problem 1 Task: to predict temperature, velocity, salinity, estimate parameters of covariance Grid: 50Mi locations on 50 levels, (X*Y*Z) + X*Y= 500*500* *500 = 50Mi. High-resolution time-dependent data about Red Sea: zonal velocity and temperature 3 / 37
4 Motivation, problem 2 Task: to predict moisture, compute covariance, estimate its parameters 2D-Grid: 2.5Mi locations with 2.1Mi observations and 278K missing values. Soil moisture latitude longitude High-resolution daily soil moisture data at the top layer of the Mississippi basin, U.S.A., (Chaney et al., in review). Important for agriculture, defense. Moisture is very heterogeneous. 4 / 37
5 Motivation, estimation of uncertain parameters cov. length H-matrix rank tion Logo Box-plots Lock-up for l = (domain [0, 1] 2 ) vs different H-matrix ranks k = {3, 7, }. Which H-matrix rank is sufficient for identification of parameters of a particular type of cov. matrix? 5 / 37
6 Motivation for H-matrices General dense matrix requires O(n 3 ) storage and time. Could be expensive! If covariance matrix is structured (diagonal, Toeplitz, circulant) then we can apply e.g. FFT with O(nlogn), but if not? 6 / 37
7 Hierarchical (H)-matrices Introduction into Hierarchical (H)-matrix technique 7 / 37
8 Examples of H-matrix approximations Figure : Three examples of H-matrix approximations, (left-middle) R n n, n = 2 10, of the discretised covariance function cov(x, y) = e r, l 1 = 0.15, l 2 = 0.2, x, y [0, 1] 2 ; (right) Soil moisture from example above with dofs. The biggest dense (dark) blocks R 32 32, max. rank k = 4 on the left, k = in the middle, and k = on the right. 8 / 37
9 Matérn covariance functions Matérn covariance functions C θ = 2σ2 ( r ν ( r ) Kν, θ = (σ Γ(ν) 2l) 2, ν, l). l / 37
10 Examples of Matérn covariance matrices ( ) ( ) 3r 3r C ν=3/2 (r) = 1 + exp l l ( ) ( 5r C ν=5/2 (r) = r ) 2 5r l 3l 2 exp l (1) (2) ν = 1/2 exponential covariance function C ν=1/2 (r) = exp( r), ν Gaussian covariance function C ν= (r) = exp( r 2 ). 10 / 37
11 Identifying uncertain parameters Identifying uncertain parameters 11 / 37
12 Identifying uncertain parameters Given: a vector of measurements z = (z 1,..., z n ) T with a covariance matrix C(θ ) = C(σ 2, ν, l). C θ = 2σ2 ( r ν ( r ) Kν, θ = (σ Γ(ν) 2l) 2, ν, l). l To identify: uncertain parameters (σ 2, ν, l). Plan: Maximize the log-likelihood function L(θ) = 1 ( ) Nlog2π + log det{c(θ)} + z T C(θ) 1 z, 2 tion Logo On each Lock-up iteration i we have a new matrix C(θ i ). 12 / 37
13 Other works 1. S. AMBIKASARAN, et al., Fast direct methods for gaussian processes and the analysis of NASA Kepler mission, arxiv: , (2014). 2. S. AMBIKASARAN, J. Y. LI, P. K. KITANIDIS, AND E. DARVE, Large-scale stochastic linear inversion using hierarchical matrices, Computational Geosciences, (20) 3. J. BALLANI AND D. KRESSNER, Sparse inverse covariance estimation with hierarchical matrices, (2015). 4. M. BEBENDORF, Why approximate LU decompositions of finite element discretizations of elliptic operators can be computed with almost linear complexity, (2007). 5. S. BOERM AND J. GARCKE, Approximating gaussian processes with H2-matrices, J. E. CASTRILLON, M. G. GENTON, AND R. YOKOTA, Multi-Level Restricted Maximum Likelihood Covariance Estimation and Kriging for Large Non-Gridded Spatial Datasets, (2015). 7. J. DOELZ, H. HARBRECHT, AND C. SCHWAB, Covariance regularity and H-matrix approximation for rough random fields, ETH-Zuerich, H. HARBRECHT et al, Efficient approximation of random fields for numerical applications, Numerical Linear Algebra with Applications, (2015).. C.-J. HSIEH, et al, Big QUIC: Sparse inverse covariance estimation for a million variables, J. QUINONERO-CANDELA, et al, A unifying view of sparse approximate gaussian process regression, (2005). 11. A. SAIBABA, S. AMBIKASARAN, J. YUE LI, P. KITANIDIS, AND E. DARVE, Application of hierarchical matrices to linear inverse problems in geostatistics, Oil & Gas Science (2012). / 37
14 Convergence of the optimization method 14 / 37
15 Details of the identification To maximize the log-likelihood function we use the Brent s method [Brent 73] (combining bisection method, secant method and inverse quadratic interpolation). 1. C(θ) C H (θ, k). 2. H-Cholesky: C H (θ, k) = LL T 3. z T C 1 z = z T (LL T ) 1 z = v T v, where v is a solution of L(θ, k)v(θ) := z(θ ). 4. Let λ i be diagonal elements of H-Cholesky factor L, then log det{c} = log det{ll T } = log det{ n λ 2 i } = 2 i=1 n logλ i, L(θ, k) = N N 2 log(2π) log{ L ii (θ, k)} 1 2 (v(θ)t v(θ)). (3) i=1 i=1 15 / 37
16 Log likelihood(θ ) Shape of Log likelihood(θ) log(det(c)) z T C 1 z Log likelihood parameter θ, truth θ*=12 Figure : Minimum of negative log-likelihood (black) is at θ = (,, l) 12 (σ 2 and ν are fixed) 16 / 37
17 Convergence of H-matrix approximations 0 5 Spectral norm, L=0.1, nu=1 Frob. norm, L=0.1 Spectral norm, L=0.2 Frob. norm, L=0.2 Spectral norm, L=0.5 Frob. norm, L= Spectral norm, L=0.1, nu=0.5 Frob. norm, L=0.1 Spectral norm, L=0.2 Frob. norm, L=0.2 Spectral norm, L=0.5 Frob. norm, L=0.5 log(rel. error) log(rel. error) rank k rank k ν = 1(left) and ν = 0.5 (right) for different cov. lengths l = {0.1, 02, 0.5} 17 / 37
18 Convergence of H-matrix approximations 0 2 Spectral norm, nu=1.5, L=0.1 Spectral norm, nu=1 Spectral norm, nu= Spectral norm, nu=1.5, L=0.5 Spectral norm, nu=1 Spectral norm, nu= log(rel. error) log(rel. error) rank k rank k ν = {1.5, 1, 0.5}, l = 0.1 (left) and ν = {1.5, 1, 0.5}, l = 0.5 (right) 18 / 37
19 What will change? Approximate C by C H 1. How the eigenvalues of C and C H differ? 2. How det(c) differs from det(c H )? [Below] 3. How L differs from L H? [Mario Bebendorf et al] 4. How C 1 differs from (C H ) 1? [Mario Bebendorf et al] 5. How L(θ, k) differs from L(θ)? [Below] 6. What is optimal H-matrix rank? [Below] 7. How θ H differs from θ? [Below] For theory, estimates for the rank and accuracy see works of Bebendorf, Grasedyck, Le Borne, Hackbusch,... 1 / 37
20 Remark For a small H-matrix rank k the H-matrix Cholesky of C H crashes when eigenvalues of C come very close to zero. A remedy is to increase the rank k. In our example for n = 65 2 we increased k from 7 to. To avoid this instability, we can modify C H m = C H + δ 2 I. Assume λ i are eigenvalues of C H. Then eigenvalues of C H m will be λ i + δ 2. n log det(cm H ) = log (λ i + δ 2 ) = i=1 n log(λ i + δ 2 ). (4) i=1 20 / 37
21 Error analysis Theorem (Existence of H-matrix inverse in [Bebendorf 11, Ballani, Kressner 14) Under certain conditions an H-matrix inverse exist C 1 H C 1 ε C 1, (5) theoretical estimations for rank k inv of C 1 H Theorem (Error in log det) are given. Let E := C C H, (C H ) 1 E := (C H ) 1 C I and for the spectral radius ρ((c H ) 1 E) = ρ((c H ) 1 C I) ε. (6) Then log det(c) log det(c H ) plog(1 ε). Proof: See [Ballani, Kressner 14], [Ipsen 05]. 21 / 37
22 How sensitive is Log-Likelihood to the H-matrix rank? It is not at all sensible. H-matrix approximation changes function L(θ, k) and estimation of θ very-very small. θ L(exact) L(7) L(20) Comparison of three likelihood functions, computed with different tion Logo H-matrix Lock-up ranks: exact, H-rank 7, H-rank 20. Exponential covariance function, with covariance length l = 0., domain G = [0, 1] / 37
23 How sensitive is Log-Likelihood to the H-matrix rank? loglikelihood θ Figure : Three negative log-likelihood functions: exact, commuted with H-matrix rank 7 and 17. One can see that even with rank 7 one can achieve very accurate results. 23 / 37
24 Do we need all measurements? Boxplots vs n cov. length number of measurements Moisture data. Boxplots with increasing of number of measurements, n = {1000,..., 32000}. The mean and median are obtained after averaging 100 simulations. 24 / 37
25 Decreasing of error bars with increasing number of measurements Error bars (mean +/- st. dev.) computed for different n. Decreasing of error bars with increasing of number of measurements/dimension, n = {17 2, 33 2, 65 2 }. The mean and median are obtained after averaging 200 simulations. 25 / 37
26 H-matrix approximation is robust w.r.t. parameter ν Figure : Dependence of H-matrix approximation error on parameter ν. Relative error C C H 2 C H 2 via smoothness parameter ν. H-matrix rank k = 8, n = 16641, Matern covariance matrix. 26 / 37
27 H-matrix approximation is robust w.r.t. cov. length l Figure : Dependence of H-matrix approximation error on cov. length l. Relative error C C H 2 C H 2 via covariance length l. H-matrix rank k = 8, n = 16641, Matern covariance matrix. 27 / 37
28 log-determinant of C rank5 ranks=3,4,5 -log-likelihood cov. length cov. length H-matrix approximation of log det(c) (left) and of the tion Logo log-likelihood Lock-up L (right); σ 2 = 1, ν = 0.5, k = 5. The red line - k = 5 and the blue - k = 3 on [0.01, 0.3], k = 4 on [0.3, 0.6] and k = 5 on [0.6, 1]. The rank k = 3 is sufficient to approximate C, but insufficient to approximate C 1 on the whole interval [0.01, 1]. The first numerical instability appears at point l i1 0.3, to avoid it the rank k is increased by 1, until the second instability appears at point l i / 37
29 10 8 log-likelihood 10 7 z T C -1 z parameter \nu H-matrix approx. of z T C 1 z and of log-likelihood L; σ 2 = 1, ν = 0.5, k increases from 5 until 12 after each jump. The red line - rank k = 5 and blue - k = 5 on [0.1, 1.42]; k = 6 on [1.42, 1.57]; k = 7 on [1.57, 1.3]; k = 8 on [1.3, 2.24] and k = {, 10, 12} between [2.24, 3.14]. k has to be increased to approximate C 1. The first numerical instability appears at point ν i1 1.42, to avoid it the rank k is increased by 1, until the second instability appears at point l i etc. 2 / 37
30 Time profiling, C-language 30 / 37
31 Parallel implementation with HLIBpro (R. Kriemann) We used to setup the exponential covariance matrix (cov. length=1), to compute Cholesky and the inverse. We used adaptive rank arithmetics with ε = 1e 4 for each block of C H and ε = 1e 8 for each block of C H. Number of processing cores is 40. We took moisture data (see above) with N points. N compute C H LL T inverse Compr. time size time size ε 1 time size ε 2 rate % sec. MB sec. MB sec. MB % e e % e e-1 Table : Here ε 1 := I (LL T ) 1 C 2, where L and C are H-matrices, I -identity matrix; ε 2 := I BC 2, where B is an H-matrix approximation of C / 37
32 Take into account the gradient L(θ i ) θ i = 1 ( ) 2 tr 1 C C 1 θ i 2 zt 1 C C C 1 z. (7) θ i For an exponential random field, have ( C(θ i ) = θ i l exp ) ( x y 2 = x y 2 ) x y l l 2 exp 2 l (8) C(θ i ) =: C 2 θ i L(θ i ) = 1 θ i 2 tr ( C 1 ) 1 C 2 2 zt C 1 C 2 C 1 z. 32 / 37
33 Conclusion Covariance matrices can be approximated in H-matrix format. Hypotes: H-matrix approximation is robust w.r.t. ν and l. Influence of H-matrix approximation error on the estimated parameters is small. With application of H-matrices we extend the class of covariance functions to work with, allows non-regular discretization of the covariance function on large spatial grids. With the maximizing algorithm we are able to identify both tion Logo parameters: Lock-up covariance lengths l and the smoothness ν 33 / 37
34 Future plans Parallel H-Cholesky for very large covariance matrices on non-regular grids Preconditioning for log-likelihood to decrease cond(c) Domain decomposition for large domains + H-matrix in each sub-domain Apply H-matrices for 1. Kriging estimate ŝ := C sy Cyy 1 y 2. Estimation of variance ˆσ, is the diagonal of conditional cov. matrix C ss y = diag ( ) C ss C sy Cyy 1 C ys, 3. Gestatistical optimal design ϕ A := n 1 tracec ss y, ϕ C := c ( ) T C ss C sy Cyy 1 C ys c, tion Logo To Lock-up implement gradient-based version Compare with the Bayesian Update (H. Matthies, H. Najm, K. Law, A. Stuart et al) 34 / 37
35 Literature 1. Application of hierarchical matrices for computing the Karhunen-Loéve expansion, B.N. Khoromskij, A. Litvinenko, H.G. Matthies, Computing 84 (1-2), 4-67, 31, Parameter identification in a probabilistic setting, B.V. Rosić, A. Kucerová, J Sýkora, O. Pajonk, A. Litvinenko, H.G. Matthies, Engineering Structures 50, 17-16, Methods for statistical data analysis with decision trees, V. Berikov, A. Litvinenko, Novosibirsk, Sobolev Institute of Mathematics, Parametric and uncertainty computations with tensor product representations, H.G. Matthies, A. Litvinenko, O. Pajonk, B.V. Rosić, E. Zander, Uncertainty in Scientific Computing, -150, Data sparse computation of the Karhunen-Loève expansion, B.N. Khoromskij, A. Litvinenko, AIP Conference Proceedings 1048 (1), 311, Kriging and spatial design accelerated by orders of magnitude: Combining low-rank covariance approximations with FFT-techniques W. Nowak, A. Litvinenko, Mathematical Geosciences 45 (4), , / 37
36 Acknowledgement 1. Lars Grasedyck (RWTH Aachen) and Steffen Boerm (Uni Kiel) for HLIB ( 2. Ronald Kriemann (MPI Leipzig) for 3. KAUST Research Computing group, KAUST Supercomputing Lab (KSL) 36 / 37
37 Matérn Fields (Whittle, 63) Taken from D. Simpson (see also Finn Lindgren, Havard Rue, David Bolin,...) Theorem The covariance function of a Matérn field c(x, y) = 1 Γ(ν + d/2)(4π) d/2 κ 2ν 2 ν 1 (κ x y )ν K ν (κ x y ) is the Green s function of the differential operator () L 2 ν = ( κ 2 ) ν+d/2. (10) 37 / 37
Uncertainty Quantification and related areas - An overview
Uncertainty Quantification and related areas - An overview Alexander Litvinenko, talk given at Department of Mathematical Sciences, Durham University Bayesian Computational Statistics & Modeling, KAUST
More informationLecture 1: Center for Uncertainty Quantification. Alexander Litvinenko. Computation of Karhunen-Loeve Expansion:
tifica Lecture 1: Computation of Karhunen-Loeve Expansion: Alexander Litvinenko http://sri-uq.kaust.edu.sa/ Stochastic PDEs We consider div(κ(x, ω) u) = f (x, ω) in G, u = 0 on G, with stochastic coefficients
More informationNON-LINEAR APPROXIMATION OF BAYESIAN UPDATE
tifica NON-LINEAR APPROXIMATION OF BAYESIAN UPDATE Alexander Litvinenko 1, Hermann G. Matthies 2, Elmar Zander 2 http://sri-uq.kaust.edu.sa/ 1 Extreme Computing Research Center, KAUST, 2 Institute of Scientific
More informationSampling and low-rank tensor approximation of the response surface
Sampling and low-rank tensor approximation of the response surface tifica Alexander Litvinenko 1,2 (joint work with Hermann G. Matthies 3 ) 1 Group of Raul Tempone, SRI UQ, and 2 Group of David Keyes,
More informationLecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko
tifica Lecture 1: Introduction to low-rank tensor representation/approximation Alexander Litvinenko http://sri-uq.kaust.edu.sa/ KAUST Figure : KAUST campus, 5 years old, approx. 7000 people (include 1400
More informationLecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko
tifica Lecture 1: Introduction to low-rank tensor representation/approximation Alexander Litvinenko http://sri-uq.kaust.edu.sa/ KAUST Figure : KAUST campus, 5 years old, approx. 7000 people (include 1400
More informationOverview of Spatial Statistics with Applications to fmri
with Applications to fmri School of Mathematics & Statistics Newcastle University April 8 th, 2016 Outline Why spatial statistics? Basic results Nonstationary models Inference for large data sets An example
More informationMultivariate Gaussian Random Fields with SPDEs
Multivariate Gaussian Random Fields with SPDEs Xiangping Hu Daniel Simpson, Finn Lindgren and Håvard Rue Department of Mathematics, University of Oslo PASI, 214 Outline The Matérn covariance function and
More informationA Spectral Approach to Linear Bayesian Updating
A Spectral Approach to Linear Bayesian Updating Oliver Pajonk 1,2, Bojana V. Rosic 1, Alexander Litvinenko 1, and Hermann G. Matthies 1 1 Institute of Scientific Computing, TU Braunschweig, Germany 2 SPT
More informationCovariance function estimation in Gaussian process regression
Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian
More informationHierarchical Low Rank Approximation of Likelihoods for Large Spatial Datasets
Hierarchical Low Rank Approximation of Likelihoods for Large Spatial Datasets Huang Huang and Ying Sun CEMSE Division, King Abdullah University of Science and Technology, Thuwal, 23955, Saudi Arabia. July
More informationLA Support for Scalable Kernel Methods. David Bindel 29 Sep 2018
LA Support for Scalable Kernel Methods David Bindel 29 Sep 2018 Collaborators Kun Dong (Cornell CAM) David Eriksson (Cornell CAM) Jake Gardner (Cornell CS) Eric Lee (Cornell CS) Hannes Nickisch (Phillips
More informationKarhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques
Institut für Numerische Mathematik und Optimierung Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Oliver Ernst Computational Methods with Applications Harrachov, CR,
More informationThe Bayesian approach to inverse problems
The Bayesian approach to inverse problems Youssef Marzouk Department of Aeronautics and Astronautics Center for Computational Engineering Massachusetts Institute of Technology ymarz@mit.edu, http://uqgroup.mit.edu
More informationSampling and Low-Rank Tensor Approximations
Sampling and Low-Rank Tensor Approximations Hermann G. Matthies Alexander Litvinenko, Tarek A. El-Moshely +, Brunswick, Germany + MIT, Cambridge, MA, USA wire@tu-bs.de http://www.wire.tu-bs.de $Id: 2_Sydney-MCQMC.tex,v.3
More informationHierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets
Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Abhirup Datta 1 Sudipto Banerjee 1 Andrew O. Finley 2 Alan E. Gelfand 3 1 University of Minnesota, Minneapolis,
More informationHierarchical Modelling for Univariate Spatial Data
Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationData are collected along transect lines, with dense data along. Spatial modelling using GMRFs. 200m. Today? Spatial modelling using GMRFs
Centre for Mathematical Sciences Lund University Engineering geology Lund University Results A non-stationary extension The model Estimation Gaussian Markov random fields Basics Approximating Mate rn covariances
More informationFast Direct Methods for Gaussian Processes
Fast Direct Methods for Gaussian Processes Mike O Neil Departments of Mathematics New York University oneil@cims.nyu.edu December 12, 2015 1 Collaborators This is joint work with: Siva Ambikasaran Dan
More informationKarhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes
TTU, October 26, 2012 p. 1/3 Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes Hao Zhang Department of Statistics Department of Forestry and Natural Resources Purdue University
More informationHierarchical Modeling for Univariate Spatial Data
Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This
More informationParametric Problems, Stochastics, and Identification
Parametric Problems, Stochastics, and Identification Hermann G. Matthies a B. Rosić ab, O. Pajonk ac, A. Litvinenko a a, b University of Kragujevac c SPT Group, Hamburg wire@tu-bs.de http://www.wire.tu-bs.de
More informationSolving Corrupted Quadratic Equations, Provably
Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin
More informationBayesian Statistics and Data Assimilation. Jonathan Stroud. Department of Statistics The George Washington University
Bayesian Statistics and Data Assimilation Jonathan Stroud Department of Statistics The George Washington University 1 Outline Motivation Bayesian Statistics Parameter Estimation in Data Assimilation Combined
More informationQuasi-optimal and adaptive sparse grids with control variates for PDEs with random diffusion coefficient
Quasi-optimal and adaptive sparse grids with control variates for PDEs with random diffusion coefficient F. Nobile, L. Tamellini, R. Tempone, F. Tesei CSQI - MATHICSE, EPFL, Switzerland Dipartimento di
More informationHilbert Space Methods for Reduced-Rank Gaussian Process Regression
Hilbert Space Methods for Reduced-Rank Gaussian Process Regression Arno Solin and Simo Särkkä Aalto University, Finland Workshop on Gaussian Process Approximation Copenhagen, Denmark, May 2015 Solin &
More informationApplications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices
Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Vahid Dehdari and Clayton V. Deutsch Geostatistical modeling involves many variables and many locations.
More informationTheory and Computation for Gaussian Processes
University of Chicago IPAM, February 2015 Funders & Collaborators US Department of Energy, US National Science Foundation (STATMOS) Mihai Anitescu, Jie Chen, Ying Sun Gaussian processes A process Z on
More informationKernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.
SINGAPORE SHANGHAI Vol TAIPEI - Interdisciplinary Mathematical Sciences 19 Kernel-based Approximation Methods using MATLAB Gregory Fasshauer Illinois Institute of Technology, USA Michael McCourt University
More informationScalable kernel methods and their use in black-box optimization
with derivatives Scalable kernel methods and their use in black-box optimization David Eriksson Center for Applied Mathematics Cornell University dme65@cornell.edu November 9, 2018 1 2 3 4 1/37 with derivatives
More informationKarhunen-Loève decomposition of Gaussian measures on Banach spaces
Karhunen-Loève decomposition of Gaussian measures on Banach spaces Jean-Charles Croix jean-charles.croix@emse.fr Génie Mathématique et Industriel (GMI) First workshop on Gaussian processes at Saint-Etienne
More informationGradient Enhanced Universal Kriging Model for Uncertainty Propagation in Nuclear Engineering
Gradient Enhanced Universal Kriging Model for Uncertainty Propagation in Nuclear Engineering Brian A. Lockwood 1 and Mihai Anitescu 2 1 Department of Mechanical Engineering University of Wyoming 2 Mathematics
More informationContents. Preface to the Third Edition (2007) Preface to the Second Edition (1992) Preface to the First Edition (1985) License and Legal Information
Contents Preface to the Third Edition (2007) Preface to the Second Edition (1992) Preface to the First Edition (1985) License and Legal Information xi xiv xvii xix 1 Preliminaries 1 1.0 Introduction.............................
More informationComputation fundamentals of discrete GMRF representations of continuous domain spatial models
Computation fundamentals of discrete GMRF representations of continuous domain spatial models Finn Lindgren September 23 2015 v0.2.2 Abstract The fundamental formulas and algorithms for Bayesian spatial
More informationMonte Carlo Methods for Uncertainty Quantification
Monte Carlo Methods for Uncertainty Quantification Mike Giles Mathematical Institute, University of Oxford Contemporary Numerical Techniques Mike Giles (Oxford) Monte Carlo methods 1 / 23 Lecture outline
More informationOn Gaussian Process Models for High-Dimensional Geostatistical Datasets
On Gaussian Process Models for High-Dimensional Geostatistical Datasets Sudipto Banerjee Joint work with Abhirup Datta, Andrew O. Finley and Alan E. Gelfand University of California, Los Angeles, USA May
More informationNon-Intrusive Solution of Stochastic and Parametric Equations
Non-Intrusive Solution of Stochastic and Parametric Equations Hermann G. Matthies a Loïc Giraldi b, Alexander Litvinenko c, Dishi Liu d, and Anthony Nouy b a,, Brunswick, Germany b École Centrale de Nantes,
More informationSpatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields
Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields 1 Introduction Jo Eidsvik Department of Mathematical Sciences, NTNU, Norway. (joeid@math.ntnu.no) February
More informationGaussian Processes 1. Schedule
1 Schedule 17 Jan: Gaussian processes (Jo Eidsvik) 24 Jan: Hands-on project on Gaussian processes (Team effort, work in groups) 31 Jan: Latent Gaussian models and INLA (Jo Eidsvik) 7 Feb: Hands-on project
More informationFixed-domain Asymptotics of Covariance Matrices and Preconditioning
Fixed-domain Asymptotics of Covariance Matrices and Preconditioning Jie Chen IBM Thomas J. Watson Research Center Presented at Preconditioning Conference, August 1, 2017 Jie Chen (IBM Research) Covariance
More information1. Structured representation of high-order tensors revisited. 2. Multi-linear algebra (MLA) with Kronecker-product data.
Lect. 4. Toward MLA in tensor-product formats B. Khoromskij, Leipzig 2007(L4) 1 Contents of Lecture 4 1. Structured representation of high-order tensors revisited. - Tucker model. - Canonical (PARAFAC)
More informationA Framework for Daily Spatio-Temporal Stochastic Weather Simulation
A Framework for Daily Spatio-Temporal Stochastic Weather Simulation, Rick Katz, Balaji Rajagopalan Geophysical Statistics Project Institute for Mathematics Applied to Geosciences National Center for Atmospheric
More informationECE521 lecture 4: 19 January Optimization, MLE, regularization
ECE521 lecture 4: 19 January 2017 Optimization, MLE, regularization First four lectures Lectures 1 and 2: Intro to ML Probability review Types of loss functions and algorithms Lecture 3: KNN Convexity
More informationAn Inversion-Free Estimating Equations Approach for. Gaussian Process Models
An Inversion-Free Estimating Equations Approach for Gaussian Process Models Mihai Anitescu Jie Chen Michael L. Stein November 29, 2015 Abstract One of the scalability bottlenecks for the large-scale usage
More informationKriging models with Gaussian processes - covariance function estimation and impact of spatial sampling
Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling François Bachoc former PhD advisor: Josselin Garnier former CEA advisor: Jean-Marc Martinez Department
More informationCurriculum Vitae. Alexander Litvenenko
Curriculum Vitae Alexander Litvinenko https://ecrc.kaust.edu.sa/pages/litvinenko.aspx http://stochastic_numerics.kaust.edu.sa/pages/litvinenkoalexander.aspx Name: Nationality: Languages: Alexander Litvenenko
More informationAdaptive Sampling of Clouds with a Fleet of UAVs: Improving Gaussian Process Regression by Including Prior Knowledge
Master s Thesis Presentation Adaptive Sampling of Clouds with a Fleet of UAVs: Improving Gaussian Process Regression by Including Prior Knowledge Diego Selle (RIS @ LAAS-CNRS, RT-TUM) Master s Thesis Presentation
More informationHierarchical Parallel Solution of Stochastic Systems
Hierarchical Parallel Solution of Stochastic Systems Second M.I.T. Conference on Computational Fluid and Solid Mechanics Contents: Simple Model of Stochastic Flow Stochastic Galerkin Scheme Resulting Equations
More informationLatent Gaussian Processes and Stochastic Partial Differential Equations
Latent Gaussian Processes and Stochastic Partial Differential Equations Johan Lindström 1 1 Centre for Mathematical Sciences Lund University Lund 2016-02-04 Johan Lindström - johanl@maths.lth.se Gaussian
More informationFast matrix algebra for dense matrices with rank-deficient off-diagonal blocks
CHAPTER 2 Fast matrix algebra for dense matrices with rank-deficient off-diagonal blocks Chapter summary: The chapter describes techniques for rapidly performing algebraic operations on dense matrices
More informationState Space Representation of Gaussian Processes
State Space Representation of Gaussian Processes Simo Särkkä Department of Biomedical Engineering and Computational Science (BECS) Aalto University, Espoo, Finland June 12th, 2013 Simo Särkkä (Aalto University)
More informationSome Areas of Recent Research
University of Chicago Department Retreat, October 2012 Funders & Collaborators NSF (STATMOS), US Department of Energy Faculty: Mihai Anitescu, Liz Moyer Postdocs: Jie Chen, Bill Leeds, Ying Sun Grad students:
More informationHierarchical Matrices. Jon Cockayne April 18, 2017
Hierarchical Matrices Jon Cockayne April 18, 2017 1 Sources Introduction to Hierarchical Matrices with Applications [Börm et al., 2003] 2 Sources Introduction to Hierarchical Matrices with Applications
More informationK-Means and Gaussian Mixture Models
K-Means and Gaussian Mixture Models David Rosenberg New York University October 29, 2016 David Rosenberg (New York University) DS-GA 1003 October 29, 2016 1 / 42 K-Means Clustering K-Means Clustering David
More informationH 2 -matrices with adaptive bases
1 H 2 -matrices with adaptive bases Steffen Börm MPI für Mathematik in den Naturwissenschaften Inselstraße 22 26, 04103 Leipzig http://www.mis.mpg.de/ Problem 2 Goal: Treat certain large dense matrices
More informationSub-kilometer-scale space-time stochastic rainfall simulation
Picture: Huw Alexander Ogilvie Sub-kilometer-scale space-time stochastic rainfall simulation Lionel Benoit (University of Lausanne) Gregoire Mariethoz (University of Lausanne) Denis Allard (INRA Avignon)
More informationCovariance Matrix Simplification For Efficient Uncertainty Management
PASEO MaxEnt 2007 Covariance Matrix Simplification For Efficient Uncertainty Management André Jalobeanu, Jorge A. Gutiérrez PASEO Research Group LSIIT (CNRS/ Univ. Strasbourg) - Illkirch, France *part
More informationNearest Neighbor Gaussian Processes for Large Spatial Data
Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns
More informationGaussian with mean ( µ ) and standard deviation ( σ)
Slide from Pieter Abbeel Gaussian with mean ( µ ) and standard deviation ( σ) 10/6/16 CSE-571: Robotics X ~ N( µ, σ ) Y ~ N( aµ + b, a σ ) Y = ax + b + + + + 1 1 1 1 1 1 1 1 1 1, ~ ) ( ) ( ), ( ~ ), (
More informationMixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate
Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means
More informationMulti-resolution models for large data sets
Multi-resolution models for large data sets Douglas Nychka, National Center for Atmospheric Research National Science Foundation NORDSTAT, Umeå, June, 2012 Credits Steve Sain, NCAR Tia LeRud, UC Davis
More informationMixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate
Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means
More informationLearning the Linear Dynamical System with ASOS ( Approximated Second-Order Statistics )
Learning the Linear Dynamical System with ASOS ( Approximated Second-Order Statistics ) James Martens University of Toronto June 24, 2010 Computer Science UNIVERSITY OF TORONTO James Martens (U of T) Learning
More informationFundamentals of Data Assimila1on
014 GSI Community Tutorial NCAR Foothills Campus, Boulder, CO July 14-16, 014 Fundamentals of Data Assimila1on Milija Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University
More informationPreface to Second Edition... vii. Preface to First Edition...
Contents Preface to Second Edition..................................... vii Preface to First Edition....................................... ix Part I Linear Algebra 1 Basic Vector/Matrix Structure and
More informationOptimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering
Optimisation under Uncertainty with Stochastic PDEs for the History Matching Problem in Reservoir Engineering Hermann G. Matthies Technische Universität Braunschweig wire@tu-bs.de http://www.wire.tu-bs.de
More informationStochastic Spectral Approaches to Bayesian Inference
Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to
More informationMultivariate modelling and efficient estimation of Gaussian random fields with application to roller data
Multivariate modelling and efficient estimation of Gaussian random fields with application to roller data Reinhard Furrer, UZH PASI, Búzios, 14-06-25 NZZ.ch Motivation Microarray data: construct alternative
More informationPrincipal Component Analysis CS498
Principal Component Analysis CS498 Today s lecture Adaptive Feature Extraction Principal Component Analysis How, why, when, which A dual goal Find a good representation The features part Reduce redundancy
More informationHands-on Matrix Algebra Using R
Preface vii 1. R Preliminaries 1 1.1 Matrix Defined, Deeper Understanding Using Software.. 1 1.2 Introduction, Why R?.................... 2 1.3 Obtaining R.......................... 4 1.4 Reference Manuals
More informationPlatzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen
Platzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen PARAMetric UNCertainties, Budapest STOCHASTIC PROCESSES AND FIELDS Noémi Friedman Institut für Wissenschaftliches Rechnen, wire@tu-bs.de
More informationMultilevel accelerated quadrature for elliptic PDEs with random diffusion. Helmut Harbrecht Mathematisches Institut Universität Basel Switzerland
Multilevel accelerated quadrature for elliptic PDEs with random diffusion Mathematisches Institut Universität Basel Switzerland Overview Computation of the Karhunen-Loéve expansion Elliptic PDE with uniformly
More informationFrequentist-Bayesian Model Comparisons: A Simple Example
Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal
More informationHierarchical Modelling for Univariate Spatial Data
Spatial omain Hierarchical Modelling for Univariate Spatial ata Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A.
More informationApproximating the Covariance Matrix with Low-rank Perturbations
Approximating the Covariance Matrix with Low-rank Perturbations Malik Magdon-Ismail and Jonathan T. Purnell Department of Computer Science Rensselaer Polytechnic Institute Troy, NY 12180 {magdon,purnej}@cs.rpi.edu
More informationStatistics & Data Sciences: First Year Prelim Exam May 2018
Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book
More informationNon-linear least squares
Non-linear least squares Concept of non-linear least squares We have extensively studied linear least squares or linear regression. We see that there is a unique regression line that can be determined
More informationCirculant embedding of approximate covariances for inference from Gaussian data on large lattices
Circulant embedding of approximate covariances for inference from Gaussian data on large lattices Joseph Guinness and Montserrat Fuentes Abstract Recently proposed computationally efficient Markov chain
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationDouglas Nychka, Soutir Bandyopadhyay, Dorit Hammerling, Finn Lindgren, and Stephan Sain. October 10, 2012
A multi-resolution Gaussian process model for the analysis of large spatial data sets. Douglas Nychka, Soutir Bandyopadhyay, Dorit Hammerling, Finn Lindgren, and Stephan Sain October 10, 2012 Abstract
More informationCross-covariance Functions for Tangent Vector Fields on the Sphere
Cross-covariance Functions for Tangent Vector Fields on the Sphere Minjie Fan 1 Tomoko Matsuo 2 1 Department of Statistics University of California, Davis 2 Cooperative Institute for Research in Environmental
More informationProbabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms
Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,
More informationLecture 5. Gaussian Models - Part 1. Luigi Freda. ALCOR Lab DIAG University of Rome La Sapienza. November 29, 2016
Lecture 5 Gaussian Models - Part 1 Luigi Freda ALCOR Lab DIAG University of Rome La Sapienza November 29, 2016 Luigi Freda ( La Sapienza University) Lecture 5 November 29, 2016 1 / 42 Outline 1 Basics
More informationGaussian processes for spatial modelling in environmental health: parameterizing for flexibility vs. computational efficiency
Gaussian processes for spatial modelling in environmental health: parameterizing for flexibility vs. computational efficiency Chris Paciorek March 11, 2005 Department of Biostatistics Harvard School of
More informationUncertainty Quantification in Discrete Fracture Network Models
Uncertainty Quantification in Discrete Fracture Network Models Claudio Canuto Dipartimento di Scienze Matematiche, Politecnico di Torino claudio.canuto@polito.it Joint work with Stefano Berrone, Sandra
More informationGaussian Process Regression
Gaussian Process Regression 4F1 Pattern Recognition, 21 Carl Edward Rasmussen Department of Engineering, University of Cambridge November 11th - 16th, 21 Rasmussen (Engineering, Cambridge) Gaussian Process
More informationBayesian and Maximum Likelihood Estimation for Gaussian Processes on an Incomplete Lattice
Bayesian and Maximum Likelihood Estimation for Gaussian Processes on an Incomplete Lattice Jonathan R. Stroud, Michael L. Stein and Shaun Lysen Georgetown University, University of Chicago, and Google,
More informationOne Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017
One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017 1 One Picture and a Thousand Words Using Matrix Approximations Dianne P. O Leary
More informationNonstationary spatial process modeling Part II Paul D. Sampson --- Catherine Calder Univ of Washington --- Ohio State University
Nonstationary spatial process modeling Part II Paul D. Sampson --- Catherine Calder Univ of Washington --- Ohio State University this presentation derived from that presented at the Pan-American Advanced
More informationModel Selection for Geostatistical Models
Model Selection for Geostatistical Models Richard A. Davis Colorado State University http://www.stat.colostate.edu/~rdavis/lectures Joint work with: Jennifer A. Hoeting, Colorado State University Andrew
More informationMultiple-step Time Series Forecasting with Sparse Gaussian Processes
Multiple-step Time Series Forecasting with Sparse Gaussian Processes Perry Groot ab Peter Lucas a Paul van den Bosch b a Radboud University, Model-Based Systems Development, Heyendaalseweg 135, 6525 AJ
More informationTests for separability in nonparametric covariance operators of random surfaces
Tests for separability in nonparametric covariance operators of random surfaces Shahin Tavakoli (joint with John Aston and Davide Pigoli) April 19, 2016 Analysis of Multidimensional Functional Data Shahin
More informationSVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning
SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning Mark Schmidt University of British Columbia, May 2016 www.cs.ubc.ca/~schmidtm/svan16 Some images from this lecture are
More informationMini-project in scientific computing
Mini-project in scientific computing Eran Treister Computer Science Department, Ben-Gurion University of the Negev, Israel. March 7, 2018 1 / 30 Scientific computing Involves the solution of large computational
More informationA short introduction to INLA and R-INLA
A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk
More informationChapter 17: Undirected Graphical Models
Chapter 17: Undirected Graphical Models The Elements of Statistical Learning Biaobin Jiang Department of Biological Sciences Purdue University bjiang@purdue.edu October 30, 2014 Biaobin Jiang (Purdue)
More informationSchwarz Preconditioner for the Stochastic Finite Element Method
Schwarz Preconditioner for the Stochastic Finite Element Method Waad Subber 1 and Sébastien Loisel 2 Preprint submitted to DD22 conference 1 Introduction The intrusive polynomial chaos approach for uncertainty
More informationExponential Family and Maximum Likelihood, Gaussian Mixture Models and the EM Algorithm. by Korbinian Schwinger
Exponential Family and Maximum Likelihood, Gaussian Mixture Models and the EM Algorithm by Korbinian Schwinger Overview Exponential Family Maximum Likelihood The EM Algorithm Gaussian Mixture Models Exponential
More informationAssociation studies and regression
Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration
More information