Fast evaluation of mixed derivatives and calculation of optimal weights for integration. Hernan Leovey

Size: px
Start display at page:

Download "Fast evaluation of mixed derivatives and calculation of optimal weights for integration. Hernan Leovey"

Transcription

1 Fast evaluation of mixed derivatives and calculation of optimal weights for integration Humboldt Universität zu Berlin MCQMC2012 Tenth International Conference on Monte Carlo and Quasi Monte Carlo Methods in Scientific Computing

2 Contents Algorithmic Differentiation (AD) 1 Algorithmic Differentiation (AD) Basics Complexity 2 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions 3 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension 4

3 Basics Algorithmic Differentiation (AD) Basics Complexity Frequently we have a program that calculates numerical values for a function, and we would like to obtain accurate values for derivatives of the function as well. The usual divided difference approach is given by : D +h f (x) f (x + h) f (x) h or D ±h f (x) f (x + h) f (x h) 2h For h small, truncation and round-off errors reduce the number of significant digits If h is not small, normally no good approximation to a derivative is expected

4 Basics Algorithmic Differentiation (AD) Basics Complexity Typically h = ɛ is taken, for ɛ = working accuracy. Expected accuracy: 1 2 of the significant digits of f for D +h 2 3 of the significant digits of f for D ±h In contrast, AD methods incur no truncation errors at all and usually yield derivatives with working accuracy. AD 0: Algorithmic Differentiation does not incur truncation errors AD 1: Difference quotients may sometimes be useful too AD 2: What is good for function values is good for their derivatives

5 Basics Algorithmic Differentiation (AD) Basics Complexity Standard setting for AD: Vector function F is the composition of a sequence of once continuously differentiable elemental functions ϕ i. Basic set of functions (polynomial core): {+,, (unary sign op.), c (const. init.)} A typical example of a library containing elemental functions: {c, +,,, /, exp, log, sin, cos, tan, tan 1,..., Φ, Φ 1,...}

6 Basics Complexity

7 Basics Complexity Basic complexity results Consider temporal complexity measure TIME, TIME{task(F )} = w WORK{task(F )} (1) with w = (w 1, w 2, w 3, w 4 ) a vector of platform dependent weights, and MOVES of fetches and stores WORK{task} ADDS MULTS of additions and subtractions of multiplications (2) NLOPS of nonlinear operations Forward mode AD: TIME{F (x), F (x)ẋ} ω tang TIME{F (x)} with a constant ω tang [2, 5/2]

8 Basics Complexity Reverse mode AD: Cheap Gradient Principle TIME{F (x), ȳ F (x)} ω grad TIME{F (x)} (3) for a constant ω grad [3, 4]. As consequence, the cost to evaluate a gradient f is bounded above by a small constant ω grad [3, 4] times the cost to evaluate the function itself. Random Access Memory requirements, in forward and reverse, are bounded multiples of those for the functions. Sequential Access Memory requirement of basic reverse mode is proportional to temporal complexity of the function.

9 Algorithmic Differentiation (AD) Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions (Automatic Evaluations of Cross-Derivatives. Griewank, L, L, Z) With the term cross derivatives we refer to those mixed partial derivatives where differentiation w.r.t. each variable is done at most once. f i (x) = j i x j f (x) = k f x i1... x ik (x), i = {i 1, i 2,..., i k }. There are 2 d cross derivatives if we take f (x) = f (x). We create a data structure with all 2 d cross derivatives of a function u in a flat array with 2 d entries. We call such data structure an d dimensional cube. d = 3 u u {1} u {2} u {1,2} u {3} u {1,3} u {2,3} u {1,2,3}

10 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions Basic Operations For a function u we denote by U its cube. For a constant function u(x) = c we set U[0] = c and zero everywhere else. For a coordinate function reps. input variable u(x) = x j we initialize its cube by U[0] = x j and U[2 j ] = 1. The rest of the entries are set to zero. Addition and Subtraction: V[i]=U[i] ± W[i] for all 0 i < 2 d. Scalar Multiplication: For v(x) = cu(x) the propagation rule is V[i]=c*U[i]. Scalar Addition/Subtraction is applied only to U[0]. The complexity of the above basic operations is (O(2 d )).

11 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions Nonlinear Operations Multiplication: The Leibniz formula for the multiplication of two functions v = u w states that: v i (x) = j i u j (x)w i j (x). Assume now that n / i. Then the above convolution sum can be split into v i {n} (x) = j i u i j (x)w j {n} (x) + j i u j {n} (x)w i j (x) Fixing the same subset i, the sums have the same structure. They all operate inside separate halves of cubes.

12 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions This leads to a possible implementation: void crossmult (int h, double U, double W, double V) { if (h == 1) { V[0] + = (U[0] W[0]); return; } h/ = 2; crossmult(h,u,w+h,v+h); crossmult(h,u+h,w,v+h); crossmult(h,u,w,v); } Due to the recursive nature of this procedure, there will be 3 d overall function calls at h = 1 resulting in 3 d multiplications and the same number of additions.

13 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions Exponential function: v = exp(u) has a very simple identity for the first partial derivatives, v k = vu k. This generalizes for k / i to: v i {k} = j i v i j (x)u j {k} (x) The second half cube of v is thus obtained by multiplying the previously computed first half cube of v and the second half cube of u. void exponent(int h, double U, double V) { int i; for(i= 0;i<h;i++) { V[i] = 0.0; } V[0]=exp(U[0]); for(i=1;i<h;i =2) { crossmult (i,v,u+i,v+i); } } There are d calls to the multiplication function and the final relative cost is 1 2 of the cost of a full multiplication.

14 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions Complexity: Nonlinear differentiable functions ϕ(u) included in math.h exhibit cost proportional to cross multiplication Given a library exhibiting cost proportional to cross multiplication, extend it by considering any nonlinear ϕ(u) satisfying differential equation ϕ (u) a(u)ϕ(u) = b(u) with functions a(.), b(.) in original library (ODE extension). Proposition The direct computation of all cross derivatives f of a function f given as an evaluation procedure (with elementals in ODE extended library) is itself an evaluation procedure with complexity OPS(f ) = O(3 d ) OPS(f ) for the runtime and with a factor of 2 d in the memory size. The unit is one multiplication, which is also the cost of addition or subtraction.

15 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions Comparison with other methods: Proposition Method of interpolation of all cross derivatives from Taylor coefficients via univariate expansions exhibits complexity OPS(f ) = O(d 2 2 d ) (OPS(f ) + c), c 4, for the runtime and with a factor of (d + 1) 2 d in the memory size. The cross over between the methods occur at d 14. For large dimensions d, the Taylor method will have better runtimes. Advantages of direct new method: more accurate than Taylor univariate method faster for d 14

16 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Quasi Monte Carlo Methods (QMC): Q N,n (f ) := 1 N N f (x i ) I (f ) := f (x)dx, [0,1] n i=1 with x 1,, x N deterministically and cleverly chosen from [0, 1] n. Lattice Rules Q N,n,z (f ) := 1 N 1 f N i=0 ({ }) i N z Where N (usually prime) is the number of selected points and z is a carefully selected integer vector in Z n. Shifted Lattice Rules Q N,n,z, (f ) := 1 N 1 ({ }) i f N N z + for [0, 1] n. i=0

17 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension (Weighted) Reproducing Kernel Hilbert spaces (Sloan&Woźniakowski 98) Integration over particular RKHS F n of functions over [0, 1] n. Reproducing kernel K n (x, t) is function defined over [0, 1] n [0, 1] n, such that K n (., t) F n for all t [0, 1] n and f (t) = f (.), K n (., t) n, f F n ; t [0, 1] n. Worst Case Error of QMC algorithm over F n e(q N,n ) := sup I (f ) Q N,n (f ) f F n: f Fn 1 Assume integration functional I (.) is continuous over F n, then e(q N,n ) is bounded and I (f ) Q N,n (f ) e(q N,n ). f Fn

18 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Weighted Unanchored Sobolev Space F n,γ Consider weights 0 γ n,i, for i {1,, n}. ( 1 K n,γ (x, y) = 1+ γ n,i 2 B 2({x j y j }) + (x j 1 2 )(y j 1 ) 2 ) f Fn,γ = =i {1,,n} i {1,,n} j i ( ) 2 γ 1 i n,i f (x i, x D i )dx D i dx i [0,1] i [0,1] x n i i 1 2 Product weights γ n,i = j i γ {n,j} Tensor Product RKHS n F n,γ = H n,γ := H 1,γ1 H 1,γn n times K n,γ (x, y) = K 1,γj (x i, y i ) j=1

19 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Theorem (Novak&Woźniakowski 10, Kuo, Sloan, Joe,..) Let 0 γ n,i, i D, D := {1,, n}, f F n,γ. Given a prime number N, there exits a shifted rank-1 lattice rule Q N,n,z, with generator vector z constructed by the Component by Component algorithm (CBC), such that ( ( ) ) i τ =i D (γ n,i) 1/(2τ) 2ζ(1/τ) ( 2π) 1/τ I (f ) Q N,d,z, (f ) (N 1) τ f F n,γ for any τ [ 1 2, 1). For fixed f, we need the weights to construct a generator vector z for a lattice rule, using CBC algorithm. How should we choose the weights in practice? What is an optimal embedding for a given function f in a practical problem?

20 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Common Approach: Choose the weights such that the integration error bound is minimized. In general case, 2 n 1 terms inside f F = n,γ Approach: =i D ( ) 2 γ 1 i n,i f (x i, x D i )dx D i dx i [0,1] i [0,1] x n i i 1 2 Very often, problems in applications exhibit low effective dimension d << n. Effective dimension refers to essential ANOVA part of the function that accumulates most of the variance ( 99%).

21 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Assume f is a square integrable function. Then we can write f as the sum of 2 n ANOVA terms: f (x) = f i (x), f i (x) = f (x i, x D i )dx D i f j (x) i D [0,1] n i j i For a given family T of subsets of D, let us define now f T (x) = i T f i (x). Then, the integration error of a QMC algorithm Q N,n is given by (I Q N,n )(f ) (I Q N,n )(f T ) + (I Q N,n) i {1,...,n},i T f i (x)

22 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Theorem Let T be a given family of subsets of D. Let f i F n,γ for i T. Then for the function f T defined above it holds f T F = n,γ =i T ( γ 1 i ) 2 n,i f (x i, x D i )dx D i dx i [0,1] x i i [0,1] n i Moreover, if f F n,γ, it holds for i D b f,i := = [0,1] i ( i ) 2 f (x i, x D i )dx D i dx i x i [0,1] n i ( ) 2 i f (x i, x D i )dx D i dx i [0,1] x n i i [0,1] i [0,1] n 1 2 ( ) i 2 f (x) dx x i

23 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Integrands with low effective dimension Remark Note that any (good) upper bound b f,i R, b f,i b f,i, conduces also to an integration error upper bound of the form ( ( ) ) i τ i D (γ n,i) 1/(2τ) 2ζ(1/τ) ( 1 2π) 1/τ (I Q N,n )(f T ) 2 γ 1 (N 1) τ n,i b f,i i T Product Weights (γ n,i = j i γ n,{j}): Let d denote effective dimension of f in truncation sense (say d 14). App.-1 f EFFTd := f i (x). i {1,...,d}

24 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Set 0 0 = 0, c 0 = + for c > 0. Assume that at least one term b i,f > 0 for some i {1,..., d}. For f EFFTd consider bound objective function ψ : R n 0 [0, + ] (where the variables are the weights). For simplicity set ψ(0) = +, and for (x 1,..., x n ) R n 0 \ {0} define ( n ( ψ(x 1,..., x n) 2τ = 1 + x 1 2ζ(1/τ) j ( 2π) 1/τ j=1 ) ) τ 1 i {1,...,d}( j i x 1 j )b i,f 1 2 Clearly we have: minimize ψ(x 1,..., x n) minimize ψ(x 1,..., x n) (x 1,...,x n) R n 0 (x 1,...,x n) R n 0 subject to x j = 0, d + 1 j n.

25 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Choice for non important weights (in ANOVA sense): Lemma For fixed τ [1/2, 1), let γn = (γn,1,..., γn,d, 0,..., 0) be an optimal feasible solution of problem above. Let ɛ 0 > 0, and let a 1,..., a n d be any sequence of nonnegative real numbers with n d i=1 a i M. Define R 0 = ( 2π) 1/τ M2ζ(1/τ) log τ ɛ 1 0 d j=1 (1 + (γ n,j) 1 2τ d j=1 (1 + (γ n,j ) 1 2τ ) 2ζ(1/τ) ( 2π) 1/τ ) 2ζ(1/τ) ( 2π) 1/τ Then it follows for γ n = (γ n,1,..., γ n,d, R 0a 1,..., R 0a n d ) that ψ(γ n) (1 + ɛ 0)ψ(γ n )

26 Option valuation problem for arithmetic average Asian options Asset S t follows the geometric Brownian motion model. ) ) S t = S 0 exp ((r σ2 t + σw t 2 Simulating asset prices reduces to simulating paths W t1,..., W td. e rt V = (2π) d/2 max 1 d S j (w) K, 0 e 1 2 wt C 1w dw det(c) R d d with w = (W t1,..., W td ). After a factorization C = AA T of the covariance matrix, transform integral using Φ 1 (.). V = e [0,1] rt max 1 d S j (AΦ 1 (x)) K, 0 dx, d d j=1 For the tests, we simplify the problem assuming K = 0. Consider principal components (PCA) and Brownian Bridge (BB) factorization of C. j=1

27 Sensitivity tests for effective dimension (Algo. Wang&Fang 03) K = 0, S 0 = 100, T = 1 real dimension n = 16, 64, 128 Domain Truncation (Kuo&Sloan&Griebel 10) ɛ = 0.1, 0.01, 0.001, ( I (f ) I (f ɛ ) ɛs 0 ) For (σ, r) [0.05, 0.35] [0.05, 0.35] (tests on 7 7 uniform grid) Using 2 16 Sobol points, all tests resulted with effective dimension in truncation sense d 3 for PCA, and d 8 for BB construction. ( ) b ˆ i 2 f,i b f,i := f (x) dx for i {1,, d} [0,1] x n i CrossAD cost for simplified examples without strike (K = 0): Example \n = Runtime(crossPCA) (d = 4) Runtime(PCA) Runtime(crossBB) (d = 8) Runtime(BB)

28 Fixed K = 0, S 0 = 100, T = 1,σ = 0.1,r = 0.1, domain truncation ɛ = 0.1, (b i,f estimates using cross AD for d first variables) Table: Weights for τ = 0.9 (runtime for opt. solver approx seconds) BB n8 S11 n8 S14 n8 acc n16 S11 n16 S14 n16 acc n128 S11 n128 S14 γn, γn, γn, γn, γn, γn, γn, γn, ψ(γ ) 2.6e e e e e e e e+04 ψ( 1, 0) 9.5e e e e e e e+05 j 3.5e+05 ψ( 1, 0) 1.0e e e e e e e+04 j 2 5.2e+04

29 Table: Weights for τ = 0.9 (runtime for opt. solver approx seconds) PCA n8 S11 n8 S14 n8 acc n16 S11 n16 S14 n16 acc n128 S11 n128 S14 γn, γn, γn, γn, ψ(γ ) 6.0e e e e e e e e+03 ψ( 1, 0) 4.0e e e e e e e+04 j 1.1e+04 ψ( 1, 0) 1.6e e e e e e e+03 j 2 4.5e+03

30 Further investigations: f EFF + T d := i {1,...,d} f i (x) + d+1 j n f {j} (x). Using cross AD + reverse mode AD for cheap gradients (Optimization problem remains n dimensional) Product and order dependent Weights for functions with low effective superposition dimension using forward and reverse AD (No need for numerical Optimization) Good bounds for functions with kinks (K 0, eff. sup. dim. d = 2 and P.O.D. weights) Improved sampling strategy for squared mixed derivatives strongly diverging at small sub-cube borders Domain truncation alternative Thank you for your attention!

Quasi-Monte Carlo integration over the Euclidean space and applications

Quasi-Monte Carlo integration over the Euclidean space and applications Quasi-Monte Carlo integration over the Euclidean space and applications f.kuo@unsw.edu.au University of New South Wales, Sydney, Australia joint work with James Nichols (UNSW) Journal of Complexity 30

More information

Quasi-Monte Carlo Methods for Applications in Statistics

Quasi-Monte Carlo Methods for Applications in Statistics Quasi-Monte Carlo Methods for Applications in Statistics Weights for QMC in Statistics Vasile Sinescu (UNSW) Weights for QMC in Statistics MCQMC February 2012 1 / 24 Quasi-Monte Carlo Methods for Applications

More information

Adapting quasi-monte Carlo methods to simulation problems in weighted Korobov spaces

Adapting quasi-monte Carlo methods to simulation problems in weighted Korobov spaces Adapting quasi-monte Carlo methods to simulation problems in weighted Korobov spaces Christian Irrgeher joint work with G. Leobacher RICAM Special Semester Workshop 1 Uniform distribution and quasi-monte

More information

QMC methods in quantitative finance. and perspectives

QMC methods in quantitative finance. and perspectives , tradition and perspectives Johannes Kepler University Linz (JKU) WU Research Seminar What is the content of the talk Valuation of financial derivatives in stochastic market models using (QMC-)simulation

More information

APPLIED MATHEMATICS REPORT AMR04/16 FINITE-ORDER WEIGHTS IMPLY TRACTABILITY OF MULTIVARIATE INTEGRATION. I.H. Sloan, X. Wang and H.

APPLIED MATHEMATICS REPORT AMR04/16 FINITE-ORDER WEIGHTS IMPLY TRACTABILITY OF MULTIVARIATE INTEGRATION. I.H. Sloan, X. Wang and H. APPLIED MATHEMATICS REPORT AMR04/16 FINITE-ORDER WEIGHTS IMPLY TRACTABILITY OF MULTIVARIATE INTEGRATION I.H. Sloan, X. Wang and H. Wozniakowski Published in Journal of Complexity, Volume 20, Number 1,

More information

Application of QMC methods to PDEs with random coefficients

Application of QMC methods to PDEs with random coefficients Application of QMC methods to PDEs with random coefficients a survey of analysis and implementation Frances Kuo f.kuo@unsw.edu.au University of New South Wales, Sydney, Australia joint work with Ivan Graham

More information

Part III. Quasi Monte Carlo methods 146/349

Part III. Quasi Monte Carlo methods 146/349 Part III Quasi Monte Carlo methods 46/349 Outline Quasi Monte Carlo methods 47/349 Quasi Monte Carlo methods Let y = (y,...,y N ) be a vector of independent uniform random variables in Γ = [0,] N, u(y)

More information

New Multilevel Algorithms Based on Quasi-Monte Carlo Point Sets

New Multilevel Algorithms Based on Quasi-Monte Carlo Point Sets New Multilevel Based on Quasi-Monte Carlo Point Sets Michael Gnewuch Institut für Informatik Christian-Albrechts-Universität Kiel 1 February 13, 2012 Based on Joint Work with Jan Baldeaux (UTS Sydney)

More information

Kernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.

Kernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt. SINGAPORE SHANGHAI Vol TAIPEI - Interdisciplinary Mathematical Sciences 19 Kernel-based Approximation Methods using MATLAB Gregory Fasshauer Illinois Institute of Technology, USA Michael McCourt University

More information

Lattice rules and sequences for non periodic smooth integrands

Lattice rules and sequences for non periodic smooth integrands Lattice rules and sequences for non periodic smooth integrands dirk.nuyens@cs.kuleuven.be Department of Computer Science KU Leuven, Belgium Joint work with Josef Dick (UNSW) and Friedrich Pillichshammer

More information

Tutorial on quasi-monte Carlo methods

Tutorial on quasi-monte Carlo methods Tutorial on quasi-monte Carlo methods Josef Dick School of Mathematics and Statistics, UNSW, Sydney, Australia josef.dick@unsw.edu.au Comparison: MCMC, MC, QMC Roughly speaking: Markov chain Monte Carlo

More information

Numerical Methods II

Numerical Methods II Numerical Methods II Prof. Mike Giles mike.giles@maths.ox.ac.uk Oxford University Mathematical Institute MC Lecture 13 p. 1 Quasi-Monte Carlo As in lecture 6, quasi-monte Carlo methods offer much greater

More information

CS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation

CS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80

More information

Topic 17. Analysis of Algorithms

Topic 17. Analysis of Algorithms Topic 17 Analysis of Algorithms Analysis of Algorithms- Review Efficiency of an algorithm can be measured in terms of : Time complexity: a measure of the amount of time required to execute an algorithm

More information

Low Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed?

Low Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed? Low Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed? Xiaoqun Wang 1,2 and Ian H. Sloan 2 1 Department of Mathematical Sciences, Tsinghua University, Beijing 100084,

More information

Introduction to Smoothing spline ANOVA models (metamodelling)

Introduction to Smoothing spline ANOVA models (metamodelling) Introduction to Smoothing spline ANOVA models (metamodelling) M. Ratto DYNARE Summer School, Paris, June 215. Joint Research Centre www.jrc.ec.europa.eu Serving society Stimulating innovation Supporting

More information

Integration of permutation-invariant. functions

Integration of permutation-invariant. functions permutation-invariant Markus Weimar Philipps-University Marburg Joint work with Dirk Nuyens and Gowri Suryanarayana (KU Leuven, Belgium) MCQMC2014, Leuven April 06-11, 2014 un Outline Permutation-invariant

More information

Applicability of Quasi-Monte Carlo for lattice systems

Applicability of Quasi-Monte Carlo for lattice systems Applicability of Quasi-Monte Carlo for lattice systems Andreas Ammon 1,2, Tobias Hartung 1,2,Karl Jansen 2, Hernan Leovey 3, Andreas Griewank 3, Michael Müller-Preussker 1 1 Humboldt-University Berlin,

More information

arxiv: v3 [math.na] 18 Sep 2016

arxiv: v3 [math.na] 18 Sep 2016 Infinite-dimensional integration and the multivariate decomposition method F. Y. Kuo, D. Nuyens, L. Plaskota, I. H. Sloan, and G. W. Wasilkowski 9 September 2016 arxiv:1501.05445v3 [math.na] 18 Sep 2016

More information

Hot new directions for QMC research in step with applications

Hot new directions for QMC research in step with applications Hot new directions for QMC research in step with applications Frances Kuo f.kuo@unsw.edu.au University of New South Wales, Sydney, Australia Frances Kuo @ UNSW Australia p.1 Outline Motivating example

More information

Finite Difference Methods for Boundary Value Problems

Finite Difference Methods for Boundary Value Problems Finite Difference Methods for Boundary Value Problems October 2, 2013 () Finite Differences October 2, 2013 1 / 52 Goals Learn steps to approximate BVPs using the Finite Difference Method Start with two-point

More information

2.20 Fall 2018 Math Review

2.20 Fall 2018 Math Review 2.20 Fall 2018 Math Review September 10, 2018 These notes are to help you through the math used in this class. This is just a refresher, so if you never learned one of these topics you should look more

More information

Lifting the Curse of Dimensionality

Lifting the Curse of Dimensionality Lifting the Curse of Dimensionality Frances Y. Kuo and Ian H. Sloan Introduction Richard Bellman [1] coined the phrase the curse of dimensionality to describe the extraordinarily rapid growth in the difficulty

More information

Numerical Solution Techniques in Mechanical and Aerospace Engineering

Numerical Solution Techniques in Mechanical and Aerospace Engineering Numerical Solution Techniques in Mechanical and Aerospace Engineering Chunlei Liang LECTURE 3 Solvers of linear algebraic equations 3.1. Outline of Lecture Finite-difference method for a 2D elliptic PDE

More information

High dimensional integration of kinks and jumps smoothing by preintegration. Andreas Griewank, Frances Y. Kuo, Hernan Leövey and Ian H.

High dimensional integration of kinks and jumps smoothing by preintegration. Andreas Griewank, Frances Y. Kuo, Hernan Leövey and Ian H. High dimensional integration of kinks and jumps smoothing by preintegration Andreas Griewank, Frances Y. Kuo, Hernan Leövey and Ian H. Sloan August 2017 arxiv:1712.00920v1 [math.na] 4 Dec 2017 Abstract

More information

ANOVA decomposition of convex piecewise linear functions

ANOVA decomposition of convex piecewise linear functions ANOVA decomposition of convex piecewise linear functions W. Römisch Abstract Piecewise linear convex functions arise as integrands in stochastic programs. They are Lipschitz continuous on their domain,

More information

CSE 421 Algorithms. T(n) = at(n/b) + n c. Closest Pair Problem. Divide and Conquer Algorithms. What you really need to know about recurrences

CSE 421 Algorithms. T(n) = at(n/b) + n c. Closest Pair Problem. Divide and Conquer Algorithms. What you really need to know about recurrences CSE 421 Algorithms Richard Anderson Lecture 13 Divide and Conquer What you really need to know about recurrences Work per level changes geometrically with the level Geometrically increasing (x > 1) The

More information

Improved Discrepancy Bounds for Hybrid Sequences. Harald Niederreiter. RICAM Linz and University of Salzburg

Improved Discrepancy Bounds for Hybrid Sequences. Harald Niederreiter. RICAM Linz and University of Salzburg Improved Discrepancy Bounds for Hybrid Sequences Harald Niederreiter RICAM Linz and University of Salzburg MC vs. QMC methods Hybrid sequences The basic sequences Deterministic discrepancy bounds The proof

More information

An Introduction to Differential Algebra

An Introduction to Differential Algebra An Introduction to Differential Algebra Alexander Wittig1, P. Di Lizia, R. Armellin, et al. 1 ESA Advanced Concepts Team (TEC-SF) SRL, Milan Dinamica Outline 1 Overview Five Views of Differential Algebra

More information

On the Behavior of the Weighted Star Discrepancy Bounds for Shifted Lattice Rules

On the Behavior of the Weighted Star Discrepancy Bounds for Shifted Lattice Rules On the Behavior of the Weighted Star Discrepancy Bounds for Shifted Lattice Rules Vasile Sinescu and Pierre L Ecuyer Abstract We examine the question of constructing shifted lattice rules of rank one with

More information

These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop

These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop Music and Machine Learning (IFT68 Winter 8) Prof. Douglas Eck, Université de Montréal These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop

More information

Optimal Randomized Algorithms for Integration on Function Spaces with underlying ANOVA decomposition

Optimal Randomized Algorithms for Integration on Function Spaces with underlying ANOVA decomposition Optimal Randomized on Function Spaces with underlying ANOVA decomposition Michael Gnewuch 1 University of Kaiserslautern, Germany October 16, 2013 Based on Joint Work with Jan Baldeaux (UTS Sydney) & Josef

More information

Differentiation and Integration

Differentiation and Integration Differentiation and Integration (Lectures on Numerical Analysis for Economists II) Jesús Fernández-Villaverde 1 and Pablo Guerrón 2 February 12, 2018 1 University of Pennsylvania 2 Boston College Motivation

More information

What s new in high-dimensional integration? designing for applications

What s new in high-dimensional integration? designing for applications What s new in high-dimensional integration? designing for applications Ian H. Sloan i.sloan@unsw.edu.au The University of New South Wales UNSW Australia ANZIAM NSW/ACT, November, 2015 The theme High dimensional

More information

Making Models with Polynomials: Taylor Series

Making Models with Polynomials: Taylor Series Making Models with Polynomials: Taylor Series Why polynomials? i a 3 x 3 + a 2 x 2 + a 1 x + a 0 How do we write a general degree n polynomial? n a i x i i=0 Why polynomials and not something else? We

More information

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

Controlled Diffusions and Hamilton-Jacobi Bellman Equations Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter

More information

Orthogonality of hat functions in Sobolev spaces

Orthogonality of hat functions in Sobolev spaces 1 Orthogonality of hat functions in Sobolev spaces Ulrich Reif Technische Universität Darmstadt A Strobl, September 18, 27 2 3 Outline: Recap: quasi interpolation Recap: orthogonality of uniform B-splines

More information

Randomized Quasi-Monte Carlo: An Introduction for Practitioners

Randomized Quasi-Monte Carlo: An Introduction for Practitioners Randomized Quasi-Monte Carlo: An Introduction for Practitioners Pierre L Ecuyer Abstract We survey basic ideas and results on randomized quasi-monte Carlo (RQMC) methods, discuss their practical aspects,

More information

STCE. Adjoint Code Design Patterns. Uwe Naumann. RWTH Aachen University, Germany. QuanTech Conference, London, April 2016

STCE. Adjoint Code Design Patterns. Uwe Naumann. RWTH Aachen University, Germany. QuanTech Conference, London, April 2016 Adjoint Code Design Patterns Uwe Naumann RWTH Aachen University, Germany QuanTech Conference, London, April 2016 Outline Why Adjoints? What Are Adjoints? Software Tool Support: dco/c++ Adjoint Code Design

More information

INTRODUCTION TO FINITE ELEMENT METHODS

INTRODUCTION TO FINITE ELEMENT METHODS INTRODUCTION TO FINITE ELEMENT METHODS LONG CHEN Finite element methods are based on the variational formulation of partial differential equations which only need to compute the gradient of a function.

More information

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis

More information

Class notes: Approximation

Class notes: Approximation Class notes: Approximation Introduction Vector spaces, linear independence, subspace The goal of Numerical Analysis is to compute approximations We want to approximate eg numbers in R or C vectors in R

More information

A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS

A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS Statistica Sinica 26 (2016), 1117-1128 doi:http://dx.doi.org/10.5705/ss.202015.0240 A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS Xu He and Peter Z. G. Qian Chinese Academy of Sciences

More information

The Proper Generalized Decomposition: A Functional Analysis Approach

The Proper Generalized Decomposition: A Functional Analysis Approach The Proper Generalized Decomposition: A Functional Analysis Approach Méthodes de réduction de modèle dans le calcul scientifique Main Goal Given a functional equation Au = f It is possible to construct

More information

Optimal Prediction for Radiative Transfer: A New Perspective on Moment Closure

Optimal Prediction for Radiative Transfer: A New Perspective on Moment Closure Optimal Prediction for Radiative Transfer: A New Perspective on Moment Closure Benjamin Seibold MIT Applied Mathematics Mar 02 nd, 2009 Collaborator Martin Frank (TU Kaiserslautern) Partial Support NSF

More information

Progress in high-dimensional numerical integration and its application to stochastic optimization

Progress in high-dimensional numerical integration and its application to stochastic optimization Progress in high-dimensional numerical integration and its application to stochastic optimization W. Römisch Humboldt-University Berlin Department of Mathematics www.math.hu-berlin.de/~romisch Short course,

More information

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Computer Representation of Numbers Counting numbers (unsigned integers) are the numbers 0,

More information

University of Houston, Department of Mathematics Numerical Analysis, Fall 2005

University of Houston, Department of Mathematics Numerical Analysis, Fall 2005 4 Interpolation 4.1 Polynomial interpolation Problem: LetP n (I), n ln, I := [a,b] lr, be the linear space of polynomials of degree n on I, P n (I) := { p n : I lr p n (x) = n i=0 a i x i, a i lr, 0 i

More information

Hierarchical Matrices. Jon Cockayne April 18, 2017

Hierarchical Matrices. Jon Cockayne April 18, 2017 Hierarchical Matrices Jon Cockayne April 18, 2017 1 Sources Introduction to Hierarchical Matrices with Applications [Börm et al., 2003] 2 Sources Introduction to Hierarchical Matrices with Applications

More information

Kernel Methods. Machine Learning A W VO

Kernel Methods. Machine Learning A W VO Kernel Methods Machine Learning A 708.063 07W VO Outline 1. Dual representation 2. The kernel concept 3. Properties of kernels 4. Examples of kernel machines Kernel PCA Support vector regression (Relevance

More information

Discretization of SDEs: Euler Methods and Beyond

Discretization of SDEs: Euler Methods and Beyond Discretization of SDEs: Euler Methods and Beyond 09-26-2006 / PRisMa 2006 Workshop Outline Introduction 1 Introduction Motivation Stochastic Differential Equations 2 The Time Discretization of SDEs Monte-Carlo

More information

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA Yoshua Bengio Pascal Vincent Jean-François Paiement University of Montreal April 2, Snowbird Learning 2003 Learning Modal Structures

More information

Schwarz Preconditioner for the Stochastic Finite Element Method

Schwarz Preconditioner for the Stochastic Finite Element Method Schwarz Preconditioner for the Stochastic Finite Element Method Waad Subber 1 and Sébastien Loisel 2 Preprint submitted to DD22 conference 1 Introduction The intrusive polynomial chaos approach for uncertainty

More information

Learning gradients: prescriptive models

Learning gradients: prescriptive models Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan

More information

Scenario generation in stochastic programming with application to energy systems

Scenario generation in stochastic programming with application to energy systems Scenario generation in stochastic programming with application to energy systems W. Römisch Humboldt-University Berlin Institute of Mathematics www.math.hu-berlin.de/~romisch SESO 2017 International Thematic

More information

Hochdimensionale Integration

Hochdimensionale Integration Oliver Ernst Institut für Numerische Mathematik und Optimierung Hochdimensionale Integration 14-tägige Vorlesung im Wintersemester 2010/11 im Rahmen des Moduls Ausgewählte Kapitel der Numerik Contents

More information

FAST QMC MATRIX-VECTOR MULTIPLICATION

FAST QMC MATRIX-VECTOR MULTIPLICATION FAST QMC MATRIX-VECTOR MULTIPLICATION JOSEF DICK, FRANCES Y. KUO, QUOC T. LE GIA, CHRISTOPH SCHWAB Abstract. Quasi-Monte Carlo (QMC) rules 1/N N 1 n=0 f(y na) can be used to approximate integrals of the

More information

Quantum Searching. Robert-Jan Slager and Thomas Beuman. 24 november 2009

Quantum Searching. Robert-Jan Slager and Thomas Beuman. 24 november 2009 Quantum Searching Robert-Jan Slager and Thomas Beuman 24 november 2009 1 Introduction Quantum computers promise a significant speed-up over classical computers, since calculations can be done simultaneously.

More information

Lecture 1. Finite difference and finite element methods. Partial differential equations (PDEs) Solving the heat equation numerically

Lecture 1. Finite difference and finite element methods. Partial differential equations (PDEs) Solving the heat equation numerically Finite difference and finite element methods Lecture 1 Scope of the course Analysis and implementation of numerical methods for pricing options. Models: Black-Scholes, stochastic volatility, exponential

More information

Quasi-Monte Carlo methods with applications in finance

Quasi-Monte Carlo methods with applications in finance Finance Stoch (2009) 13: 307 349 DOI 10.1007/s00780-009-0095-y Quasi-Monte Carlo methods with applications in finance Pierre L Ecuyer Received: 30 May 2008 / Accepted: 20 August 2008 / Published online:

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Weierstraß-Institut. für Angewandte Analysis und Stochastik. Leibniz-Institut im Forschungsverbund Berlin e. V. Preprint ISSN

Weierstraß-Institut. für Angewandte Analysis und Stochastik. Leibniz-Institut im Forschungsverbund Berlin e. V. Preprint ISSN Weierstraß-Institut für Angewandte Analysis und Stochastik Leibniz-Institut im Forschungsverbund Berlin e. V. Preprint ISSN 2198-5855 Are Quasi-Monte Carlo algorithms efficient for two-stage stochastic

More information

THE WAVE EQUATION. d = 1: D Alembert s formula We begin with the initial value problem in 1 space dimension { u = utt u xx = 0, in (0, ) R, (2)

THE WAVE EQUATION. d = 1: D Alembert s formula We begin with the initial value problem in 1 space dimension { u = utt u xx = 0, in (0, ) R, (2) THE WAVE EQUATION () The free wave equation takes the form u := ( t x )u = 0, u : R t R d x R In the literature, the operator := t x is called the D Alembertian on R +d. Later we shall also consider the

More information

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 = Chapter 5 Sequences and series 5. Sequences Definition 5. (Sequence). A sequence is a function which is defined on the set N of natural numbers. Since such a function is uniquely determined by its values

More information

Poisson Solvers. William McLean. April 21, Return to Math3301/Math5315 Common Material.

Poisson Solvers. William McLean. April 21, Return to Math3301/Math5315 Common Material. Poisson Solvers William McLean April 21, 2004 Return to Math3301/Math5315 Common Material 1 Introduction Many problems in applied mathematics lead to a partial differential equation of the form a 2 u +

More information

Introduction and Review of Power Series

Introduction and Review of Power Series Introduction and Review of Power Series Definition: A power series in powers of x a is an infinite series of the form c n (x a) n = c 0 + c 1 (x a) + c 2 (x a) 2 +...+c n (x a) n +... If a = 0, this is

More information

10.1 Sequences. Example: A sequence is a function f(n) whose domain is a subset of the integers. Notation: *Note: n = 0 vs. n = 1.

10.1 Sequences. Example: A sequence is a function f(n) whose domain is a subset of the integers. Notation: *Note: n = 0 vs. n = 1. 10.1 Sequences Example: A sequence is a function f(n) whose domain is a subset of the integers. Notation: *Note: n = 0 vs. n = 1 Examples: EX1: Find a formula for the general term a n of the sequence,

More information

SCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES. Kinjal Basu Art B. Owen

SCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES. Kinjal Basu Art B. Owen SCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES By Kinjal Basu Art B. Owen Technical Report No. 2015-07 March 2015 Department of Statistics STANFORD UNIVERSITY Stanford, California 94305-4065

More information

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count

More information

Introduction CSE 541

Introduction CSE 541 Introduction CSE 541 1 Numerical methods Solving scientific/engineering problems using computers. Root finding, Chapter 3 Polynomial Interpolation, Chapter 4 Differentiation, Chapter 4 Integration, Chapters

More information

Importance Sampling to Accelerate the Convergence of Quasi-Monte Carlo

Importance Sampling to Accelerate the Convergence of Quasi-Monte Carlo Importance Sampling to Accelerate the Convergence of Quasi-Monte Carlo Wolfgang Hörmann, Josef Leydold Department of Statistics and Mathematics Wirtschaftsuniversität Wien Research Report Series Report

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

Variance Reduction s Greatest Hits

Variance Reduction s Greatest Hits 1 Variance Reduction s Greatest Hits Pierre L Ecuyer CIRRELT, GERAD, and Département d Informatique et de Recherche Opérationnelle Université de Montréal, Canada ESM 2007, Malta, October 22, 2007 Context

More information

Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space

Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space Pierre L Ecuyer 1, Christian Lécot 2, and Bruno Tuffin 3 1 Département d informatique et de recherche opérationnelle,

More information

Numerical Methods. King Saud University

Numerical Methods. King Saud University Numerical Methods King Saud University Aims In this lecture, we will... find the approximate solutions of derivative (first- and second-order) and antiderivative (definite integral only). Numerical Differentiation

More information

Kernel Method: Data Analysis with Positive Definite Kernels

Kernel Method: Data Analysis with Positive Definite Kernels Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University

More information

Stability of optimization problems with stochastic dominance constraints

Stability of optimization problems with stochastic dominance constraints Stability of optimization problems with stochastic dominance constraints D. Dentcheva and W. Römisch Stevens Institute of Technology, Hoboken Humboldt-University Berlin www.math.hu-berlin.de/~romisch SIAM

More information

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Jim Lambers MAT 610 Summer Session Lecture 2 Notes Jim Lambers MAT 610 Summer Session 2009-10 Lecture 2 Notes These notes correspond to Sections 2.2-2.4 in the text. Vector Norms Given vectors x and y of length one, which are simply scalars x and y, the

More information

FILTERING IN THE FREQUENCY DOMAIN

FILTERING IN THE FREQUENCY DOMAIN 1 FILTERING IN THE FREQUENCY DOMAIN Lecture 4 Spatial Vs Frequency domain 2 Spatial Domain (I) Normal image space Changes in pixel positions correspond to changes in the scene Distances in I correspond

More information

Malliavin Calculus in Finance

Malliavin Calculus in Finance Malliavin Calculus in Finance Peter K. Friz 1 Greeks and the logarithmic derivative trick Model an underlying assent by a Markov process with values in R m with dynamics described by the SDE dx t = b(x

More information

Neural Networks Learning the network: Backprop , Fall 2018 Lecture 4

Neural Networks Learning the network: Backprop , Fall 2018 Lecture 4 Neural Networks Learning the network: Backprop 11-785, Fall 2018 Lecture 4 1 Recap: The MLP can represent any function The MLP can be constructed to represent anything But how do we construct it? 2 Recap:

More information

Partial Differential Equations

Partial Differential Equations Part II Partial Differential Equations Year 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2015 Paper 4, Section II 29E Partial Differential Equations 72 (a) Show that the Cauchy problem for u(x,

More information

Application of Taylor Models to the Worst-Case Analysis of Stripline Interconnects

Application of Taylor Models to the Worst-Case Analysis of Stripline Interconnects Application of Taylor Models to the Worst-Case Analysis of Stripline Interconnects Paolo Manfredi, Riccardo Trinchero, Igor Stievano, Flavio Canavero Email: paolo.manfredi@ugent.be riccardo.trinchero@to.infn.it,

More information

A Belgian view on lattice rules

A Belgian view on lattice rules A Belgian view on lattice rules Ronald Cools Dept. of Computer Science, KU Leuven Linz, Austria, October 14 18, 2013 Atomium Brussels built in 1958 height 103m figure = 2e coin 5 10 6 in circulation Body

More information

2 Two-Point Boundary Value Problems

2 Two-Point Boundary Value Problems 2 Two-Point Boundary Value Problems Another fundamental equation, in addition to the heat eq. and the wave eq., is Poisson s equation: n j=1 2 u x 2 j The unknown is the function u = u(x 1, x 2,..., x

More information

Error Analysis for Quasi-Monte Carlo Methods

Error Analysis for Quasi-Monte Carlo Methods Error Analysis for Quasi-Monte Carlo Methods Fred J. Hickernell Department of Applied Mathematics, Illinois Institute of Technology hickernell@iit.edu mypages.iit.edu/~hickernell Thanks to the Guaranteed

More information

STRONG TRACTABILITY OF MULTIVARIATE INTEGRATION USING QUASI MONTE CARLO ALGORITHMS

STRONG TRACTABILITY OF MULTIVARIATE INTEGRATION USING QUASI MONTE CARLO ALGORITHMS MATHEMATICS OF COMPUTATION Volume 72, Number 242, Pages 823 838 S 0025-5718(02)01440-0 Article electronically published on June 25, 2002 STRONG TRACTABILITY OF MULTIVARIATE INTEGRATION USING QUASI MONTE

More information

Quasi-Monte Carlo Algorithms with Applications in Numerical Analysis and Finance

Quasi-Monte Carlo Algorithms with Applications in Numerical Analysis and Finance Quasi-Monte Carlo Algorithms with Applications in Numerical Analysis and Finance Reinhold F. Kainhofer Rigorosumsvortrag, Institut für Mathematik Graz, 16. Mai 2003 Inhalt 1. Quasi-Monte Carlo Methoden

More information

Ian Sloan and Lattice Rules

Ian Sloan and Lattice Rules www.oeaw.ac.at Ian Sloan and Lattice Rules P. Kritzer, H. Niederreiter, F. Pillichshammer RICAM-Report 2016-34 www.ricam.oeaw.ac.at Ian Sloan and Lattice Rules Peter Kritzer, Harald Niederreiter, Friedrich

More information

Before you begin read these instructions carefully.

Before you begin read these instructions carefully. MATHEMATICAL TRIPOS Part IA Tuesday, 4 June, 2013 1:30 pm to 4:30 pm PAPER 3 Before you begin read these instructions carefully. The examination paper is divided into two sections. Each question in Section

More information

Convergence of Particle Filtering Method for Nonlinear Estimation of Vortex Dynamics

Convergence of Particle Filtering Method for Nonlinear Estimation of Vortex Dynamics Convergence of Particle Filtering Method for Nonlinear Estimation of Vortex Dynamics Meng Xu Department of Mathematics University of Wyoming February 20, 2010 Outline 1 Nonlinear Filtering Stochastic Vortex

More information

ABSTRACT NONSINGULAR CURVES

ABSTRACT NONSINGULAR CURVES ABSTRACT NONSINGULAR CURVES Affine Varieties Notation. Let k be a field, such as the rational numbers Q or the complex numbers C. We call affine n-space the collection A n k of points P = a 1, a,..., a

More information

Multi-Factor Finite Differences

Multi-Factor Finite Differences February 17, 2017 Aims and outline Finite differences for more than one direction The θ-method, explicit, implicit, Crank-Nicolson Iterative solution of discretised equations Alternating directions implicit

More information

CONVERGENCE THEORY. G. ALLAIRE CMAP, Ecole Polytechnique. 1. Maximum principle. 2. Oscillating test function. 3. Two-scale convergence

CONVERGENCE THEORY. G. ALLAIRE CMAP, Ecole Polytechnique. 1. Maximum principle. 2. Oscillating test function. 3. Two-scale convergence 1 CONVERGENCE THEOR G. ALLAIRE CMAP, Ecole Polytechnique 1. Maximum principle 2. Oscillating test function 3. Two-scale convergence 4. Application to homogenization 5. General theory H-convergence) 6.

More information

Dynamic Risk Measures and Nonlinear Expectations with Markov Chain noise

Dynamic Risk Measures and Nonlinear Expectations with Markov Chain noise Dynamic Risk Measures and Nonlinear Expectations with Markov Chain noise Robert J. Elliott 1 Samuel N. Cohen 2 1 Department of Commerce, University of South Australia 2 Mathematical Insitute, University

More information

Introduction to (randomized) quasi-monte Carlo

Introduction to (randomized) quasi-monte Carlo 1 aft Introduction to (randomized) quasi-monte Carlo Dr Pierre L Ecuyer MCQMC Conference, Stanford University, August 2016 Program Monte Carlo, Quasi-Monte Carlo, Randomized quasi-monte Carlo QMC point

More information

Stochastic Spectral Approaches to Bayesian Inference

Stochastic Spectral Approaches to Bayesian Inference Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to

More information

Algorithms for Uncertainty Quantification

Algorithms for Uncertainty Quantification Algorithms for Uncertainty Quantification Lecture 9: Sensitivity Analysis ST 2018 Tobias Neckel Scientific Computing in Computer Science TUM Repetition of Previous Lecture Sparse grids in Uncertainty Quantification

More information

Trivariate polynomial approximation on Lissajous curves 1

Trivariate polynomial approximation on Lissajous curves 1 Trivariate polynomial approximation on Lissajous curves 1 Stefano De Marchi DMV2015, Minisymposium on Mathematical Methods for MPI 22 September 2015 1 Joint work with Len Bos (Verona) and Marco Vianello

More information

State Space Representation of Gaussian Processes

State Space Representation of Gaussian Processes State Space Representation of Gaussian Processes Simo Särkkä Department of Biomedical Engineering and Computational Science (BECS) Aalto University, Espoo, Finland June 12th, 2013 Simo Särkkä (Aalto University)

More information