Fast evaluation of mixed derivatives and calculation of optimal weights for integration. Hernan Leovey
|
|
- Abigail Chambers
- 5 years ago
- Views:
Transcription
1 Fast evaluation of mixed derivatives and calculation of optimal weights for integration Humboldt Universität zu Berlin MCQMC2012 Tenth International Conference on Monte Carlo and Quasi Monte Carlo Methods in Scientific Computing
2 Contents Algorithmic Differentiation (AD) 1 Algorithmic Differentiation (AD) Basics Complexity 2 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions 3 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension 4
3 Basics Algorithmic Differentiation (AD) Basics Complexity Frequently we have a program that calculates numerical values for a function, and we would like to obtain accurate values for derivatives of the function as well. The usual divided difference approach is given by : D +h f (x) f (x + h) f (x) h or D ±h f (x) f (x + h) f (x h) 2h For h small, truncation and round-off errors reduce the number of significant digits If h is not small, normally no good approximation to a derivative is expected
4 Basics Algorithmic Differentiation (AD) Basics Complexity Typically h = ɛ is taken, for ɛ = working accuracy. Expected accuracy: 1 2 of the significant digits of f for D +h 2 3 of the significant digits of f for D ±h In contrast, AD methods incur no truncation errors at all and usually yield derivatives with working accuracy. AD 0: Algorithmic Differentiation does not incur truncation errors AD 1: Difference quotients may sometimes be useful too AD 2: What is good for function values is good for their derivatives
5 Basics Algorithmic Differentiation (AD) Basics Complexity Standard setting for AD: Vector function F is the composition of a sequence of once continuously differentiable elemental functions ϕ i. Basic set of functions (polynomial core): {+,, (unary sign op.), c (const. init.)} A typical example of a library containing elemental functions: {c, +,,, /, exp, log, sin, cos, tan, tan 1,..., Φ, Φ 1,...}
6 Basics Complexity
7 Basics Complexity Basic complexity results Consider temporal complexity measure TIME, TIME{task(F )} = w WORK{task(F )} (1) with w = (w 1, w 2, w 3, w 4 ) a vector of platform dependent weights, and MOVES of fetches and stores WORK{task} ADDS MULTS of additions and subtractions of multiplications (2) NLOPS of nonlinear operations Forward mode AD: TIME{F (x), F (x)ẋ} ω tang TIME{F (x)} with a constant ω tang [2, 5/2]
8 Basics Complexity Reverse mode AD: Cheap Gradient Principle TIME{F (x), ȳ F (x)} ω grad TIME{F (x)} (3) for a constant ω grad [3, 4]. As consequence, the cost to evaluate a gradient f is bounded above by a small constant ω grad [3, 4] times the cost to evaluate the function itself. Random Access Memory requirements, in forward and reverse, are bounded multiples of those for the functions. Sequential Access Memory requirement of basic reverse mode is proportional to temporal complexity of the function.
9 Algorithmic Differentiation (AD) Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions (Automatic Evaluations of Cross-Derivatives. Griewank, L, L, Z) With the term cross derivatives we refer to those mixed partial derivatives where differentiation w.r.t. each variable is done at most once. f i (x) = j i x j f (x) = k f x i1... x ik (x), i = {i 1, i 2,..., i k }. There are 2 d cross derivatives if we take f (x) = f (x). We create a data structure with all 2 d cross derivatives of a function u in a flat array with 2 d entries. We call such data structure an d dimensional cube. d = 3 u u {1} u {2} u {1,2} u {3} u {1,3} u {2,3} u {1,2,3}
10 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions Basic Operations For a function u we denote by U its cube. For a constant function u(x) = c we set U[0] = c and zero everywhere else. For a coordinate function reps. input variable u(x) = x j we initialize its cube by U[0] = x j and U[2 j ] = 1. The rest of the entries are set to zero. Addition and Subtraction: V[i]=U[i] ± W[i] for all 0 i < 2 d. Scalar Multiplication: For v(x) = cu(x) the propagation rule is V[i]=c*U[i]. Scalar Addition/Subtraction is applied only to U[0]. The complexity of the above basic operations is (O(2 d )).
11 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions Nonlinear Operations Multiplication: The Leibniz formula for the multiplication of two functions v = u w states that: v i (x) = j i u j (x)w i j (x). Assume now that n / i. Then the above convolution sum can be split into v i {n} (x) = j i u i j (x)w j {n} (x) + j i u j {n} (x)w i j (x) Fixing the same subset i, the sums have the same structure. They all operate inside separate halves of cubes.
12 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions This leads to a possible implementation: void crossmult (int h, double U, double W, double V) { if (h == 1) { V[0] + = (U[0] W[0]); return; } h/ = 2; crossmult(h,u,w+h,v+h); crossmult(h,u+h,w,v+h); crossmult(h,u,w,v); } Due to the recursive nature of this procedure, there will be 3 d overall function calls at h = 1 resulting in 3 d multiplications and the same number of additions.
13 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions Exponential function: v = exp(u) has a very simple identity for the first partial derivatives, v k = vu k. This generalizes for k / i to: v i {k} = j i v i j (x)u j {k} (x) The second half cube of v is thus obtained by multiplying the previously computed first half cube of v and the second half cube of u. void exponent(int h, double U, double V) { int i; for(i= 0;i<h;i++) { V[i] = 0.0; } V[0]=exp(U[0]); for(i=1;i<h;i =2) { crossmult (i,v,u+i,v+i); } } There are d calls to the multiplication function and the final relative cost is 1 2 of the cost of a full multiplication.
14 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions Complexity: Nonlinear differentiable functions ϕ(u) included in math.h exhibit cost proportional to cross multiplication Given a library exhibiting cost proportional to cross multiplication, extend it by considering any nonlinear ϕ(u) satisfying differential equation ϕ (u) a(u)ϕ(u) = b(u) with functions a(.), b(.) in original library (ODE extension). Proposition The direct computation of all cross derivatives f of a function f given as an evaluation procedure (with elementals in ODE extended library) is itself an evaluation procedure with complexity OPS(f ) = O(3 d ) OPS(f ) for the runtime and with a factor of 2 d in the memory size. The unit is one multiplication, which is also the cost of addition or subtraction.
15 Arithmetic operations and nonlinear functions Complexity Comparison with other methods: univariate Taylor polynomial expansions Comparison with other methods: Proposition Method of interpolation of all cross derivatives from Taylor coefficients via univariate expansions exhibits complexity OPS(f ) = O(d 2 2 d ) (OPS(f ) + c), c 4, for the runtime and with a factor of (d + 1) 2 d in the memory size. The cross over between the methods occur at d 14. For large dimensions d, the Taylor method will have better runtimes. Advantages of direct new method: more accurate than Taylor univariate method faster for d 14
16 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Quasi Monte Carlo Methods (QMC): Q N,n (f ) := 1 N N f (x i ) I (f ) := f (x)dx, [0,1] n i=1 with x 1,, x N deterministically and cleverly chosen from [0, 1] n. Lattice Rules Q N,n,z (f ) := 1 N 1 f N i=0 ({ }) i N z Where N (usually prime) is the number of selected points and z is a carefully selected integer vector in Z n. Shifted Lattice Rules Q N,n,z, (f ) := 1 N 1 ({ }) i f N N z + for [0, 1] n. i=0
17 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension (Weighted) Reproducing Kernel Hilbert spaces (Sloan&Woźniakowski 98) Integration over particular RKHS F n of functions over [0, 1] n. Reproducing kernel K n (x, t) is function defined over [0, 1] n [0, 1] n, such that K n (., t) F n for all t [0, 1] n and f (t) = f (.), K n (., t) n, f F n ; t [0, 1] n. Worst Case Error of QMC algorithm over F n e(q N,n ) := sup I (f ) Q N,n (f ) f F n: f Fn 1 Assume integration functional I (.) is continuous over F n, then e(q N,n ) is bounded and I (f ) Q N,n (f ) e(q N,n ). f Fn
18 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Weighted Unanchored Sobolev Space F n,γ Consider weights 0 γ n,i, for i {1,, n}. ( 1 K n,γ (x, y) = 1+ γ n,i 2 B 2({x j y j }) + (x j 1 2 )(y j 1 ) 2 ) f Fn,γ = =i {1,,n} i {1,,n} j i ( ) 2 γ 1 i n,i f (x i, x D i )dx D i dx i [0,1] i [0,1] x n i i 1 2 Product weights γ n,i = j i γ {n,j} Tensor Product RKHS n F n,γ = H n,γ := H 1,γ1 H 1,γn n times K n,γ (x, y) = K 1,γj (x i, y i ) j=1
19 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Theorem (Novak&Woźniakowski 10, Kuo, Sloan, Joe,..) Let 0 γ n,i, i D, D := {1,, n}, f F n,γ. Given a prime number N, there exits a shifted rank-1 lattice rule Q N,n,z, with generator vector z constructed by the Component by Component algorithm (CBC), such that ( ( ) ) i τ =i D (γ n,i) 1/(2τ) 2ζ(1/τ) ( 2π) 1/τ I (f ) Q N,d,z, (f ) (N 1) τ f F n,γ for any τ [ 1 2, 1). For fixed f, we need the weights to construct a generator vector z for a lattice rule, using CBC algorithm. How should we choose the weights in practice? What is an optimal embedding for a given function f in a practical problem?
20 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Common Approach: Choose the weights such that the integration error bound is minimized. In general case, 2 n 1 terms inside f F = n,γ Approach: =i D ( ) 2 γ 1 i n,i f (x i, x D i )dx D i dx i [0,1] i [0,1] x n i i 1 2 Very often, problems in applications exhibit low effective dimension d << n. Effective dimension refers to essential ANOVA part of the function that accumulates most of the variance ( 99%).
21 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Assume f is a square integrable function. Then we can write f as the sum of 2 n ANOVA terms: f (x) = f i (x), f i (x) = f (x i, x D i )dx D i f j (x) i D [0,1] n i j i For a given family T of subsets of D, let us define now f T (x) = i T f i (x). Then, the integration error of a QMC algorithm Q N,n is given by (I Q N,n )(f ) (I Q N,n )(f T ) + (I Q N,n) i {1,...,n},i T f i (x)
22 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Theorem Let T be a given family of subsets of D. Let f i F n,γ for i T. Then for the function f T defined above it holds f T F = n,γ =i T ( γ 1 i ) 2 n,i f (x i, x D i )dx D i dx i [0,1] x i i [0,1] n i Moreover, if f F n,γ, it holds for i D b f,i := = [0,1] i ( i ) 2 f (x i, x D i )dx D i dx i x i [0,1] n i ( ) 2 i f (x i, x D i )dx D i dx i [0,1] x n i i [0,1] i [0,1] n 1 2 ( ) i 2 f (x) dx x i
23 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Integrands with low effective dimension Remark Note that any (good) upper bound b f,i R, b f,i b f,i, conduces also to an integration error upper bound of the form ( ( ) ) i τ i D (γ n,i) 1/(2τ) 2ζ(1/τ) ( 1 2π) 1/τ (I Q N,n )(f T ) 2 γ 1 (N 1) τ n,i b f,i i T Product Weights (γ n,i = j i γ n,{j}): Let d denote effective dimension of f in truncation sense (say d 14). App.-1 f EFFTd := f i (x). i {1,...,d}
24 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Set 0 0 = 0, c 0 = + for c > 0. Assume that at least one term b i,f > 0 for some i {1,..., d}. For f EFFTd consider bound objective function ψ : R n 0 [0, + ] (where the variables are the weights). For simplicity set ψ(0) = +, and for (x 1,..., x n ) R n 0 \ {0} define ( n ( ψ(x 1,..., x n) 2τ = 1 + x 1 2ζ(1/τ) j ( 2π) 1/τ j=1 ) ) τ 1 i {1,...,d}( j i x 1 j )b i,f 1 2 Clearly we have: minimize ψ(x 1,..., x n) minimize ψ(x 1,..., x n) (x 1,...,x n) R n 0 (x 1,...,x n) R n 0 subject to x j = 0, d + 1 j n.
25 Reproducing Kernel Hilbert Spaces (RKHS) Integrands with low effective dimension Choice for non important weights (in ANOVA sense): Lemma For fixed τ [1/2, 1), let γn = (γn,1,..., γn,d, 0,..., 0) be an optimal feasible solution of problem above. Let ɛ 0 > 0, and let a 1,..., a n d be any sequence of nonnegative real numbers with n d i=1 a i M. Define R 0 = ( 2π) 1/τ M2ζ(1/τ) log τ ɛ 1 0 d j=1 (1 + (γ n,j) 1 2τ d j=1 (1 + (γ n,j ) 1 2τ ) 2ζ(1/τ) ( 2π) 1/τ ) 2ζ(1/τ) ( 2π) 1/τ Then it follows for γ n = (γ n,1,..., γ n,d, R 0a 1,..., R 0a n d ) that ψ(γ n) (1 + ɛ 0)ψ(γ n )
26 Option valuation problem for arithmetic average Asian options Asset S t follows the geometric Brownian motion model. ) ) S t = S 0 exp ((r σ2 t + σw t 2 Simulating asset prices reduces to simulating paths W t1,..., W td. e rt V = (2π) d/2 max 1 d S j (w) K, 0 e 1 2 wt C 1w dw det(c) R d d with w = (W t1,..., W td ). After a factorization C = AA T of the covariance matrix, transform integral using Φ 1 (.). V = e [0,1] rt max 1 d S j (AΦ 1 (x)) K, 0 dx, d d j=1 For the tests, we simplify the problem assuming K = 0. Consider principal components (PCA) and Brownian Bridge (BB) factorization of C. j=1
27 Sensitivity tests for effective dimension (Algo. Wang&Fang 03) K = 0, S 0 = 100, T = 1 real dimension n = 16, 64, 128 Domain Truncation (Kuo&Sloan&Griebel 10) ɛ = 0.1, 0.01, 0.001, ( I (f ) I (f ɛ ) ɛs 0 ) For (σ, r) [0.05, 0.35] [0.05, 0.35] (tests on 7 7 uniform grid) Using 2 16 Sobol points, all tests resulted with effective dimension in truncation sense d 3 for PCA, and d 8 for BB construction. ( ) b ˆ i 2 f,i b f,i := f (x) dx for i {1,, d} [0,1] x n i CrossAD cost for simplified examples without strike (K = 0): Example \n = Runtime(crossPCA) (d = 4) Runtime(PCA) Runtime(crossBB) (d = 8) Runtime(BB)
28 Fixed K = 0, S 0 = 100, T = 1,σ = 0.1,r = 0.1, domain truncation ɛ = 0.1, (b i,f estimates using cross AD for d first variables) Table: Weights for τ = 0.9 (runtime for opt. solver approx seconds) BB n8 S11 n8 S14 n8 acc n16 S11 n16 S14 n16 acc n128 S11 n128 S14 γn, γn, γn, γn, γn, γn, γn, γn, ψ(γ ) 2.6e e e e e e e e+04 ψ( 1, 0) 9.5e e e e e e e+05 j 3.5e+05 ψ( 1, 0) 1.0e e e e e e e+04 j 2 5.2e+04
29 Table: Weights for τ = 0.9 (runtime for opt. solver approx seconds) PCA n8 S11 n8 S14 n8 acc n16 S11 n16 S14 n16 acc n128 S11 n128 S14 γn, γn, γn, γn, ψ(γ ) 6.0e e e e e e e e+03 ψ( 1, 0) 4.0e e e e e e e+04 j 1.1e+04 ψ( 1, 0) 1.6e e e e e e e+03 j 2 4.5e+03
30 Further investigations: f EFF + T d := i {1,...,d} f i (x) + d+1 j n f {j} (x). Using cross AD + reverse mode AD for cheap gradients (Optimization problem remains n dimensional) Product and order dependent Weights for functions with low effective superposition dimension using forward and reverse AD (No need for numerical Optimization) Good bounds for functions with kinks (K 0, eff. sup. dim. d = 2 and P.O.D. weights) Improved sampling strategy for squared mixed derivatives strongly diverging at small sub-cube borders Domain truncation alternative Thank you for your attention!
Quasi-Monte Carlo integration over the Euclidean space and applications
Quasi-Monte Carlo integration over the Euclidean space and applications f.kuo@unsw.edu.au University of New South Wales, Sydney, Australia joint work with James Nichols (UNSW) Journal of Complexity 30
More informationQuasi-Monte Carlo Methods for Applications in Statistics
Quasi-Monte Carlo Methods for Applications in Statistics Weights for QMC in Statistics Vasile Sinescu (UNSW) Weights for QMC in Statistics MCQMC February 2012 1 / 24 Quasi-Monte Carlo Methods for Applications
More informationAdapting quasi-monte Carlo methods to simulation problems in weighted Korobov spaces
Adapting quasi-monte Carlo methods to simulation problems in weighted Korobov spaces Christian Irrgeher joint work with G. Leobacher RICAM Special Semester Workshop 1 Uniform distribution and quasi-monte
More informationQMC methods in quantitative finance. and perspectives
, tradition and perspectives Johannes Kepler University Linz (JKU) WU Research Seminar What is the content of the talk Valuation of financial derivatives in stochastic market models using (QMC-)simulation
More informationAPPLIED MATHEMATICS REPORT AMR04/16 FINITE-ORDER WEIGHTS IMPLY TRACTABILITY OF MULTIVARIATE INTEGRATION. I.H. Sloan, X. Wang and H.
APPLIED MATHEMATICS REPORT AMR04/16 FINITE-ORDER WEIGHTS IMPLY TRACTABILITY OF MULTIVARIATE INTEGRATION I.H. Sloan, X. Wang and H. Wozniakowski Published in Journal of Complexity, Volume 20, Number 1,
More informationApplication of QMC methods to PDEs with random coefficients
Application of QMC methods to PDEs with random coefficients a survey of analysis and implementation Frances Kuo f.kuo@unsw.edu.au University of New South Wales, Sydney, Australia joint work with Ivan Graham
More informationPart III. Quasi Monte Carlo methods 146/349
Part III Quasi Monte Carlo methods 46/349 Outline Quasi Monte Carlo methods 47/349 Quasi Monte Carlo methods Let y = (y,...,y N ) be a vector of independent uniform random variables in Γ = [0,] N, u(y)
More informationNew Multilevel Algorithms Based on Quasi-Monte Carlo Point Sets
New Multilevel Based on Quasi-Monte Carlo Point Sets Michael Gnewuch Institut für Informatik Christian-Albrechts-Universität Kiel 1 February 13, 2012 Based on Joint Work with Jan Baldeaux (UTS Sydney)
More informationKernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.
SINGAPORE SHANGHAI Vol TAIPEI - Interdisciplinary Mathematical Sciences 19 Kernel-based Approximation Methods using MATLAB Gregory Fasshauer Illinois Institute of Technology, USA Michael McCourt University
More informationLattice rules and sequences for non periodic smooth integrands
Lattice rules and sequences for non periodic smooth integrands dirk.nuyens@cs.kuleuven.be Department of Computer Science KU Leuven, Belgium Joint work with Josef Dick (UNSW) and Friedrich Pillichshammer
More informationTutorial on quasi-monte Carlo methods
Tutorial on quasi-monte Carlo methods Josef Dick School of Mathematics and Statistics, UNSW, Sydney, Australia josef.dick@unsw.edu.au Comparison: MCMC, MC, QMC Roughly speaking: Markov chain Monte Carlo
More informationNumerical Methods II
Numerical Methods II Prof. Mike Giles mike.giles@maths.ox.ac.uk Oxford University Mathematical Institute MC Lecture 13 p. 1 Quasi-Monte Carlo As in lecture 6, quasi-monte Carlo methods offer much greater
More informationCS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation
Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80
More informationTopic 17. Analysis of Algorithms
Topic 17 Analysis of Algorithms Analysis of Algorithms- Review Efficiency of an algorithm can be measured in terms of : Time complexity: a measure of the amount of time required to execute an algorithm
More informationLow Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed?
Low Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed? Xiaoqun Wang 1,2 and Ian H. Sloan 2 1 Department of Mathematical Sciences, Tsinghua University, Beijing 100084,
More informationIntroduction to Smoothing spline ANOVA models (metamodelling)
Introduction to Smoothing spline ANOVA models (metamodelling) M. Ratto DYNARE Summer School, Paris, June 215. Joint Research Centre www.jrc.ec.europa.eu Serving society Stimulating innovation Supporting
More informationIntegration of permutation-invariant. functions
permutation-invariant Markus Weimar Philipps-University Marburg Joint work with Dirk Nuyens and Gowri Suryanarayana (KU Leuven, Belgium) MCQMC2014, Leuven April 06-11, 2014 un Outline Permutation-invariant
More informationApplicability of Quasi-Monte Carlo for lattice systems
Applicability of Quasi-Monte Carlo for lattice systems Andreas Ammon 1,2, Tobias Hartung 1,2,Karl Jansen 2, Hernan Leovey 3, Andreas Griewank 3, Michael Müller-Preussker 1 1 Humboldt-University Berlin,
More informationarxiv: v3 [math.na] 18 Sep 2016
Infinite-dimensional integration and the multivariate decomposition method F. Y. Kuo, D. Nuyens, L. Plaskota, I. H. Sloan, and G. W. Wasilkowski 9 September 2016 arxiv:1501.05445v3 [math.na] 18 Sep 2016
More informationHot new directions for QMC research in step with applications
Hot new directions for QMC research in step with applications Frances Kuo f.kuo@unsw.edu.au University of New South Wales, Sydney, Australia Frances Kuo @ UNSW Australia p.1 Outline Motivating example
More informationFinite Difference Methods for Boundary Value Problems
Finite Difference Methods for Boundary Value Problems October 2, 2013 () Finite Differences October 2, 2013 1 / 52 Goals Learn steps to approximate BVPs using the Finite Difference Method Start with two-point
More information2.20 Fall 2018 Math Review
2.20 Fall 2018 Math Review September 10, 2018 These notes are to help you through the math used in this class. This is just a refresher, so if you never learned one of these topics you should look more
More informationLifting the Curse of Dimensionality
Lifting the Curse of Dimensionality Frances Y. Kuo and Ian H. Sloan Introduction Richard Bellman [1] coined the phrase the curse of dimensionality to describe the extraordinarily rapid growth in the difficulty
More informationNumerical Solution Techniques in Mechanical and Aerospace Engineering
Numerical Solution Techniques in Mechanical and Aerospace Engineering Chunlei Liang LECTURE 3 Solvers of linear algebraic equations 3.1. Outline of Lecture Finite-difference method for a 2D elliptic PDE
More informationHigh dimensional integration of kinks and jumps smoothing by preintegration. Andreas Griewank, Frances Y. Kuo, Hernan Leövey and Ian H.
High dimensional integration of kinks and jumps smoothing by preintegration Andreas Griewank, Frances Y. Kuo, Hernan Leövey and Ian H. Sloan August 2017 arxiv:1712.00920v1 [math.na] 4 Dec 2017 Abstract
More informationANOVA decomposition of convex piecewise linear functions
ANOVA decomposition of convex piecewise linear functions W. Römisch Abstract Piecewise linear convex functions arise as integrands in stochastic programs. They are Lipschitz continuous on their domain,
More informationCSE 421 Algorithms. T(n) = at(n/b) + n c. Closest Pair Problem. Divide and Conquer Algorithms. What you really need to know about recurrences
CSE 421 Algorithms Richard Anderson Lecture 13 Divide and Conquer What you really need to know about recurrences Work per level changes geometrically with the level Geometrically increasing (x > 1) The
More informationImproved Discrepancy Bounds for Hybrid Sequences. Harald Niederreiter. RICAM Linz and University of Salzburg
Improved Discrepancy Bounds for Hybrid Sequences Harald Niederreiter RICAM Linz and University of Salzburg MC vs. QMC methods Hybrid sequences The basic sequences Deterministic discrepancy bounds The proof
More informationAn Introduction to Differential Algebra
An Introduction to Differential Algebra Alexander Wittig1, P. Di Lizia, R. Armellin, et al. 1 ESA Advanced Concepts Team (TEC-SF) SRL, Milan Dinamica Outline 1 Overview Five Views of Differential Algebra
More informationOn the Behavior of the Weighted Star Discrepancy Bounds for Shifted Lattice Rules
On the Behavior of the Weighted Star Discrepancy Bounds for Shifted Lattice Rules Vasile Sinescu and Pierre L Ecuyer Abstract We examine the question of constructing shifted lattice rules of rank one with
More informationThese slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop
Music and Machine Learning (IFT68 Winter 8) Prof. Douglas Eck, Université de Montréal These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop
More informationOptimal Randomized Algorithms for Integration on Function Spaces with underlying ANOVA decomposition
Optimal Randomized on Function Spaces with underlying ANOVA decomposition Michael Gnewuch 1 University of Kaiserslautern, Germany October 16, 2013 Based on Joint Work with Jan Baldeaux (UTS Sydney) & Josef
More informationDifferentiation and Integration
Differentiation and Integration (Lectures on Numerical Analysis for Economists II) Jesús Fernández-Villaverde 1 and Pablo Guerrón 2 February 12, 2018 1 University of Pennsylvania 2 Boston College Motivation
More informationWhat s new in high-dimensional integration? designing for applications
What s new in high-dimensional integration? designing for applications Ian H. Sloan i.sloan@unsw.edu.au The University of New South Wales UNSW Australia ANZIAM NSW/ACT, November, 2015 The theme High dimensional
More informationMaking Models with Polynomials: Taylor Series
Making Models with Polynomials: Taylor Series Why polynomials? i a 3 x 3 + a 2 x 2 + a 1 x + a 0 How do we write a general degree n polynomial? n a i x i i=0 Why polynomials and not something else? We
More informationControlled Diffusions and Hamilton-Jacobi Bellman Equations
Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationOrthogonality of hat functions in Sobolev spaces
1 Orthogonality of hat functions in Sobolev spaces Ulrich Reif Technische Universität Darmstadt A Strobl, September 18, 27 2 3 Outline: Recap: quasi interpolation Recap: orthogonality of uniform B-splines
More informationRandomized Quasi-Monte Carlo: An Introduction for Practitioners
Randomized Quasi-Monte Carlo: An Introduction for Practitioners Pierre L Ecuyer Abstract We survey basic ideas and results on randomized quasi-monte Carlo (RQMC) methods, discuss their practical aspects,
More informationSTCE. Adjoint Code Design Patterns. Uwe Naumann. RWTH Aachen University, Germany. QuanTech Conference, London, April 2016
Adjoint Code Design Patterns Uwe Naumann RWTH Aachen University, Germany QuanTech Conference, London, April 2016 Outline Why Adjoints? What Are Adjoints? Software Tool Support: dco/c++ Adjoint Code Design
More informationINTRODUCTION TO FINITE ELEMENT METHODS
INTRODUCTION TO FINITE ELEMENT METHODS LONG CHEN Finite element methods are based on the variational formulation of partial differential equations which only need to compute the gradient of a function.
More informationMonte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan
Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis
More informationClass notes: Approximation
Class notes: Approximation Introduction Vector spaces, linear independence, subspace The goal of Numerical Analysis is to compute approximations We want to approximate eg numbers in R or C vectors in R
More informationA CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS
Statistica Sinica 26 (2016), 1117-1128 doi:http://dx.doi.org/10.5705/ss.202015.0240 A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS Xu He and Peter Z. G. Qian Chinese Academy of Sciences
More informationThe Proper Generalized Decomposition: A Functional Analysis Approach
The Proper Generalized Decomposition: A Functional Analysis Approach Méthodes de réduction de modèle dans le calcul scientifique Main Goal Given a functional equation Au = f It is possible to construct
More informationOptimal Prediction for Radiative Transfer: A New Perspective on Moment Closure
Optimal Prediction for Radiative Transfer: A New Perspective on Moment Closure Benjamin Seibold MIT Applied Mathematics Mar 02 nd, 2009 Collaborator Martin Frank (TU Kaiserslautern) Partial Support NSF
More informationProgress in high-dimensional numerical integration and its application to stochastic optimization
Progress in high-dimensional numerical integration and its application to stochastic optimization W. Römisch Humboldt-University Berlin Department of Mathematics www.math.hu-berlin.de/~romisch Short course,
More informationEAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science
EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Computer Representation of Numbers Counting numbers (unsigned integers) are the numbers 0,
More informationUniversity of Houston, Department of Mathematics Numerical Analysis, Fall 2005
4 Interpolation 4.1 Polynomial interpolation Problem: LetP n (I), n ln, I := [a,b] lr, be the linear space of polynomials of degree n on I, P n (I) := { p n : I lr p n (x) = n i=0 a i x i, a i lr, 0 i
More informationHierarchical Matrices. Jon Cockayne April 18, 2017
Hierarchical Matrices Jon Cockayne April 18, 2017 1 Sources Introduction to Hierarchical Matrices with Applications [Börm et al., 2003] 2 Sources Introduction to Hierarchical Matrices with Applications
More informationKernel Methods. Machine Learning A W VO
Kernel Methods Machine Learning A 708.063 07W VO Outline 1. Dual representation 2. The kernel concept 3. Properties of kernels 4. Examples of kernel machines Kernel PCA Support vector regression (Relevance
More informationDiscretization of SDEs: Euler Methods and Beyond
Discretization of SDEs: Euler Methods and Beyond 09-26-2006 / PRisMa 2006 Workshop Outline Introduction 1 Introduction Motivation Stochastic Differential Equations 2 The Time Discretization of SDEs Monte-Carlo
More informationLearning Eigenfunctions: Links with Spectral Clustering and Kernel PCA
Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA Yoshua Bengio Pascal Vincent Jean-François Paiement University of Montreal April 2, Snowbird Learning 2003 Learning Modal Structures
More informationSchwarz Preconditioner for the Stochastic Finite Element Method
Schwarz Preconditioner for the Stochastic Finite Element Method Waad Subber 1 and Sébastien Loisel 2 Preprint submitted to DD22 conference 1 Introduction The intrusive polynomial chaos approach for uncertainty
More informationLearning gradients: prescriptive models
Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan
More informationScenario generation in stochastic programming with application to energy systems
Scenario generation in stochastic programming with application to energy systems W. Römisch Humboldt-University Berlin Institute of Mathematics www.math.hu-berlin.de/~romisch SESO 2017 International Thematic
More informationHochdimensionale Integration
Oliver Ernst Institut für Numerische Mathematik und Optimierung Hochdimensionale Integration 14-tägige Vorlesung im Wintersemester 2010/11 im Rahmen des Moduls Ausgewählte Kapitel der Numerik Contents
More informationFAST QMC MATRIX-VECTOR MULTIPLICATION
FAST QMC MATRIX-VECTOR MULTIPLICATION JOSEF DICK, FRANCES Y. KUO, QUOC T. LE GIA, CHRISTOPH SCHWAB Abstract. Quasi-Monte Carlo (QMC) rules 1/N N 1 n=0 f(y na) can be used to approximate integrals of the
More informationQuantum Searching. Robert-Jan Slager and Thomas Beuman. 24 november 2009
Quantum Searching Robert-Jan Slager and Thomas Beuman 24 november 2009 1 Introduction Quantum computers promise a significant speed-up over classical computers, since calculations can be done simultaneously.
More informationLecture 1. Finite difference and finite element methods. Partial differential equations (PDEs) Solving the heat equation numerically
Finite difference and finite element methods Lecture 1 Scope of the course Analysis and implementation of numerical methods for pricing options. Models: Black-Scholes, stochastic volatility, exponential
More informationQuasi-Monte Carlo methods with applications in finance
Finance Stoch (2009) 13: 307 349 DOI 10.1007/s00780-009-0095-y Quasi-Monte Carlo methods with applications in finance Pierre L Ecuyer Received: 30 May 2008 / Accepted: 20 August 2008 / Published online:
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More informationWeierstraß-Institut. für Angewandte Analysis und Stochastik. Leibniz-Institut im Forschungsverbund Berlin e. V. Preprint ISSN
Weierstraß-Institut für Angewandte Analysis und Stochastik Leibniz-Institut im Forschungsverbund Berlin e. V. Preprint ISSN 2198-5855 Are Quasi-Monte Carlo algorithms efficient for two-stage stochastic
More informationTHE WAVE EQUATION. d = 1: D Alembert s formula We begin with the initial value problem in 1 space dimension { u = utt u xx = 0, in (0, ) R, (2)
THE WAVE EQUATION () The free wave equation takes the form u := ( t x )u = 0, u : R t R d x R In the literature, the operator := t x is called the D Alembertian on R +d. Later we shall also consider the
More information1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =
Chapter 5 Sequences and series 5. Sequences Definition 5. (Sequence). A sequence is a function which is defined on the set N of natural numbers. Since such a function is uniquely determined by its values
More informationPoisson Solvers. William McLean. April 21, Return to Math3301/Math5315 Common Material.
Poisson Solvers William McLean April 21, 2004 Return to Math3301/Math5315 Common Material 1 Introduction Many problems in applied mathematics lead to a partial differential equation of the form a 2 u +
More informationIntroduction and Review of Power Series
Introduction and Review of Power Series Definition: A power series in powers of x a is an infinite series of the form c n (x a) n = c 0 + c 1 (x a) + c 2 (x a) 2 +...+c n (x a) n +... If a = 0, this is
More information10.1 Sequences. Example: A sequence is a function f(n) whose domain is a subset of the integers. Notation: *Note: n = 0 vs. n = 1.
10.1 Sequences Example: A sequence is a function f(n) whose domain is a subset of the integers. Notation: *Note: n = 0 vs. n = 1 Examples: EX1: Find a formula for the general term a n of the sequence,
More informationSCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES. Kinjal Basu Art B. Owen
SCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES By Kinjal Basu Art B. Owen Technical Report No. 2015-07 March 2015 Department of Statistics STANFORD UNIVERSITY Stanford, California 94305-4065
More informationLECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel
LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count
More informationIntroduction CSE 541
Introduction CSE 541 1 Numerical methods Solving scientific/engineering problems using computers. Root finding, Chapter 3 Polynomial Interpolation, Chapter 4 Differentiation, Chapter 4 Integration, Chapters
More informationImportance Sampling to Accelerate the Convergence of Quasi-Monte Carlo
Importance Sampling to Accelerate the Convergence of Quasi-Monte Carlo Wolfgang Hörmann, Josef Leydold Department of Statistics and Mathematics Wirtschaftsuniversität Wien Research Report Series Report
More informationFinite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product
Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )
More informationVariance Reduction s Greatest Hits
1 Variance Reduction s Greatest Hits Pierre L Ecuyer CIRRELT, GERAD, and Département d Informatique et de Recherche Opérationnelle Université de Montréal, Canada ESM 2007, Malta, October 22, 2007 Context
More informationRandomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space
Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space Pierre L Ecuyer 1, Christian Lécot 2, and Bruno Tuffin 3 1 Département d informatique et de recherche opérationnelle,
More informationNumerical Methods. King Saud University
Numerical Methods King Saud University Aims In this lecture, we will... find the approximate solutions of derivative (first- and second-order) and antiderivative (definite integral only). Numerical Differentiation
More informationKernel Method: Data Analysis with Positive Definite Kernels
Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University
More informationStability of optimization problems with stochastic dominance constraints
Stability of optimization problems with stochastic dominance constraints D. Dentcheva and W. Römisch Stevens Institute of Technology, Hoboken Humboldt-University Berlin www.math.hu-berlin.de/~romisch SIAM
More informationJim Lambers MAT 610 Summer Session Lecture 2 Notes
Jim Lambers MAT 610 Summer Session 2009-10 Lecture 2 Notes These notes correspond to Sections 2.2-2.4 in the text. Vector Norms Given vectors x and y of length one, which are simply scalars x and y, the
More informationFILTERING IN THE FREQUENCY DOMAIN
1 FILTERING IN THE FREQUENCY DOMAIN Lecture 4 Spatial Vs Frequency domain 2 Spatial Domain (I) Normal image space Changes in pixel positions correspond to changes in the scene Distances in I correspond
More informationMalliavin Calculus in Finance
Malliavin Calculus in Finance Peter K. Friz 1 Greeks and the logarithmic derivative trick Model an underlying assent by a Markov process with values in R m with dynamics described by the SDE dx t = b(x
More informationNeural Networks Learning the network: Backprop , Fall 2018 Lecture 4
Neural Networks Learning the network: Backprop 11-785, Fall 2018 Lecture 4 1 Recap: The MLP can represent any function The MLP can be constructed to represent anything But how do we construct it? 2 Recap:
More informationPartial Differential Equations
Part II Partial Differential Equations Year 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2015 Paper 4, Section II 29E Partial Differential Equations 72 (a) Show that the Cauchy problem for u(x,
More informationApplication of Taylor Models to the Worst-Case Analysis of Stripline Interconnects
Application of Taylor Models to the Worst-Case Analysis of Stripline Interconnects Paolo Manfredi, Riccardo Trinchero, Igor Stievano, Flavio Canavero Email: paolo.manfredi@ugent.be riccardo.trinchero@to.infn.it,
More informationA Belgian view on lattice rules
A Belgian view on lattice rules Ronald Cools Dept. of Computer Science, KU Leuven Linz, Austria, October 14 18, 2013 Atomium Brussels built in 1958 height 103m figure = 2e coin 5 10 6 in circulation Body
More information2 Two-Point Boundary Value Problems
2 Two-Point Boundary Value Problems Another fundamental equation, in addition to the heat eq. and the wave eq., is Poisson s equation: n j=1 2 u x 2 j The unknown is the function u = u(x 1, x 2,..., x
More informationError Analysis for Quasi-Monte Carlo Methods
Error Analysis for Quasi-Monte Carlo Methods Fred J. Hickernell Department of Applied Mathematics, Illinois Institute of Technology hickernell@iit.edu mypages.iit.edu/~hickernell Thanks to the Guaranteed
More informationSTRONG TRACTABILITY OF MULTIVARIATE INTEGRATION USING QUASI MONTE CARLO ALGORITHMS
MATHEMATICS OF COMPUTATION Volume 72, Number 242, Pages 823 838 S 0025-5718(02)01440-0 Article electronically published on June 25, 2002 STRONG TRACTABILITY OF MULTIVARIATE INTEGRATION USING QUASI MONTE
More informationQuasi-Monte Carlo Algorithms with Applications in Numerical Analysis and Finance
Quasi-Monte Carlo Algorithms with Applications in Numerical Analysis and Finance Reinhold F. Kainhofer Rigorosumsvortrag, Institut für Mathematik Graz, 16. Mai 2003 Inhalt 1. Quasi-Monte Carlo Methoden
More informationIan Sloan and Lattice Rules
www.oeaw.ac.at Ian Sloan and Lattice Rules P. Kritzer, H. Niederreiter, F. Pillichshammer RICAM-Report 2016-34 www.ricam.oeaw.ac.at Ian Sloan and Lattice Rules Peter Kritzer, Harald Niederreiter, Friedrich
More informationBefore you begin read these instructions carefully.
MATHEMATICAL TRIPOS Part IA Tuesday, 4 June, 2013 1:30 pm to 4:30 pm PAPER 3 Before you begin read these instructions carefully. The examination paper is divided into two sections. Each question in Section
More informationConvergence of Particle Filtering Method for Nonlinear Estimation of Vortex Dynamics
Convergence of Particle Filtering Method for Nonlinear Estimation of Vortex Dynamics Meng Xu Department of Mathematics University of Wyoming February 20, 2010 Outline 1 Nonlinear Filtering Stochastic Vortex
More informationABSTRACT NONSINGULAR CURVES
ABSTRACT NONSINGULAR CURVES Affine Varieties Notation. Let k be a field, such as the rational numbers Q or the complex numbers C. We call affine n-space the collection A n k of points P = a 1, a,..., a
More informationMulti-Factor Finite Differences
February 17, 2017 Aims and outline Finite differences for more than one direction The θ-method, explicit, implicit, Crank-Nicolson Iterative solution of discretised equations Alternating directions implicit
More informationCONVERGENCE THEORY. G. ALLAIRE CMAP, Ecole Polytechnique. 1. Maximum principle. 2. Oscillating test function. 3. Two-scale convergence
1 CONVERGENCE THEOR G. ALLAIRE CMAP, Ecole Polytechnique 1. Maximum principle 2. Oscillating test function 3. Two-scale convergence 4. Application to homogenization 5. General theory H-convergence) 6.
More informationDynamic Risk Measures and Nonlinear Expectations with Markov Chain noise
Dynamic Risk Measures and Nonlinear Expectations with Markov Chain noise Robert J. Elliott 1 Samuel N. Cohen 2 1 Department of Commerce, University of South Australia 2 Mathematical Insitute, University
More informationIntroduction to (randomized) quasi-monte Carlo
1 aft Introduction to (randomized) quasi-monte Carlo Dr Pierre L Ecuyer MCQMC Conference, Stanford University, August 2016 Program Monte Carlo, Quasi-Monte Carlo, Randomized quasi-monte Carlo QMC point
More informationStochastic Spectral Approaches to Bayesian Inference
Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to
More informationAlgorithms for Uncertainty Quantification
Algorithms for Uncertainty Quantification Lecture 9: Sensitivity Analysis ST 2018 Tobias Neckel Scientific Computing in Computer Science TUM Repetition of Previous Lecture Sparse grids in Uncertainty Quantification
More informationTrivariate polynomial approximation on Lissajous curves 1
Trivariate polynomial approximation on Lissajous curves 1 Stefano De Marchi DMV2015, Minisymposium on Mathematical Methods for MPI 22 September 2015 1 Joint work with Len Bos (Verona) and Marco Vianello
More informationState Space Representation of Gaussian Processes
State Space Representation of Gaussian Processes Simo Särkkä Department of Biomedical Engineering and Computational Science (BECS) Aalto University, Espoo, Finland June 12th, 2013 Simo Särkkä (Aalto University)
More information