Introduction to (randomized) quasi-monte Carlo
|
|
- Clementine Banks
- 6 years ago
- Views:
Transcription
1 1 aft Introduction to (randomized) quasi-monte Carlo Dr Pierre L Ecuyer MCQMC Conference, Stanford University, August 2016
2 Program Monte Carlo, Quasi-Monte Carlo, Randomized quasi-monte Carlo QMC point sets and randomizations Error and variance bounds, convergence rates Transforming the integrand to make it more QMC friendly (smoother, smaller effective dimension, etc.). Numerical illustrations RQMC for Markov chains Focus on ideas, insight, and examples. 2
3 Example: A stochastic activity network Gives precedence relations between activities. Activity k has random duration Y k (also length of arc k) with known cumulative distribution function (cdf) F k (y) := P[Y k y]. Project duration T = (random) length of longest path from source to sink. May want to estimate E[T ], P[T > x], a quantile, density of T, etc. Y 1 Y 2 source 0 1 Y 0 2 Y 5 Y 3 Y Y 9 Y 7 Y 10 Y 8 Y 6 8 sink 7 6 Y 12 Y 11 3
4 Monte Carlo (simulation) Algorithm: Monte Carlo to estimate E[T ] for i = 0,..., n 1 do for k = 0,..., 12 do Generate U k U(0, 1) and let Y k = F 1 k (U k ) Compute X i = T = h(y 0,..., Y 12 ) = f (U 0,..., U 12 ) Estimate E[T ] = (0,1) f (u)du by X s n = 1 n 1 n i=0 X i, etc. Can also compute confidence interval on E[T ], a histogram to estimate the distribution of T, etc. Numerical illustration from Elmaghraby (1977): Y k N(µ k, σ 2 k ) for k = 0, 1, 3, 10, 11, and V k Expon(1/µ k ) otherwise. µ 0,..., µ 12 : 13.0, 5.5, 7.0, 5.2, 16.5, 14.7, 10.3, 6.0, 4.0, 20.0, 3.2, 3.2, We may pay a penalty if T > 90, for example. 4
5 Naive idea: replace each Y k by its expectation. Gives T =
6 Naive idea: replace each Y k by its expectation. Gives T = Results of an experiment with n = Histogram of values of T gives more information than confidence interval on E[T ] or P[T > x]. Values from 14.4 to 268.6; 11.57% exceed x = 90. Frequency mean = 64.2 T = 48.2 T = x = 90 ˆξ 0.99 = T 5
7 6 Sample path of hurricane Sandy for the next 5 days As Forecasts Go, You Can Bet on Monte Carlo - WSJ This copy is for your personal, non-commercial use only. To order presentation-ready copies for distribution to your colleagues, clients or customers visit U.S. THE NUMBERS As Forecasts Go, You Can Bet on Monte Carlo From Super Bowls to hurricanes, this simulation method helps predict them all Monte Carlo simulations helped give emergency workers advance warning that Hurricane Sandy would make landfall in
8 7 Sample path of hurricane Sandy for the next 5 days
9 Monte Carlo to estimate an expectation Want to estimate µ = E[X ] where X = f (U) = f (U 0,..., U s 1 ), and the U j are i.i.d. U(0, 1) random numbers. We have µ = E[X ] = f (u)du. [0,1) s Monte Carlo estimator: X n = 1 n 1 n where X i = f (U i ) and U 0,..., U n 1 i.i.d. uniform over [0, 1) s. i=0 We have E[ X n ] = µ and Var[ X n ] = σ 2 /n = Var[X ]/n. X i 8
10 Convergence Theorem. Suppose σ 2 <. When n : (i) Strong law of large numbers: lim n ˆµ n = µ with probability 1. 9
11 Convergence Theorem. Suppose σ 2 <. When n : (i) Strong law of large numbers: lim n ˆµ n = µ with probability 1. (ii) Central limit theorem (CLT): n(ˆµn µ) S n N(0, 1), where Sn 2 = 1 n 1 (X i X n ) 2. n 1 i=0 9
12 Confidence interval at level α (we want Φ(x) = 1 α/2): (ˆµ n ± z α/2 S n / n), where z α/2 = Φ 1 (1 α/2). Example: z α/ for α = z α/2 z α/2 α/2 α/2 1 α The width of the confidence interval is asymptotically proportional to σ/ n, so it converges as O(n 1/2 ). Relative error: σ/(µ n). For one more decimal digit of accuracy, we must multiply n by
13 Confidence interval at level α (we want Φ(x) = 1 α/2): (ˆµ n ± z α/2 S n / n), where z α/2 = Φ 1 (1 α/2). Example: z α/ for α = z α/2 z α/2 α/2 α/2 1 α The width of the confidence interval is asymptotically proportional to σ/ n, so it converges as O(n 1/2 ). Relative error: σ/(µ n). For one more decimal digit of accuracy, we must multiply n by 100. Warning: If the X i have an asymmetric law, these confidence intervals can have very bad coverage (convergence to normal can be very slow). 10
14 Alternative estimator of P[T > x] = E[I(T > x)] for SAN. Naive estimator: Generate T and compute X = I[T > x]. Repeat n times and average. Y 1 Y 2 source 0 1 Y 0 2 Y 5 Y 3 Y Y 9 Y 7 Y 10 Y 8 Y 6 8 sink 7 6 Y 12 Y 11 11
15 Conditional Monte Carlo estimator of P[T > x]. Generate the Y j s only for the 8 arcs that do not belong to the cut L = {4, 5, 6, 8, 9}, and replace I[T > x] by its conditional expectation given those Y j s, X e = P[T > x {Y j, j L}]. This makes the integrand continuous in the U j s. 12
16 Conditional Monte Carlo estimator of P[T > x]. Generate the Y j s only for the 8 arcs that do not belong to the cut L = {4, 5, 6, 8, 9}, and replace I[T > x] by its conditional expectation given those Y j s, X e = P[T > x {Y j, j L}]. This makes the integrand continuous in the U j s. To compute X e : for each l L, say from a l to b l, compute the length α l of the longest path from 1 to a l, and the length β l of the longest path from b l to the destination. The longest path that passes through link l does not exceed x iff α l + Y l + β l x, which occurs with probability P[Y l x α l β l ] = F l [x α l β l ]. 12
17 Conditional Monte Carlo estimator of P[T > x]. Generate the Y j s only for the 8 arcs that do not belong to the cut L = {4, 5, 6, 8, 9}, and replace I[T > x] by its conditional expectation given those Y j s, X e = P[T > x {Y j, j L}]. This makes the integrand continuous in the U j s. To compute X e : for each l L, say from a l to b l, compute the length α l of the longest path from 1 to a l, and the length β l of the longest path from b l to the destination. The longest path that passes through link l does not exceed x iff α l + Y l + β l x, which occurs with probability P[Y l x α l β l ] = F l [x α l β l ]. Since the Y l are independent, we obtain X e = 1 l L F l [x α l β l ]. Can be faster to compute than X, and always has less variance. 12
18 Example: Pricing a financial derivative. Market price of some asset (e.g., one share of a stock) evolves in time as stochastic process {S(t), t 0} with (supposedly) known probability law (estimated from data). A financial contract gives owner net payoff g(s(t 1 ),..., S(t d )) at time T = t d, where g : R d R, and 0 t 1 < < t d are fixed observation times. Under a no-arbitrage assumption, present value (fair price) of contract at time 0, when S(0) = s 0, can be written as v(s 0, T ) = E [ e rt g(s(t 1 ),..., S(t d )) ], where E is under a risk-neutral measure and e rt is the discount factor. This expectation can be written as an integral over [0, 1) s and estimated by the average of n i.i.d. replicates of X = e rt g(s(t 1 ),..., S(t d )). 13
19 A simple model for S: geometric Brownian motion (GBM): S(t) = s 0 e (r σ2 /2)t+σB(t) where r is the interest rate, σ is the volatility, and B( ) is a standard Brownian motion: for any t 2 > t 1 0, B(t 2 ) B(t 1 ) N(0, t 2 t 1 ), and the increments over disjoint intervals are independent. 14
20 A simple model for S: geometric Brownian motion (GBM): S(t) = s 0 e (r σ2 /2)t+σB(t) where r is the interest rate, σ is the volatility, and B( ) is a standard Brownian motion: for any t 2 > t 1 0, B(t 2 ) B(t 1 ) N(0, t 2 t 1 ), and the increments over disjoint intervals are independent. Algorithm: Option pricing under GBM model for i = 0,..., n 1 do Let t 0 = 0 and B(t 0 ) = 0 for j = 1,..., d do Generate U j U(0, 1) and let Z j = Φ 1 (U j ) Let B(t j ) = B(t j 1 ) + t j t j 1 Z j Let S(t j ) = s 0 exp [ (r σ 2 /2)t j + σb(t j ) ] Compute X i = e rt g(s(t 1 ),..., S(t d )) Return X n = 1 n 1 n i=0 X i, estimator of v(s 0, T ). 14
21 Example of contract: Discretely-monitored Asian call option: g(s(t 1 ),..., S(t d )) = max 0, 1 d S(t j ) K. d Option price written as an integral over the unit hypercube: Let Z j = Φ 1 (U j ) where the U j are i.i.d. U(0, 1). Here we have s = d and v(s 0, T ) = ( e rt max 0, 1 s s 0 [0,1) s s i=1 i exp (r σ 2 /2)t i + σ tj t j 1 Φ 1 (u j ) K du 1... du s = j=1 [0,1) s f (u 1,..., u s )du 1... du s. j=1 15
22 16 Numerical illustration: Bermudean Asian option with d = 12, T = 1 (one year), t j = j/12 for j = 0,..., 12, K = 100, s 0 = 100, r = 0.05, σ = 0.5. We performed n = 10 6 independent simulation runs. In 53.47% of cases, the payoff is 0. Mean: Max = Histogram of the 46.53% positive values: Frequency ( 10 3 ) Payoff
23 Reducing the variance by changing f If we replace the arithmetic average by a geometric average in the payoff, we obtain d C = e rt max 0, (S(t j )) 1/d K, j=1 whose expectation ν = E[C] has a closed-form formula. When estimating the mean E[X ] = v(s 0, T ), we can then use C as a control variate (CV): Replace the estimator X by the corrected version X c = X β(c ν) for some well-chosen constant β. Optimal β is β = Cov[C, X ]/Var[C]. Using a CV makes the integrand f smoother. Can provide a huge variance reduction, e.g., by a factor of over a million in some examples. 17
24 Quasi-Monte Carlo (QMC) Replace the independent random points U i by a set of deterministic points P n = {u 0,..., u n 1 } that cover [0, 1) s more evenly. Estimate Integration error E n = µ µ. µ = f (u)du by µ n = 1 n 1 f (u i ). [0,1) s n P n is called a highly-uniform point set or low-discrepancy point set if some measure of discrepancy between the empirical distribution of P n and the uniform distribution converges to 0 faster than O(n 1/2 ) (the typical rate for independent random points). i=0 18
25 Quasi-Monte Carlo (QMC) Replace the independent random points U i by a set of deterministic points P n = {u 0,..., u n 1 } that cover [0, 1) s more evenly. Estimate Integration error E n = µ µ. µ = f (u)du by µ n = 1 n 1 f (u i ). [0,1) s n P n is called a highly-uniform point set or low-discrepancy point set if some measure of discrepancy between the empirical distribution of P n and the uniform distribution converges to 0 faster than O(n 1/2 ) (the typical rate for independent random points). i=0 Main construction methods: lattice rules and digital nets (Korobov, Hammersley, Halton, Sobol, Faure, Niederreiter, etc.) 18
26 Simple case: one dimension (s = 1) Obvious solutions: P n = Z n /n = {0, 1/n,..., (n 1)/n} (left Riemann sum): which gives µ n = 1 n 1 f (i/n), and E n = O(n 1 ) if f is bounded, n i=0 19
27 Simple case: one dimension (s = 1) Obvious solutions: P n = Z n /n = {0, 1/n,..., (n 1)/n} (left Riemann sum): which gives µ n = 1 n 1 f (i/n), and E n = O(n 1 ) if f is bounded, n i=0 or P n = {1/(2n), 3/(2n),..., (2n 1)/(2n)} (midpoint rule): for which E n = O(n 2 ) if f is bounded. 19
28 If we allow different weights on the f (u i ), we have the trapezoidal rule: [ ] 1 f (0) + f (1) n 1 + f (i/n), n 2 for which E n = O(n 2 ) if f is bounded, i=1 20
29 If we allow different weights on the f (u i ), we have the trapezoidal rule: [ ] 1 f (0) + f (1) n 1 + f (i/n), n 2 for which E n = O(n 2 ) if f is bounded, or the Simpson rule, f (0) + 4f (1/n) + 2f (2/n) + + 2f ((n 2)/n) + 4f ((n 1)/n) + f (1), 3n i=1 which gives E n = O(n 4 ) if f (4) is bounded, etc. 20
30 If we allow different weights on the f (u i ), we have the trapezoidal rule: [ ] 1 f (0) + f (1) n 1 + f (i/n), n 2 for which E n = O(n 2 ) if f is bounded, or the Simpson rule, f (0) + 4f (1/n) + 2f (2/n) + + 2f ((n 2)/n) + 4f ((n 1)/n) + f (1), 3n i=1 which gives E n = O(n 4 ) if f (4) is bounded, etc. Here, for QMC and RQMC, we restrict ourselves to equal weight rules. For the RQMC points that we will examine, one can prove that equal weights are optimal. 20
31 Simplistic solution for s > 1: rectangular grid P n = {(i 1 /d,..., i s /d) such that 0 i j < d j} where n = d s. u i, u i,2 21
32 Simplistic solution for s > 1: rectangular grid P n = {(i 1 /d,..., i s /d) such that 0 i j < d j} where n = d s. u i, Midpoint rule in s dimensions. Quickly becomes impractical when s increases. Moreover, each one-dimensional projection has only d distinct points, each two-dimensional projections has only d 2 distinct points, etc. u i,2 21
33 Lattice rules (Korobov, Sloan, etc.) Integration lattice: L s = v = s z j v j such that each z j Z, j=1 where v 1,..., v s R s are linearly independent over R and where L s contains Z s. Lattice rule: Take P n = {u 0,..., u n 1 } = L s [0, 1) s. 22
34 Lattice rules (Korobov, Sloan, etc.) Integration lattice: L s = v = s z j v j such that each z j Z, j=1 where v 1,..., v s R s are linearly independent over R and where L s contains Z s. Lattice rule: Take P n = {u 0,..., u n 1 } = L s [0, 1) s. Lattice rule of rank 1: u i = iv 1 mod 1 for i = 0,..., n 1, where nv 1 = a = (a 1,..., a s ) {0, 1,..., n 1} s. Korobov rule: a = (1, a, a 2 mod n,... ). 22
35 Lattice rules (Korobov, Sloan, etc.) Integration lattice: L s = v = s z j v j such that each z j Z, j=1 where v 1,..., v s R s are linearly independent over R and where L s contains Z s. Lattice rule: Take P n = {u 0,..., u n 1 } = L s [0, 1) s. Lattice rule of rank 1: u i = iv 1 mod 1 for i = 0,..., n 1, where nv 1 = a = (a 1,..., a s ) {0, 1,..., n 1} s. Korobov rule: a = (1, a, a 2 mod n,... ). For any u {1,..., s}, the projection L s (u) of L s is also a lattice. 22
36 Example: lattice with s = 2, n = 101, v 1 = (1, 12)/n P n = {u i = iv 1 mod 1) : i = 0,..., n 1} = {(0, 0), (1/101, 12/101), (2/101, 43/101),... }. 1 u i,1 v Here, each one-dimensional projection is {0, 1/n,..., (n 1)/n}. u i,2 23
37 Example: lattice with s = 2, n = 101, v 1 = (1, 12)/n P n = {u i = iv 1 mod 1) : i = 0,..., n 1} = {(0, 0), (1/101, 12/101), (2/101, 43/101),... }. 1 u i,1 v Here, each one-dimensional projection is {0, 1/n,..., (n 1)/n}. u i,2 23
38 Example: lattice with s = 2, n = 101, v 1 = (1, 12)/n P n = {u i = iv 1 mod 1) : i = 0,..., n 1} = {(0, 0), (1/101, 12/101), (2/101, 43/101),... }. 1 u i,1 v Here, each one-dimensional projection is {0, 1/n,..., (n 1)/n}. u i,2 23
39 Example: lattice with s = 2, n = 101, v 1 = (1, 12)/n P n = {u i = iv 1 mod 1) : i = 0,..., n 1} = {(0, 0), (1/101, 12/101), (2/101, 43/101),... }. 1 u i,1 v Here, each one-dimensional projection is {0, 1/n,..., (n 1)/n}. u i,2 23
40 Example: lattice with s = 2, n = 101, v 1 = (1, 12)/n P n = {u i = iv 1 mod 1) : i = 0,..., n 1} = {(0, 0), (1/101, 12/101), (2/101, 43/101),... }. 1 u i,1 v Here, each one-dimensional projection is {0, 1/n,..., (n 1)/n}. u i,2 23
41 Another example: s = 2, n = 1021, v 1 = (1, 90)/n P n = {u i = iv 1 mod 1 : i = 0,..., n 1} = {(i/1021, (90i/1021) mod 1) : i = 0,..., 1020}. u i,1 1 v u i,2 24
42 A bad lattice: s = 2, n = 101, v 1 = (1, 51)/n u i,1 1 v u i,2 Good uniformity in one dimension, but not in two! 25
43 Digital net in base b (Niederreiter) Gives n = b k points. For i = 0,..., b k 1 and j = 1,..., s: u i,j,1. u i,j,w i = a i,0 + a i,1 b + + a i,k 1 b k 1 = a i,k 1 a i,1 a i,0, = C j u i,j = a i,0. a i,k 1 mod b, w u i,j,l b l, u i = (u i,1,..., u i,s ), l=1 where the generating matrices C j are w k with elements in Z b. In practice, w and k are finite, but there is no limit. Digital sequence: infinite sequence. Can stop at n = b k for any k. 26
44 Digital net in base b (Niederreiter) Gives n = b k points. For i = 0,..., b k 1 and j = 1,..., s: u i,j,1. u i,j,w i = a i,0 + a i,1 b + + a i,k 1 b k 1 = a i,k 1 a i,1 a i,0, = C j u i,j = a i,0. a i,k 1 mod b, w u i,j,l b l, u i = (u i,1,..., u i,s ), l=1 where the generating matrices C j are w k with elements in Z b. In practice, w and k are finite, but there is no limit. Digital sequence: infinite sequence. Can stop at n = b k for any k. Can also multiply in some ring R, with bijections between Z b and R. Each one-dim projection truncated to first k digits is Z n /n = {0, 1/n,..., (n 1)/n}. Each C j defines a permutation of Z n /n. 26
45 Small example: Hammersley in two dimensions Let n = 2 8 = 256 and s = 2. Take the points (in binary): i u 1,i u 2,i Right side: van der Corput sequence in base 2. 27
46 u i,1 Hammersley point set, n = 2 8 = 256, s = u i,2 28
47 u i,1 Hammersley point set, n = 2 8 = 256, s = u i,2 28
48 u i,1 Hammersley point set, n = 2 8 = 256, s = u i,2 28
49 u i,1 Hammersley point set, n = 2 8 = 256, s = u i,2 28
50 u i,1 Hammersley point set, n = 2 8 = 256, s = u i,2 28
51 In general, can take n = 2 k points. If we partition [0, 1) 2 in rectangles of sizes 2 k 1 by 2 k 2 where k 1 + k 2 k, each rectangle will contain exactly the same number of points. We say that the points are equidistributed for this partition. 29
52 In general, can take n = 2 k points. If we partition [0, 1) 2 in rectangles of sizes 2 k 1 by 2 k 2 where k 1 + k 2 k, each rectangle will contain exactly the same number of points. We say that the points are equidistributed for this partition. For a digital net in base b in s dimensions, we choose s permutations of {0, 1,..., 2 b 1}, then divide each coordinate by b k. Can also have s = and/or n = (infinite sequence of points). 29
53 Suppose we divide axis j in b q j equal parts, for each j. This determines a partition of [0, 1) s into 2 q 1+ +q s rectangles of equal sizes. If each rectangle contains exactly the same number of points, we say that the point set P n is (q 1,..., q s )-equidistributed in base b. This occurs iff the matrix formed by the first q 1 rows of C 1, the first q 2 rows of C 2,..., the first q s rows of C s, is of full rank (mod b). To verify equidistribution, we can construct these matrices and compute their rank. P n is a (t, k, s)-net iff it is (q 1,..., q s )-equidistributed whenever q q s = k t. This is possible for t = 0 only if b s 1. t-value of a net: smallest t for which it is a (t, k, s)-net. 30
54 Suppose we divide axis j in b q j equal parts, for each j. This determines a partition of [0, 1) s into 2 q 1+ +q s rectangles of equal sizes. If each rectangle contains exactly the same number of points, we say that the point set P n is (q 1,..., q s )-equidistributed in base b. This occurs iff the matrix formed by the first q 1 rows of C 1, the first q 2 rows of C 2,..., the first q s rows of C s, is of full rank (mod b). To verify equidistribution, we can construct these matrices and compute their rank. P n is a (t, k, s)-net iff it is (q 1,..., q s )-equidistributed whenever q q s = k t. This is possible for t = 0 only if b s 1. t-value of a net: smallest t for which it is a (t, k, s)-net. An infinite sequence {u 0, u 1,..., } in [0, 1) s is a (t, s)-sequence in base b if for all k > 0 and ν 0, Q(k, ν) = {u i : i = νb k,..., (ν + 1)b k 1}, is a (t, k, s)-net in base b. 30
55 Suppose we divide axis j in b q j equal parts, for each j. This determines a partition of [0, 1) s into 2 q 1+ +q s rectangles of equal sizes. If each rectangle contains exactly the same number of points, we say that the point set P n is (q 1,..., q s )-equidistributed in base b. This occurs iff the matrix formed by the first q 1 rows of C 1, the first q 2 rows of C 2,..., the first q s rows of C s, is of full rank (mod b). To verify equidistribution, we can construct these matrices and compute their rank. P n is a (t, k, s)-net iff it is (q 1,..., q s )-equidistributed whenever q q s = k t. This is possible for t = 0 only if b s 1. t-value of a net: smallest t for which it is a (t, k, s)-net. An infinite sequence {u 0, u 1,..., } in [0, 1) s is a (t, s)-sequence in base b if for all k > 0 and ν 0, Q(k, ν) = {u i : i = νb k,..., (ν + 1)b k 1}, is a (t, k, s)-net in base b. This is possible for t = 0 only if b s. 30
56 Sobol nets and sequences Sobol (1967) proposed a digital net in base b = 2 where 1 v j,2,1... v j,c, v j,c,2... C j =
57 Sobol nets and sequences Sobol (1967) proposed a digital net in base b = 2 where 1 v j,2,1... v j,c, v j,c,2... C j = Column c of C j is represented by an odd integer c m j,c = v j,c,l 2 c l = v j,c,1 2 c v j,c,c < 2 c. l=1 The integers m j,c are selected as follows. 31
58 For each j, we choose a primitive polynomial over F 2, f j (z) = z d j + a j,1 z d j a j,dj, and we choose d j integers m j,0,..., m j,dj 1 (the first d j columns). 32
59 For each j, we choose a primitive polynomial over F 2, f j (z) = z d j + a j,1 z d j a j,dj, and we choose d j integers m j,0,..., m j,dj 1 (the first d j columns). Then, m j,dj, m j,dj +1,... are determined by the recurrence m j,c = 2a j,1 m j,c 1 2 d j 1 a j,dj 1m j,c dj +1 2 d j m j,c dj m j,c dj Proposition. If the polynomials f j (z) are all distinct, we obtain a (t, s)-sequence with t d d s s. 32
60 For each j, we choose a primitive polynomial over F 2, f j (z) = z d j + a j,1 z d j a j,dj, and we choose d j integers m j,0,..., m j,dj 1 (the first d j columns). Then, m j,dj, m j,dj +1,... are determined by the recurrence m j,c = 2a j,1 m j,c 1 2 d j 1 a j,dj 1m j,c dj +1 2 d j m j,c dj m j,c dj Proposition. If the polynomials f j (z) are all distinct, we obtain a (t, s)-sequence with t d d s s. Sobol suggests to list all primitive polynomials over F 2 by increasing order of degree, starting with f 0 (z) 1 (which gives C 0 = I), and to take f j (z) as the (j + 1)-th polynomial in the list. There are many ways of selecting the first m j,c s, which are called the direction numbers. They can be selected to minimize some discrepancy (or figure of merit). The values proposed by Sobol give an (s, l)-equidistribution for l = 1 and l = 2 (only the first two bits). For n = 2 k fixed, we can gain one dimension as for the Faure sequence. Joe and Kuo (2008) tabulated direction numbers giving the best t-value for the two-dimensional projections, for given s and k. 32
61 Other constructions Faure nets and sequences Niederreiter-Xing point sets and sequences Polynomial lattice rules (special case of digital nets) Halton sequence Etc. 33
62 Worst-case error bounds Koksma-Hlawka-type inequalities (Koksma, Hlawka, Hickernell, etc.): ˆµ n,rqmc µ V (f ) D(P n ) for all f in some Hilbert space or Banach space H, where V (f ) = f µ H is the variation of f, and D(P n ) is the discrepancy of P n. 34
63 Worst-case error bounds Koksma-Hlawka-type inequalities (Koksma, Hlawka, Hickernell, etc.): ˆµ n,rqmc µ V (f ) D(P n ) for all f in some Hilbert space or Banach space H, where V (f ) = f µ H is the variation of f, and D(P n ) is the discrepancy of P n. Lattice rules: For certain Hilbert spaces of smooth periodic functions f with square-integrable partial derivatives of order up to α: D(P n ) = O(n α+ɛ ) for arbitrary small ɛ. Digital nets: Classical Koksma-Hlawka inequality for QMC: f must have finite variation in the sense of Hardy and Krause (implies no discontinuity not aligned with the axes). Popular constructions achieve D(P n ) = O(n 1 (ln n) s ) = O(n 1+ɛ ) for arbitrary small ɛ. More recent constructions offer better rates for smooth functions. 34
64 Worst-case error bounds Koksma-Hlawka-type inequalities (Koksma, Hlawka, Hickernell, etc.): ˆµ n,rqmc µ V (f ) D(P n ) for all f in some Hilbert space or Banach space H, where V (f ) = f µ H is the variation of f, and D(P n ) is the discrepancy of P n. Lattice rules: For certain Hilbert spaces of smooth periodic functions f with square-integrable partial derivatives of order up to α: D(P n ) = O(n α+ɛ ) for arbitrary small ɛ. Digital nets: Classical Koksma-Hlawka inequality for QMC: f must have finite variation in the sense of Hardy and Krause (implies no discontinuity not aligned with the axes). Popular constructions achieve D(P n ) = O(n 1 (ln n) s ) = O(n 1+ɛ ) for arbitrary small ɛ. More recent constructions offer better rates for smooth functions. Bounds are conservative and too hard to compute in practice. 34
65 Randomized quasi-monte Carlo (RQMC) ˆµ n,rqmc = 1 n 1 f (U i ), n i=0 with P n = {U 0,..., U n 1 } (0, 1) s an RQMC point set: (i) each point U i has the uniform distribution over (0, 1) s ; (ii) P n as a whole is a low-discrepancy point set. E[ˆµ n,rqmc ] = µ Var[ˆµ n,rqmc ] = Var[f (U i)] n (unbiased). + 2 n 2 Cov[f (U i ), f (U j )]. i<j We want to make the last sum as negative as possible. 35
66 Randomized quasi-monte Carlo (RQMC) ˆµ n,rqmc = 1 n 1 f (U i ), n i=0 with P n = {U 0,..., U n 1 } (0, 1) s an RQMC point set: (i) each point U i has the uniform distribution over (0, 1) s ; (ii) P n as a whole is a low-discrepancy point set. E[ˆµ n,rqmc ] = µ Var[ˆµ n,rqmc ] = Var[f (U i)] n (unbiased). + 2 n 2 Cov[f (U i ), f (U j )]. i<j We want to make the last sum as negative as possible. Weaker attempts to do the same: antithetic variates (n = 2), Latin hypercube sampling (LHS), stratification,... 35
67 Variance estimation: Can compute m independent realizations X 1,..., X m of ˆµ n,rqmc, then estimate µ and Var[ˆµ n,rqmc ] by their sample mean X m and sample variance S 2 m. Could be used to compute a confidence interval. Temptation: assume that X m has the normal distribution. Beware: usually wrong unless m. 36
68 Stratification of the unit hypercube Partition axis j in k j 1 equal parts, for j = 1,..., s. Draw n = k 1 k s random points, one per box, independently. Example, s = 2, k 1 = 12, k 2 = 8, n = 12 8 = u i,1 0 1 u i,2 37
69 Stratification of the unit hypercube u i,1 Example, s = 2, k 1 = 24, k 2 = 16, n = u i,2 38
70 Stratified estimator: X s,n = 1 n 1 f (U j ). n The crude MC variance with n points can be decomposed as j=0 Var[ X n ] = Var[X s,n ] + 1 n where µ j is the mean over box j. n 1 (µ j µ) 2 j=0 The more the µ j differ, the more the variance is reduced. 39
71 Stratified estimator: X s,n = 1 n 1 f (U j ). n The crude MC variance with n points can be decomposed as j=0 Var[ X n ] = Var[X s,n ] + 1 n where µ j is the mean over box j. n 1 (µ j µ) 2 j=0 The more the µ j differ, the more the variance is reduced. If f is continuous and bounded, and all k j are equal, then Var[X s,n ] = O(n 1 2/s ). 39
72 Stratified estimator: X s,n = 1 n 1 f (U j ). n The crude MC variance with n points can be decomposed as where µ j is the mean over box j. j=0 Var[ X n ] = Var[X s,n ] + 1 n 1 (µ j µ) 2 n j=0 The more the µ j differ, the more the variance is reduced. If f is continuous and bounded, and all k j are equal, then Var[X s,n ] = O(n 1 2/s ). For large s, not practical. For small s, not really better than midpoint rule with a grid when f is smooth. But can still be applied to a few important random variables. Also, gives an unbiased estimator, and variance can be estimated by replicating m 2 times. 39
73 Randomly-Shifted Lattice Example: lattice with s = 2, n = 101, v 1 = (1, 12)/101 u i, u i,2 40
74 Randomly-Shifted Lattice Example: lattice with s = 2, n = 101, v 1 = (1, 12)/101 u i,1 1 U 0 1 u i,2 40
75 Randomly-Shifted Lattice Example: lattice with s = 2, n = 101, v 1 = (1, 12)/101 u i, u i,2 40
76 Randomly-Shifted Lattice Example: lattice with s = 2, n = 101, v 1 = (1, 12)/101 u i, u i,2 40
77 Random digital shift for digital net Equidistribution in digital boxes is lost with random shift modulo 1, but can be kept with a random digital shift in base b. In base 2: Generate U U(0, 1) s and XOR it bitwise with each u i. Example for s = 2: u i = ( , ) 2 U = ( , ) 2 u i U = ( , ) 2. Each point has U(0, 1) distribution. Preservation of the equidistribution (k 1 = 3, k 2 = 5): u i = (0.***, 0.*****) U = (0.010, ) 2 u i U = (0.***, 0.*****) 41
78 Example with U = ( , ) 10 = ( , ) 2. Changes the bits 3, 9, 15, 16, 17, 18 of u i,1 and the bits 2, 4, 8, 9, 13, 15, 16 of u i,2. u n+1 1 u n+1 0 u n Red and green squares are permuted (k 1 = k 2 = 4, first 4 bits of U). 1 u n 42
79 Random digital shift in base b We have u i,j = w l=1 u i,j,lb l. Let U = (U 1,..., U s ) U[0, 1) s where U j = w l=1 U j,l b l. We replace each u i,j by Ũ i,j = w l=1 [(u i,j,l + U j,l ) mod b]b l. Proposition. P n is (q 1,..., q s )-equidistributed in base b iff P n is. For w =, each point Ũ i has the uniform distribution over (0, 1) s. 43
80 Other permutations that preserve equidistribution and may help reduce the variance further: Linear matrix scrambling (Matoušek, Hickernell et Hong, Tezuka, Owen): We left-multiply each matrix C j by a random w w matrix M j, non-singular and lower triangular, mod b. Several variants. We then apply a random digital shift in base b to obtain uniform distribution for each point (unbiasedness). 44
81 Other permutations that preserve equidistribution and may help reduce the variance further: Linear matrix scrambling (Matoušek, Hickernell et Hong, Tezuka, Owen): We left-multiply each matrix C j by a random w w matrix M j, non-singular and lower triangular, mod b. Several variants. We then apply a random digital shift in base b to obtain uniform distribution for each point (unbiasedness). Nested uniform scrambling (Owen 1995). More costly. But provably reduces the variance to O(n 3 (log n) s ) when f is sufficiently smooth! 44
82 Asian option example T = 1 (year), t j = j/d, K = 100, s 0 = 100, r = 0.05, σ = 0.5. s = d = 2. Exact value: µ MC Variance: Lattice: Korobov with a from old table + random shift. Sobol: left matrix scramble + random digital shift. Variance estimated from m = 1000 indep. randomizations. VRF = (MC variance) / (nvar[x s,n ]) method n X m nsm 2 VRF stratif lattice Sobol stratif lattice Sobol ,330 stratif lattice ,318 Sobol ,000 45
83 s = d = 12. µ MC variance: Lattice: Korobov + random shift. Sobol: left matrix scramble + random digital shift. Variance estimated from m = 1000 indep. randomizations. method n X m nsm 2 VRF lattice Sobol lattice Sobol lattice Sobol
84 Variance for randomly-shifted lattice rules Suppose f has Fourier expansion f (u) = h Z s ˆf (h)e 2π 1h tu. For a randomly shifted lattice, the exact variance is always Var[ˆµ n,rqmc ] = ˆf (h) 2, 0 h L s where L s = {h R s : h t v Z for all v L s } Z s is the dual lattice. From the viewpoint of variance reduction, an optimal lattice for f minimizes Var[ˆµ n,rqmc ]. 47
85 Var[ˆµ n,rqmc ] = 0 h L s ˆf (h) 2. Let α > 0 be an even integer. If f has square-integrable mixed partial derivatives up to order α/2 > 0, and the periodic continuation of its derivatives up to order α/2 1 is continuous across the unit cube boundaries, then ˆf (h) 2 = O((max(1, h 1 ) max(1, h s )) α ). Moreover, there is a vector v 1 = v 1 (n) such that P α := (max(1, h 1 ) max(1, h s )) α = O(n α+ɛ ). 0 h L s This P α has been proposed long ago as a figure of merit, often with α = 2. It is the variance for a worst-case f having ˆf (h) 2 = (max(1, h 1 ) max(1, h s )) α. A larger α means a smoother f and a faster convergence rate. 48
86 For even integer α, this worst-case f is f (u) = u {1,...,s} j u (2π) α/2 (α/2)! B α/2(u j ). where B α/2 is the Bernoulli polynomial of degree α/2. In particular, B 1 (u) = u 1/2 and B 2 (u) = u 2 u + 1/6. Easy to compute P α and search for good lattices in this case! However: This worst-case function is not necessarily representative of what happens in applications. Also, the hidden factor in O increases quickly with s, so this result is not very useful for large s. To get a bound that is uniform in s, the Fourier coefficients must decrease faster with the dimension and size of vectors h; that is, f must be smoother in high-dimensional projections. This is typically what happens in applications for which RQMC is effective! 49
87 Baker s (or tent) transformation To make the periodic continuation of f continuous. If f (0) f (1), define f by f (1 u) = f (u) = f (2u) for 0 u 1/2. This f has the same integral as f and f (0) = f (1) /2. 50
88 Baker s (or tent) transformation To make the periodic continuation of f continuous. If f (0) f (1), define f by f (1 u) = f (u) = f (2u) for 0 u 1/2. This f has the same integral as f and f (0) = f (1) /2. 50
89 Baker s (or tent) transformation To make the periodic continuation of f continuous. If f (0) f (1), define f by f (1 u) = f (u) = f (2u) for 0 u 1/2. This f has the same integral as f and f (0) = f (1) /2. 50
90 Baker s (or tent) transformation To make the periodic continuation of f continuous. If f (0) f (1), define f by f (1 u) = f (u) = f (2u) for 0 u 1/2. This f has the same integral as f and f (0) = f (1) /2 For smooth f, can reduce the variance to O(n 4+ɛ ) (Hickernell 2002). The resulting f is symmetric with respect to u = 1/2. In practice, we transform the points U i instead of f. 50
91 One-dimensional case Random shift followed by baker s transformation. Along each coordinate, stretch everything by a factor of 2 and fold. Same as replacing U j by min[2u j, 2(1 U j )]
92 One-dimensional case Random shift followed by baker s transformation. Along each coordinate, stretch everything by a factor of 2 and fold. Same as replacing U j by min[2u j, 2(1 U j )] U/n 51
93 One-dimensional case Random shift followed by baker s transformation. Along each coordinate, stretch everything by a factor of 2 and fold. Same as replacing U j by min[2u j, 2(1 U j )]
94 One-dimensional case Random shift followed by baker s transformation. Along each coordinate, stretch everything by a factor of 2 and fold. Same as replacing U j by min[2u j, 2(1 U j )] Gives locally antithetic points in intervals of size 2/n. This implies that linear pieces over these intervals are integrated exactly. Intuition: when f is smooth, it is well-approximated by a piecewise linear function, which is integrated exactly, so the error is small. 51
95 ANOVA decomposition The Fourier expansion has too many terms to handle. As a cruder expansion, we can write f (u) = f (u 1,..., u s ) as: f (u) = where u {1,...,s} f u (u) = µ + s f {i} (u i ) + i=1 s f {i,j} (u i, u j ) + i,j=1 f u (u) = f (u) duū f v (u v ), [0,1) ū v u and the Monte Carlo variance decomposes as σ 2 = σu, 2 where σu 2 = Var[f u (U)]. u {1,...,s} The σ 2 u s can be estimated by MC or RQMC. Heuristic intuition: Make sure the projections P n (u) are very uniform for the important subsets u (i.e., with larger σu). 2 52
96 53 Weighted P γ,α with projection-dependent weights γ u Denote u(h) = u(h 1,..., h s ) the set of indices j for which h j 0. P γ,α = 0 h L s γ u(h) (max(1, h 1 ) max(1, h s )) α. For α/2 integer > 0, with u i = (u i,1,..., u i,s ) = iv 1 mod 1, P γ,α = =u {1,...,s} and the corresponding variation is V 2 γ (f ) = =u {1,...,s} 1 n 1 [ ( 4π 2 ) α/2 ] u γ u B α (u i,j ), n (α)! i=0 1 γ u (4π 2 ) α u /2 [0,1] u j u α u /2 u f u(u) α/2 for f : [0, 1) s R smooth enough. Then, Var[ˆµ n,rqmc ] = Var[ˆµ n,rqmc (f u )] Vγ 2 (f )P γ,α. u {1,...,s} 2 du,
97 P γ,α with α = 2 and properly chosen weights γ is a good practical choice of figure of merit. Simple choices of weights: order-dependent or product. Lattice Builder: Software to search for good lattices with arbitrary n, s, weights, etc. See my web page. 54
98 ANOVA Variances for estimator of P[T > x] in Stochastic Activity Network x = 64 x = 100 CMC, x = 64 CMC, x = 100 Stochastic Activity Network % of total variance for each cardinality of u 55
99 Variance for estimator of P[T > x] for SAN variance Stochastic Activity Network (x = 64) MC Sobol Lattice (P 2) + baker n Variance decreases roughly as O(n 1.2 ). For E[T ], we observe O(n 1.4 ). n 56
100 Variance for estimator of P[T > x] with CMC variance Stochastic Activity Network (CMC x = 64) MC Sobol Lattice (P 2) + baker n n 57
101 Histograms probability probability probability single MC draw (x = 100) MC estimator (x = 100) 6 7 RQMC estimator (x = 100)
102 Histograms probability probability probability single MC draw (CMC x = 100) MC estimator (CMC x = 100) RQMC estimator (CMC x = 100)
103 Effective dimension (Caflisch, Morokoff, and Owen 1997). A function f has effective dimension d in proportion ρ in the superposition sense if σu 2 ρσ 2. u d It has effective dimension d in the truncation sense if σu 2 ρσ 2. u {1,...,d} High-dimensional functions with low effective dimension are frequent. One may change f to make this happen. 60
104 Example: Function of a Multinormal vector Let µ = E[f (U)] = E[g(Y)] where Y = (Y 1,..., Y s ) N(0, Σ). 61
105 Example: Function of a Multinormal vector Let µ = E[f (U)] = E[g(Y)] where Y = (Y 1,..., Y s ) N(0, Σ). For example, if the payoff of a financial derivative is a function of the values taken by a c-dimensional geometric Brownian motion (GMB) at d observations times 0 < t 1 < < t d = T, then we have s = cd. 61
106 Example: Function of a Multinormal vector Let µ = E[f (U)] = E[g(Y)] where Y = (Y 1,..., Y s ) N(0, Σ). For example, if the payoff of a financial derivative is a function of the values taken by a c-dimensional geometric Brownian motion (GMB) at d observations times 0 < t 1 < < t d = T, then we have s = cd. To generate Y: Decompose Σ = AA t, generate Z = (Z 1,..., Z s ) N(0, I) where the (independent) Z j s are generated by inversion: Z j = Φ 1 (U j ), and return Y = AZ. 61
107 Example: Function of a Multinormal vector Let µ = E[f (U)] = E[g(Y)] where Y = (Y 1,..., Y s ) N(0, Σ). For example, if the payoff of a financial derivative is a function of the values taken by a c-dimensional geometric Brownian motion (GMB) at d observations times 0 < t 1 < < t d = T, then we have s = cd. To generate Y: Decompose Σ = AA t, generate Z = (Z 1,..., Z s ) N(0, I) where the (independent) Z j s are generated by inversion: Z j = Φ 1 (U j ), and return Y = AZ. Choice of A? 61
108 Example: Function of a Multinormal vector Let µ = E[f (U)] = E[g(Y)] where Y = (Y 1,..., Y s ) N(0, Σ). For example, if the payoff of a financial derivative is a function of the values taken by a c-dimensional geometric Brownian motion (GMB) at d observations times 0 < t 1 < < t d = T, then we have s = cd. To generate Y: Decompose Σ = AA t, generate Z = (Z 1,..., Z s ) N(0, I) where the (independent) Z j s are generated by inversion: Z j = Φ 1 (U j ), and return Y = AZ. Choice of A? Cholesky factorization: A is lower triangular. 61
109 Principal component decomposition (PCA) (Ackworth et al. 1998): A = PD 1/2 where D = diag(λ s,..., λ 1 ) (eigenvalues of Σ in decreasing order) and the columns of P are the corresponding unit-length eigenvectors. 62
110 Principal component decomposition (PCA) (Ackworth et al. 1998): A = PD 1/2 where D = diag(λ s,..., λ 1 ) (eigenvalues of Σ in decreasing order) and the columns of P are the corresponding unit-length eigenvectors. With this A, Z 1 accounts for the max amount of variance of Y, then Z 2 the max amount of variance cond. on Z 1, etc. 62
111 Principal component decomposition (PCA) (Ackworth et al. 1998): A = PD 1/2 where D = diag(λ s,..., λ 1 ) (eigenvalues of Σ in decreasing order) and the columns of P are the corresponding unit-length eigenvectors. With this A, Z 1 accounts for the max amount of variance of Y, then Z 2 the max amount of variance cond. on Z 1, etc. Function of a Brownian motion (or other Lévy process): Payoff depends on c-dimensional Brownian motion {X(t), t 0} observed at times 0 = t 0 < t 1 < < t d = T. 62
112 Principal component decomposition (PCA) (Ackworth et al. 1998): A = PD 1/2 where D = diag(λ s,..., λ 1 ) (eigenvalues of Σ in decreasing order) and the columns of P are the corresponding unit-length eigenvectors. With this A, Z 1 accounts for the max amount of variance of Y, then Z 2 the max amount of variance cond. on Z 1, etc. Function of a Brownian motion (or other Lévy process): Payoff depends on c-dimensional Brownian motion {X(t), t 0} observed at times 0 = t 0 < t 1 < < t d = T. Sequential (or random walk) method: generate X(t 1 ), then X(t 2 ) X(t 1 ), then X(t 3 ) X(t 2 ), etc. 62
113 Principal component decomposition (PCA) (Ackworth et al. 1998): A = PD 1/2 where D = diag(λ s,..., λ 1 ) (eigenvalues of Σ in decreasing order) and the columns of P are the corresponding unit-length eigenvectors. With this A, Z 1 accounts for the max amount of variance of Y, then Z 2 the max amount of variance cond. on Z 1, etc. Function of a Brownian motion (or other Lévy process): Payoff depends on c-dimensional Brownian motion {X(t), t 0} observed at times 0 = t 0 < t 1 < < t d = T. Sequential (or random walk) method: generate X(t 1 ), then X(t 2 ) X(t 1 ), then X(t 3 ) X(t 2 ), etc. Bridge sampling (Moskowitz and Caflisch 1996). Suppose d = 2 m. generate X(t d ), then X(t d/2 ) conditional on (X(0), X(t d )), 62
114 Principal component decomposition (PCA) (Ackworth et al. 1998): A = PD 1/2 where D = diag(λ s,..., λ 1 ) (eigenvalues of Σ in decreasing order) and the columns of P are the corresponding unit-length eigenvectors. With this A, Z 1 accounts for the max amount of variance of Y, then Z 2 the max amount of variance cond. on Z 1, etc. Function of a Brownian motion (or other Lévy process): Payoff depends on c-dimensional Brownian motion {X(t), t 0} observed at times 0 = t 0 < t 1 < < t d = T. Sequential (or random walk) method: generate X(t 1 ), then X(t 2 ) X(t 1 ), then X(t 3 ) X(t 2 ), etc. Bridge sampling (Moskowitz and Caflisch 1996). Suppose d = 2 m. generate X(t d ), then X(t d/2 ) conditional on (X(0), X(t d )), then X(t d/4 ) conditional on (X(0), X(t d/2 )), and so on. The first few N(0, 1) r.v. s already sketch the path trajectory. 62
115 Principal component decomposition (PCA) (Ackworth et al. 1998): A = PD 1/2 where D = diag(λ s,..., λ 1 ) (eigenvalues of Σ in decreasing order) and the columns of P are the corresponding unit-length eigenvectors. With this A, Z 1 accounts for the max amount of variance of Y, then Z 2 the max amount of variance cond. on Z 1, etc. Function of a Brownian motion (or other Lévy process): Payoff depends on c-dimensional Brownian motion {X(t), t 0} observed at times 0 = t 0 < t 1 < < t d = T. Sequential (or random walk) method: generate X(t 1 ), then X(t 2 ) X(t 1 ), then X(t 3 ) X(t 2 ), etc. Bridge sampling (Moskowitz and Caflisch 1996). Suppose d = 2 m. generate X(t d ), then X(t d/2 ) conditional on (X(0), X(t d )), then X(t d/4 ) conditional on (X(0), X(t d/2 )), and so on. The first few N(0, 1) r.v. s already sketch the path trajectory. Each of these methods corresponds to some matrix A. Choice has a large impact on the ANOVA decomposition of f. 62
116 Example: Pricing an Asian basket option We have c assets, d observation times. f (U) = e rt 1 max 0, cd Want to estimate E[f (U)], where c d S i (t j ) K i=1 j=1 is the net discounted payoff and S i (t j ) is the price of asset i at time t j. 63
117 Example: Pricing an Asian basket option We have c assets, d observation times. f (U) = e rt 1 max 0, cd Want to estimate E[f (U)], where c d S i (t j ) K i=1 j=1 is the net discounted payoff and S i (t j ) is the price of asset i at time t j. Suppose (S 1 (t),..., S c (t)) obeys a geometric Brownian motion. Then, f (U) = g(y) where Y = (Y 1,..., Y s ) N(0, Σ) and s = cd. 63
118 Example: Pricing an Asian basket option We have c assets, d observation times. f (U) = e rt 1 max 0, cd Want to estimate E[f (U)], where c d S i (t j ) K i=1 j=1 is the net discounted payoff and S i (t j ) is the price of asset i at time t j. Suppose (S 1 (t),..., S c (t)) obeys a geometric Brownian motion. Then, f (U) = g(y) where Y = (Y 1,..., Y s ) N(0, Σ) and s = cd. Even with Cholesky decompositions of Σ, the two-dimensional projections often account for more than 99% of the variance: low effective dimension in the superposition sense. With PCA or bridge sampling, we get low effective dimension in the truncation sense. In realistic examples, the first two coordinates Z 1 and Z 2 often account for more than 99.99% of the variance! 63
119 Numerical experiment with c = 10 and d = 25 This gives a 250-dimensional integration problem. Let ρ i,j = 0.4 for all i j, T = 1, σ i = (i 1)/9 for all i, r = 0.04, S(0) = 100, and K = 100. (Imai and Tan 2002). 64
120 Numerical experiment with c = 10 and d = 25 This gives a 250-dimensional integration problem. Let ρ i,j = 0.4 for all i j, T = 1, σ i = (i 1)/9 for all i, r = 0.04, S(0) = 100, and K = 100. (Imai and Tan 2002). Variance reduction factors for Cholesky (left) and PCA (right) (experiment from 2003): Korobov Lattice Rules n = n = n = a = 5693 a = 944 a = Lattice+shift Lattice+shift+baker Sobol Nets n = 2 14 n = 2 16 n = 2 18 Sobol+Shift Sobol+LMS+Shift Note: The payoff function is not smooth and also unbounded! 64
121 ANOVA Variances for ordinary Asian Option s = 3, seq. s = 3, BB s = 3, PCA s = 6, seq. s = 6, BB s = 6, PCA s = 12, seq. s = 12, BB s = 12, PCA Asian Option with S(0) = 100, K = 100, r = 0.05, σ = % of total variance for each cardinality of u 65
122 66 Total Variance per Coordinate for the Asian Option Asian Option (s = 6) with S(0) = 100, K = 100, r = 0.05, σ = 0.5 sequential BB PCA % of total variance Coordinate 1 Coordinate 2 Coordinate 3 Coordinate 4 Coordinate 5 Coordinate 6
123 Variance with good lattices rules and Sobol points variance Asian Option (PCA) s = 12, S(0) = 100, K = 100, r = 0.05, σ = 0.5 MC Sobol Lattice (P 2) + baker n n 67
124 68 Asian Option on a Single Asset, with control variate Let c = 1, S(0) = 100, r = ln(1.09), σ i = 0.2, T = 120/365, t j = D 1 /365 + (T D 1 /365)(j 1)/(d 1) for j = 1,..., d,
125 68 Asian Option on a Single Asset, with control variate Let c = 1, S(0) = 100, r = ln(1.09), σ i = 0.2, T = 120/365, t j = D 1 /365 + (T D 1 /365)(j 1)/(d 1) for j = 1,..., d, We estimated the optimal CV coefficient by pilot runs for MC and for each combination of sampling scheme, RQMC method, and n.
126 68 Asian Option on a Single Asset, with control variate Let c = 1, S(0) = 100, r = ln(1.09), σ i = 0.2, T = 120/365, t j = D 1 /365 + (T D 1 /365)(j 1)/(d 1) for j = 1,..., d, We estimated the optimal CV coefficient by pilot runs for MC and for each combination of sampling scheme, RQMC method, and n. d D 1 K µ σ 2 VRF of CV
127 VRFs (per run) for RQMC vs MC, with n Sequential sampling (left), bridge sampling (middle), and PCA (right). d D 1 K P n without CV with CV SEQ BBS PCA SEQ BBS PCA Kor+S Kor+S+B Sob+DS Kor+S Kor+S+B Sob+DS Kor+S Kor+S+B Sob+DS For d = 10, Sobol with PCA combined with CV reduces the variance approximately by a factor of , without increasing the CPU time. For d = 120, PCA is slower than SEQ by a factor of 2 or 3, but worth it. 69
128 Array-RQMC for Markov Chains Setting: A Markov chain with state space X R l, evolves as X 0 = x 0, X j = ϕ j (X j 1, U j ), j 1, where the U j are i.i.d. uniform r.v. s over (0, 1) d. Want to estimate µ = E[Y ] where Y = τ g j (X j ). Ordinary MC: n i.i.d. realizations of Y. Requires τs uniforms. Array-RQMC: L., Lécot, Tuffin, et al. [2004, 2006, 2008, etc.] Simulate an array (or population) of n chains in parallel. Goal: Want small discrepancy between empirical distribution of states S n,j = {X 0,j,..., X n 1,j } and theoretical distribution of X j, at each step j. At each step, use RQMC point set to advance all the chains by one step. j=1 70
129 Some RQMC insight: To simplify, suppose X j U(0, 1) l. We estimate µ j = E[g j (X j )] = E[g j (ϕ j (X j 1, U))] = g j (ϕ j (x, u))dxdu [0,1) l+d by ˆµ arqmc,j,n = 1 n 1 g j (X i,j ) = 1 n 1 g j (ϕ j (X i,j 1, U i,j )). n n i=0 This is (roughly) RQMC with the point set Q n = {(X i,j 1, U i,j ), 0 i < n}. We want Q n to have low discrepancy (LD) over [0, 1) l+d. i=0 71
130 Some RQMC insight: To simplify, suppose X j U(0, 1) l. We estimate µ j = E[g j (X j )] = E[g j (ϕ j (X j 1, U))] = g j (ϕ j (x, u))dxdu [0,1) l+d by ˆµ arqmc,j,n = 1 n 1 g j (X i,j ) = 1 n 1 g j (ϕ j (X i,j 1, U i,j )). n n i=0 This is (roughly) RQMC with the point set Q n = {(X i,j 1, U i,j ), 0 i < n}. We want Q n to have low discrepancy (LD) over [0, 1) l+d. We do not choose the X i,j 1 s in Q n : they come from the simulation. We select a LD point set i=0 Q n = {(w 0, U 0,j ),..., (w n 1, U n 1,j )}, where the w i [0, 1) l are fixed and each U i,j U(0, 1) d. Permute the states X i,j 1 so that X πj (i),j 1 is close to w i for each i (LD between the two sets), and compute X i,j = ϕ j (X πj (i),j 1, U i,j ) for each i. Example: If l = 1, can take w i = (i + 0.5)/n and just sort the states. For l > 1, there are various ways to define the matching (multivariate sort). 71
131 Array-RQMC algorithm X i,0 x 0 (or X i,0 x i,0 ) for i = 0,..., n 1; for j = 1, 2,..., τ do Compute the permutation π j of the states (for matching); Randomize afresh {U 0,j,..., U n 1,j } in Q n ; X i,j = ϕ j (X πj (i),j 1, U i,j ), for i = 0,..., n 1; n 1 ˆµ arqmc,j,n = Ȳ n,j = 1 n i=0 g(x i,j); Estimate µ by the average Ȳn = ˆµ arqmc,n = τ j=1 ˆµ arqmc,j,n. 72
132 Array-RQMC algorithm X i,0 x 0 (or X i,0 x i,0 ) for i = 0,..., n 1; for j = 1, 2,..., τ do Compute the permutation π j of the states (for matching); Randomize afresh {U 0,j,..., U n 1,j } in Q n ; X i,j = ϕ j (X πj (i),j 1, U i,j ), for i = 0,..., n 1; n 1 ˆµ arqmc,j,n = Ȳ n,j = 1 n i=0 g(x i,j); Estimate µ by the average Ȳn = ˆµ arqmc,n = τ j=1 ˆµ arqmc,j,n. Proposition: (i) The average Ȳn is an unbiased estimator of µ. (ii) The empirical variance of m independent realizations gives an unbiased estimator of Var[Ȳ n ]. 72
Randomized Quasi-Monte Carlo: An Introduction for Practitioners
Randomized Quasi-Monte Carlo: An Introduction for Practitioners Pierre L Ecuyer Abstract We survey basic ideas and results on randomized quasi-monte Carlo (RQMC) methods, discuss their practical aspects,
More informationRandomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space
Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space Pierre L Ecuyer 1, Christian Lécot 2, and Bruno Tuffin 3 1 Département d informatique et de recherche opérationnelle,
More informationVariance Reduction s Greatest Hits
1 Variance Reduction s Greatest Hits Pierre L Ecuyer CIRRELT, GERAD, and Département d Informatique et de Recherche Opérationnelle Université de Montréal, Canada ESM 2007, Malta, October 22, 2007 Context
More informationA Randomized Quasi-Monte Carlo Simulation Method for Markov Chains
A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains Pierre L Ecuyer GERAD and Département d Informatique et de Recherche Opérationnelle Université de Montréal, C.P. 6128, Succ. Centre-Ville,
More informationA Randomized Quasi-Monte Carlo Simulation Method for Markov Chains
A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains Pierre L Ecuyer GERAD and Département d Informatique et de Recherche Opérationnelle Université de Montréal, C.P. 6128, Succ. Centre-Ville,
More informationQuasi- Monte Carlo Multiple Integration
Chapter 6 Quasi- Monte Carlo Multiple Integration Introduction In some sense, this chapter fits within Chapter 4 on variance reduction; in some sense it is stratification run wild. Quasi-Monte Carlo methods
More informationTutorial on quasi-monte Carlo methods
Tutorial on quasi-monte Carlo methods Josef Dick School of Mathematics and Statistics, UNSW, Sydney, Australia josef.dick@unsw.edu.au Comparison: MCMC, MC, QMC Roughly speaking: Markov chain Monte Carlo
More informationQUASI-MONTE CARLO VIA LINEAR SHIFT-REGISTER SEQUENCES. Pierre L Ecuyer Christiane Lemieux
Proceedings of the 1999 Winter Simulation Conference P. A. Farrington, H. B. Nembhard, D. T. Sturrock, and G. W. Evans, eds. QUASI-MONTE CARLO VIA LINEAR SHIFT-REGISTER SEQUENCES Pierre L Ecuyer Christiane
More informationQuasi-Monte Carlo methods with applications in finance
Finance Stoch (2009) 13: 307 349 DOI 10.1007/s00780-009-0095-y Quasi-Monte Carlo methods with applications in finance Pierre L Ecuyer Received: 30 May 2008 / Accepted: 20 August 2008 / Published online:
More informationOn the Behavior of the Weighted Star Discrepancy Bounds for Shifted Lattice Rules
On the Behavior of the Weighted Star Discrepancy Bounds for Shifted Lattice Rules Vasile Sinescu and Pierre L Ecuyer Abstract We examine the question of constructing shifted lattice rules of rank one with
More informationLow Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed?
Low Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed? Xiaoqun Wang 1,2 and Ian H. Sloan 2 1 Department of Mathematical Sciences, Tsinghua University, Beijing 100084,
More informationA Coding Theoretic Approach to Building Nets with Well-Equidistributed Projections
A Coding Theoretic Approach to Building Nets with Well-Equidistributed Projections Yves Edel 1 and Pierre L Ecuyer 2 1 Mathematisches Institut der Universität Im Neuenheimer Feld 288, 69120 Heidelberg,
More informationMathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )
Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca October 22nd, 2014 E. Tanré (INRIA - Team Tosca) Mathematical
More informationFast evaluation of mixed derivatives and calculation of optimal weights for integration. Hernan Leovey
Fast evaluation of mixed derivatives and calculation of optimal weights for integration Humboldt Universität zu Berlin 02.14.2012 MCQMC2012 Tenth International Conference on Monte Carlo and Quasi Monte
More informationarxiv: v1 [math.pr] 3 Oct 2007
Monte Carlo Methods and Path-Generation techniques for Pricing Multi-asset Path-dependent Options arxiv:0710.0850v1 [math.pr] 3 Oct 2007 Piergiacomo Sabino Dipartimento di Matematica Università degli Studi
More informationVariance and discrepancy with alternative scramblings
Variance and discrepancy with alternative scramblings ART B. OWEN Stanford University This paper analyzes some schemes for reducing the computational burden of digital scrambling. Some such schemes have
More informationQMC methods in quantitative finance. and perspectives
, tradition and perspectives Johannes Kepler University Linz (JKU) WU Research Seminar What is the content of the talk Valuation of financial derivatives in stochastic market models using (QMC-)simulation
More informationProgress in high-dimensional numerical integration and its application to stochastic optimization
Progress in high-dimensional numerical integration and its application to stochastic optimization W. Römisch Humboldt-University Berlin Department of Mathematics www.math.hu-berlin.de/~romisch Short course,
More informationSCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES. Kinjal Basu Art B. Owen
SCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES By Kinjal Basu Art B. Owen Technical Report No. 2015-07 March 2015 Department of Statistics STANFORD UNIVERSITY Stanford, California 94305-4065
More informationQuasi-Monte Carlo integration over the Euclidean space and applications
Quasi-Monte Carlo integration over the Euclidean space and applications f.kuo@unsw.edu.au University of New South Wales, Sydney, Australia joint work with James Nichols (UNSW) Journal of Complexity 30
More informationMonte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan
Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis
More informationRandomized Quasi-Monte Carlo for MCMC
Randomized Quasi-Monte Carlo for MCMC Radu Craiu 1 Christiane Lemieux 2 1 Department of Statistics, Toronto 2 Department of Statistics, Waterloo Third Workshop on Monte Carlo Methods Harvard, May 2007
More informationScientific Computing: Monte Carlo
Scientific Computing: Monte Carlo Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 April 5th and 12th, 2012 A. Donev (Courant Institute)
More informationExtensible grids: uniform sampling on a space-filling curve
Extensible grids: uniform sampling on a space-filling curve Zhijian He Tsinghua University Art B. Owen Stanford University June 2014 Abstract We study the properties of points in [0, 1] d generated by
More informationScenario Generation and Sampling Methods
Scenario Generation and Sampling Methods Tito Homem-de-Mello Güzin Bayraksan SVAN 2016 IMPA 9 13 May 2106 Let s Start with a Recap Sample Average Approximation We want to solve the true problem min {g(x)
More informationRandom and Quasi-Random Point Sets
Peter Hellekalek Gerhard Larcher (Editors) Random and Quasi-Random Point Sets With contributions by JözsefBeck Peter Hellekalek Fred J. Hickemell Gerhard Larcher Pierre L'Ecuyer Harald Niederreiter Shu
More informationResolution-Stationary Random Number Generators
Resolution-Stationary Random Number Generators Francois Panneton Caisse Centrale Desjardins, 1 Complexe Desjardins, bureau 2822 Montral (Québec), H5B 1B3, Canada Pierre L Ecuyer Département d Informatique
More informationChapter 4: Monte-Carlo Methods
Chapter 4: Monte-Carlo Methods A Monte-Carlo method is a technique for the numerical realization of a stochastic process by means of normally distributed random variables. In financial mathematics, it
More informationQuasi-Random Simulation of Discrete Choice Models
Quasi-Random Simulation of Discrete Choice Models by Zsolt Sándor and Kenneth Train Erasmus University Rotterdam and University of California, Berkeley June 12, 2002 Abstract We describe the properties
More informationPart III. Quasi Monte Carlo methods 146/349
Part III Quasi Monte Carlo methods 46/349 Outline Quasi Monte Carlo methods 47/349 Quasi Monte Carlo methods Let y = (y,...,y N ) be a vector of independent uniform random variables in Γ = [0,] N, u(y)
More informationPseudo-Random Numbers Generators. Anne GILLE-GENEST. March 1, Premia Introduction Definitions Good generators...
14 pages 1 Pseudo-Random Numbers Generators Anne GILLE-GENEST March 1, 2012 Contents Premia 14 1 Introduction 2 1.1 Definitions............................. 2 1.2 Good generators..........................
More informationImproved Discrepancy Bounds for Hybrid Sequences. Harald Niederreiter. RICAM Linz and University of Salzburg
Improved Discrepancy Bounds for Hybrid Sequences Harald Niederreiter RICAM Linz and University of Salzburg MC vs. QMC methods Hybrid sequences The basic sequences Deterministic discrepancy bounds The proof
More informationNonuniform Random Variate Generation
Nonuniform Random Variate Generation 1 Suppose we have a generator of i.i.d. U(0, 1) r.v. s. We want to generate r.v. s from other distributions, such as normal, Weibull, Poisson, etc. Or other random
More informationarxiv: v2 [stat.co] 22 Jun 2017
A randomized Halton algorithm in R arxiv:1706.02808v2 [stat.co] 22 Jun 2017 Art B. Owen Stanford University May 2017 Abstract Randomized quasi-monte Carlo (RQMC) sampling can bring orders of magnitude
More informationtypical random point set. There are several ways of dening a discrepancy, many of them leading to an error bound of the form je n j V (f)d(p n ) for a
QUASI-MONTE CARLO VIA LINEAR SHIFT-REGISTER SEQUENCES Pierre L'Ecuyer and Christiane Lemieux Departement d'informatique et de Recherche Operationnelle Universite de Montreal, C.P. 6128, Succ. Centre-Ville
More informationAnalysis and Improvements of Path-Based Methods for Monte Carlo Reliability Evaluation of Static Models
Chapter Analysis and Improvements of Path-Based Methods for Monte Carlo Reliability Evaluation of Static Models Héctor Cancela, Pierre L Ecuyer, Matías Lee, Gerardo Rubino and Bruno Tuffin Abstract Many
More informationImportance Sampling to Accelerate the Convergence of Quasi-Monte Carlo
Importance Sampling to Accelerate the Convergence of Quasi-Monte Carlo Wolfgang Hörmann, Josef Leydold Department of Statistics and Mathematics Wirtschaftsuniversität Wien Research Report Series Report
More informationDiscrepancy Bounds for Deterministic Acceptance-Rejection Samplers
Discrepancy Bounds for Deterministic Acceptance-Rejection Samplers Houying Zhu Athesisinfulfilmentoftherequirementsforthedegreeof Doctor of Philosophy School of Mathematics and Statistics Faculty of Science
More informationImportance Sampling and Stratification Techniques for Multivariate Models with Low-Dimensional Structures
Importance Sampling and Stratification Techniques for Multivariate Models with Low-Dimensional Structures by Yoshihiro Taniguchi A thesis presented to the University of Waterloo in fulfillment of the thesis
More informationHochdimensionale Integration
Oliver Ernst Institut für Numerische Mathematik und Optimierung Hochdimensionale Integration 14-tägige Vorlesung im Wintersemester 2010/11 im Rahmen des Moduls Ausgewählte Kapitel der Numerik Contents
More informationGaussian, Markov and stationary processes
Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November
More informationNumerical Methods II
Numerical Methods II Prof. Mike Giles mike.giles@maths.ox.ac.uk Oxford University Mathematical Institute MC Lecture 13 p. 1 Quasi-Monte Carlo As in lecture 6, quasi-monte Carlo methods offer much greater
More informationStat 890 Design of computer experiments
Stat 890 Design of computer experiments Will introduce design concepts for computer experiments Will look at more elaborate constructions next day Experiment design In computer experiments, as in many
More informationQuasi-Monte Carlo Algorithms with Applications in Numerical Analysis and Finance
Quasi-Monte Carlo Algorithms with Applications in Numerical Analysis and Finance Reinhold F. Kainhofer Rigorosumsvortrag, Institut für Mathematik Graz, 16. Mai 2003 Inhalt 1. Quasi-Monte Carlo Methoden
More informationLatin Hypercube Sampling with Multidimensional Uniformity
Latin Hypercube Sampling with Multidimensional Uniformity Jared L. Deutsch and Clayton V. Deutsch Complex geostatistical models can only be realized a limited number of times due to large computational
More informationA CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS
Statistica Sinica 26 (2016), 1117-1128 doi:http://dx.doi.org/10.5705/ss.202015.0240 A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS Xu He and Peter Z. G. Qian Chinese Academy of Sciences
More informationUniform Random Number Generation
Uniform Random Number Generation Pierre L Ecuyer Département d Informatique et de Recherche Opérationnelle, Université de Montréal, C.P. 6128, Succ. Centre-Ville, Montréal (Québec), H9S 5B8, Canada. http://www.iro.umontreal.ca/
More informationNumerical Methods in Economics MIT Press, Chapter 9 Notes Quasi-Monte Carlo Methods. Kenneth L. Judd Hoover Institution.
1 Numerical Methods in Economics MIT Press, 1998 Chapter 9 Notes Quasi-Monte Carlo Methods Kenneth L. Judd Hoover Institution October 26, 2002 2 Quasi-Monte Carlo Methods Observation: MC uses random sequences
More informationError Analysis for Quasi-Monte Carlo Methods
Error Analysis for Quasi-Monte Carlo Methods Fred J. Hickernell Department of Applied Mathematics, Illinois Institute of Technology hickernell@iit.edu mypages.iit.edu/~hickernell Thanks to the Guaranteed
More informationQUASI-MONTE CARLO METHODS FOR SIMULATION. Pierre L Ecuyer
Proceedings of the 2003 Winter Simulation Conference S Chick, P J Sánchez, D Ferrin, and D J Morrice, eds QUASI-MONTE CARLO METHODS FOR SIMULATION Pierre L Ecuyer Département d Informatique et de Recherche
More informationMarkov Chain Monte Carlo Lecture 1
What are Monte Carlo Methods? The subject of Monte Carlo methods can be viewed as a branch of experimental mathematics in which one uses random numbers to conduct experiments. Typically the experiments
More informationThe square root rule for adaptive importance sampling
The square root rule for adaptive importance sampling Art B. Owen Stanford University Yi Zhou January 2019 Abstract In adaptive importance sampling, and other contexts, we have unbiased and uncorrelated
More informationLecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes
Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities
More informationLow Discrepancy Toolbox Manual. Michael Baudin
Low Discrepancy Toolbox Manual Michael Baudin Version 0.1 May 2013 Abstract This document is an introduction to the Low Discrepancy module v0.5. Contents 1 Gallery 4 1.1 The Van Der Corput sequence.............................
More informationPolynomial Integration Lattices
Polynomial Integration Lattices Pierre L Ecuyer Département d informatique et de recherche opérationnelle Université de Montréal, C.P. 6128, Succ. Centre-Ville Montréal (Québec), H3C 3J7, Canada lecuyer@iro.umontreal.ca
More information1 Simulating normal (Gaussian) rvs with applications to simulating Brownian motion and geometric Brownian motion in one and two dimensions
Copyright c 2007 by Karl Sigman 1 Simulating normal Gaussian rvs with applications to simulating Brownian motion and geometric Brownian motion in one and two dimensions Fundamental to many applications
More informationLes Cahiers du GERAD ISSN:
Les Cahiers du GERAD ISSN: 0711 2440 On Array-RQMC for Markov Chains: Mapping Alternatives and Convergence Rates P. L Ecuyer, C. Lécot, A. L Archevêque-Gaudet G 2009 03 January 2009 Les textes publiés
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More information1. Introduction. In this paper we consider stochastic optimization problems of the form
O RATES OF COVERGECE FOR STOCHASTIC OPTIMIZATIO PROBLEMS UDER O-I.I.D. SAMPLIG TITO HOMEM-DE-MELLO Abstract. In this paper we discuss the issue of solving stochastic optimization problems by means of sample
More information2 Random Variable Generation
2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods
More informationMultivariate Distributions
IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate
More informationTRB 2005: For Presentation and Publication Paper # Final Submission: April 1, 2005 Word Count: 8,191
Simulation Estimation of Mixed Discrete Choice Models Using Randomized Quasi-Monte Carlo Sequences: A Comparison of Alternative Sequences, Scrambling Methods, and Uniform-to-Normal Variate Transformation
More informationMonte Carlo Method for Numerical Integration based on Sobol s Sequences
Monte Carlo Method for Numerical Integration based on Sobol s Sequences Ivan Dimov and Rayna Georgieva Institute of Information and Communication Technologies, Bulgarian Academy of Sciences, 1113 Sofia,
More informationConstruction of Equidistributed Generators Based on Linear Recurrences Modulo 2
Construction of Equidistributed Generators Based on Linear Recurrences Modulo 2 Pierre L Ecuyer and François Panneton Département d informatique et de recherche opérationnelle Université de Montréal C.P.
More informationUniform Random Number Generators
JHU 553.633/433: Monte Carlo Methods J. C. Spall 25 September 2017 CHAPTER 2 RANDOM NUMBER GENERATION Motivation and criteria for generators Linear generators (e.g., linear congruential generators) Multiple
More informationApplicability of Quasi-Monte Carlo for lattice systems
Applicability of Quasi-Monte Carlo for lattice systems Andreas Ammon 1,2, Tobias Hartung 1,2,Karl Jansen 2, Hernan Leovey 3, Andreas Griewank 3, Michael Müller-Preussker 1 1 Humboldt-University Berlin,
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationQMC tutorial 1. Quasi-Monte Carlo. Art B. Owen Stanford University. A tutorial reflecting personal views on what is important in QMC.
QMC tutorial 1 Quasi-Monte Carlo Art B. Owen Stanford University A tutorial reflecting personal views on what is important in QMC. MCQMC 2016 will be at Stanford, August 14-19 mcqmc2016.stanford.edu QMC
More informationDigital lattice rules
Digital lattice rules Multivariate integration and discrepancy estimates A thesis presented to The University of New South Wales in fulfilment of the thesis requirement for the degree of Doctor of Philosophy
More informationANOVA decomposition of convex piecewise linear functions
ANOVA decomposition of convex piecewise linear functions W. Römisch Abstract Piecewise linear convex functions arise as integrands in stochastic programs. They are Lipschitz continuous on their domain,
More informationSTRONG TRACTABILITY OF MULTIVARIATE INTEGRATION USING QUASI MONTE CARLO ALGORITHMS
MATHEMATICS OF COMPUTATION Volume 72, Number 242, Pages 823 838 S 0025-5718(02)01440-0 Article electronically published on June 25, 2002 STRONG TRACTABILITY OF MULTIVARIATE INTEGRATION USING QUASI MONTE
More informationOptimal Randomized Algorithms for Integration on Function Spaces with underlying ANOVA decomposition
Optimal Randomized on Function Spaces with underlying ANOVA decomposition Michael Gnewuch 1 University of Kaiserslautern, Germany October 16, 2013 Based on Joint Work with Jan Baldeaux (UTS Sydney) & Josef
More informationQUASI-RANDOM NUMBER GENERATION IN THE VIEW OF PARALLEL IMPLEMENTATION. 1. Introduction
Trends in Mathematics Information Center for Mathematical Sciences Volume 6, Number 2, December, 2003, Pages 155 162 QUASI-RANDOM NUMBER GENERATION IN THE VIEW OF PARALLEL IMPLEMENTATION CHI-OK HWANG Abstract.
More informationModel Counting for Logical Theories
Model Counting for Logical Theories Wednesday Dmitry Chistikov Rayna Dimitrova Department of Computer Science University of Oxford, UK Max Planck Institute for Software Systems (MPI-SWS) Kaiserslautern
More informationSolutions of the Financial Risk Management Examination
Solutions of the Financial Risk Management Examination Thierry Roncalli January 9 th 03 Remark The first five questions are corrected in TR-GDR and in the document of exercise solutions, which is available
More information6.1 Moment Generating and Characteristic Functions
Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,
More informationBayesian Methods with Monte Carlo Markov Chains II
Bayesian Methods with Monte Carlo Markov Chains II Henry Horng-Shing Lu Institute of Statistics National Chiao Tung University hslu@stat.nctu.edu.tw http://tigpbp.iis.sinica.edu.tw/courses.htm 1 Part 3
More informationLecture 2: Linear Algebra Review
EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1
More informationComplexity of two and multi-stage stochastic programming problems
Complexity of two and multi-stage stochastic programming problems A. Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332-0205, USA The concept
More informationToday: Fundamentals of Monte Carlo
Today: Fundamentals of Monte Carlo What is Monte Carlo? Named at Los Alamos in 1940 s after the casino. Any method which uses (pseudo)random numbers as an essential part of the algorithm. Stochastic -
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationSelected Exercises on Expectations and Some Probability Inequalities
Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ
More informationMultivariate Time Series: VAR(p) Processes and Models
Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with
More informationAPPLIED MATHEMATICS REPORT AMR04/16 FINITE-ORDER WEIGHTS IMPLY TRACTABILITY OF MULTIVARIATE INTEGRATION. I.H. Sloan, X. Wang and H.
APPLIED MATHEMATICS REPORT AMR04/16 FINITE-ORDER WEIGHTS IMPLY TRACTABILITY OF MULTIVARIATE INTEGRATION I.H. Sloan, X. Wang and H. Wozniakowski Published in Journal of Complexity, Volume 20, Number 1,
More informationThe rotor-router mechanism for quasirandom walk and aggregation. Jim Propp (U. Mass. Lowell) July 10, 2008
The rotor-router mechanism for quasirandom walk and aggregation Jim Propp (U. Mass. Lowell) July 10, 2008 (based on articles in progress with Ander Holroyd and Lionel Levine; with thanks also to Hal Canary,
More informationThe strictly 1/2-stable example
The strictly 1/2-stable example 1 Direct approach: building a Lévy pure jump process on R Bert Fristedt provided key mathematical facts for this example. A pure jump Lévy process X is a Lévy process such
More informationSTAT232B Importance and Sequential Importance Sampling
STAT232B Importance and Sequential Importance Sampling Gianfranco Doretto Andrea Vedaldi June 7, 2004 1 Monte Carlo Integration Goal: computing the following integral µ = h(x)π(x) dx χ Standard numerical
More informationSYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions
SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationStatistical Data Analysis
DS-GA 0 Lecture notes 8 Fall 016 1 Descriptive statistics Statistical Data Analysis In this section we consider the problem of analyzing a set of data. We describe several techniques for visualizing the
More informationToday: Fundamentals of Monte Carlo
Today: Fundamentals of Monte Carlo What is Monte Carlo? Named at Los Alamos in 940 s after the casino. Any method which uses (pseudo)random numbers as an essential part of the algorithm. Stochastic - not
More informationTest Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics
Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics The candidates for the research course in Statistics will have to take two shortanswer type tests
More informationLattice rules and sequences for non periodic smooth integrands
Lattice rules and sequences for non periodic smooth integrands dirk.nuyens@cs.kuleuven.be Department of Computer Science KU Leuven, Belgium Joint work with Josef Dick (UNSW) and Friedrich Pillichshammer
More informationFlorida State University Libraries
Florida State University Libraries Electronic Theses, Treatises and Dissertations The Graduate School 29 Monte Carlo and Quasi-Monte Carlo Methods in Financial Derivative Pricing Ahmet Göncü Follow this
More information6 Markov Chain Monte Carlo (MCMC)
6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution
More informationEco517 Fall 2004 C. Sims MIDTERM EXAM
Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering
More informationThe Multivariate Normal Distribution. In this case according to our theorem
The Multivariate Normal Distribution Defn: Z R 1 N(0, 1) iff f Z (z) = 1 2π e z2 /2. Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) T with the Z i independent and each Z i N(0, 1). In this
More informationThe Error Analysis through Dual Net and Walsh Function in Quasi-Monte Carlo Method
The Error Analysis through Dual Net and Walsh Function in Quasi-Monte Carlo Method Siyuan Deng Department of Applied Mathematics Illinois Institute of Technology Apr 25, 2013 The Basic Concept of Quasi-Monte
More informationStat 516, Homework 1
Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball
More informationA GENERAL THEORY FOR ORTHOGONAL ARRAY BASED LATIN HYPERCUBE SAMPLING
Statistica Sinica 26 (2016), 761-777 doi:http://dx.doi.org/10.5705/ss.202015.0029 A GENERAL THEORY FOR ORTHOGONAL ARRAY BASED LATIN HYPERCUBE SAMPLING Mingyao Ai, Xiangshun Kong and Kang Li Peking University
More information