New results for the functional ANOVA and. Sobol indices

Size: px
Start display at page:

Download "New results for the functional ANOVA and. Sobol indices"

Transcription

1 Generalized Sobol Indices 1 New results for the functional ANOVA and Sobol indices Art B. Owen Stanford University

2 Generalized Sobol Indices 2 Ilya Meerovich Sobol At MCM 2001, Salzburg Known for Sobol sequences and Sobol indices Every time I read one of his papers, I wish I d read it earlier

3 Generalized Sobol Indices 3 Simplified Saint-Venant flood model Overflow in meters (Lamboni, Iooss, Popelin, Gamboa, 2012) at a dyke S = Z v + H H d C b, ( Q H = BK s (Zm Z v )/L where ) 3/5 (max annual river height) Q Maximal annual flow m 3 /s Gumbel(1013, 558) [500, 3000] K s Strickler coefficient m 1/3 /s N (30, 8) [15, ) Z v River downstream level m Triangle(49, 50, 51) Z m River upstream level m Triangle(54, 55, 56) H d Dyke height m U[7, 9] C b Bank level m Triangle(55, 55.5, 56) L Length of river stretch m Triangle(4990, 5000, 5010) B River width m Triangle(295, 300, 305) Reduced from a Navier-Stokes model

4 Generalized Sobol Indices 4 The cost model C p = 1 S>0m (flood cost) + 1 S 0m ( (1 e 1000m4 /S 4 )) (dyke maintenance) min(h d m 1, 8) (investment cost, from construction) in millions of Euros A discontinuity at S = 0 would be better. Quibble

5 Generalized Sobol Indices 5 Variable importance Which of these variables is most important? How important is any given subset of variables taken together? How should one define/measure variable importance? Similar problems come up in black box models for aerospace, semiconductor manufacturing, climate modeling, etc.

6 Generalized Sobol Indices 6 Why do they care? The ultimate goal may be to optimize something (a mean or variance or max or min) by choosing levels for those variables under your control. Understanding a function is an intermediate goal, not an ultimate goal. Yet variable importance is widely applied. Similar question Why compute R 2? Sobol indices yield many analogues of R 2. Quantitative measures of importance aid discussions.

7 Generalized Sobol Indices 7 Quasi-Monte Carlo QMC sometimes gives accurate estimates for high dimensional integrals. Despite the curse of dimensionality. When it happens, we usually find that the integrand was dominated by low order interactions among the variables. That understanding motivates trying to lower the effective dimension of our integrands. (Analogous to variance reduction.) which never actually said that all high dim problems are insoluble.

8 Generalized Sobol Indices 8 ANOVA Fisher (1923) to Hoeffding (1948) to Sobol (1967) to Efron & Stein (1981) then back to Sobol (1990/3) Recall X ij = X + ( X i X ) + ( X j X ) + (X ij X i X j + X ) Take a grand average X Subtract it out and average over j, getting X i X Similarly get X j X Generally: subtract away all lower order effects and average out the extra variables.

9 Generalized Sobol Indices 9 Extensions of ANOVA From functions on {1, 2,..., I} {1, 2,..., J} to {1, 2,..., I} {1, 2,..., J} {1, 2,..., Z} To functions in L 2 [0, 1] d. To L 2 functions of d arbitrary independent inputs. Also d = works via martingales.

10 Generalized Sobol Indices 10 ANOVA for L 2 [0, 1] d Goes back to Hoeffding (1948) for U -statistics (skillful reading may be required) & Efron & Stein (1981) for jackknife f(x) = f () () + = f () () + d f (j) (x j ) + f (j,k) (x j, x k ) + + f (1,2,...,d) (x 1,..., x d ) j=1 j<k d f (j1,j 2,...,j d )(x j1, x j2,..., x jd ) r=1 1 j 1 <j 2 < <j r d More simply f(x) = u f u (x) Sum over all u D = {1, 2,..., d}

11 Generalized Sobol Indices 11 Notation For u D {1,..., d} u = card(u) u = u c = {1, 2,..., d} u v u strict subset i.e. If u = {j 1, j 2,..., j u } then x u = (x j1,..., x j u ) and dx u = j u dx j

12 Generalized Sobol Indices 12 Recursive definition For u {1,..., d}, f u (x) only depends on x j for j u. Overall mean µ f (x) = f(x) dx (f(x) Main effect j f {j} (x) = f (x) ) dx {j} (f(x) Interaction u f u (x) = f v (x) ) dx u = Dependence v u f(x) dx u v u f v (x) f u (x) is a function of x that happens to only depend on x u f u (x) + f v (x) makes sense because they re on the same domain

13 Generalized Sobol Indices 13 ANOVA properties j u = u v = & 1 f u (x) dx j = 0 0 f u (x)f v (x) dx = 0 f u (x)g v (x) dx = 0 Variances Var(f) (f(x) µ) 2 dx = u {1,...,d} σu 2 = σu(f) 2 fu (x) 2 dx u = 0 u =. σ 2 u

14 Generalized Sobol Indices 14 Sobol s decomposition Let φ 0, φ 1, φ 2... be a complete orthonormal basis of L 2 [0, 1] with φ 0 (x) 1. Sobol (1969) expanded f(x) in a tensor product basis (Haar wavelets). He grouped the terms into 2 d subsets depending on which inputs are active. Sobol has a synthesis not an analysis for this decomposition. Thanks to A. Chouldechova for translation.

15 Generalized Sobol Indices 15 How important is x u? Variable importance Larger σ 2 u means that f u(x) contributes more. We also want to count σ 2 v for v u. Sobol s importance measures τ 2 u = v u σ 2 v v contained in u τ 2 u = v u σ 2 v v touches u, so interactions count Identity: τ 2 u = σ 2 τ 2 u Normalized versions: τ 2 u σ 2 and τ 2 u σ 2

16 Generalized Sobol Indices 16 More derived importance measures Superset importance Υ 2 u = v u σ 2 v Liu & O (2006) Small Υ 2 u means deleting f u and f v for v u (to stay hierarchical) makes little difference. Relevant to Hooker (2004) s simplifications of black box functions. Mean dimension u D σ 2 u σ 2 u Measures dimensionality of f. Liu & O (2006) Higher dimensionality makes for harder numerical handling. Many quadrature problems have mean dimension near 1

17 Generalized Sobol Indices 17 Estimation of τ 2 u and τ 2 u Naive approach for τ 2 u : 1) Sample x i [0, 1] d and get y i = f(x i ) for i = 1,..., n. 2) Statistical machine learning estimate ˆf v (x) for all necessary v. 3) Put ˆσ v 2 = ˆfu (x) 2 dx, u. 4) Sum: τ 2 u = v u ˆσ2 v. This is expensive and has many biases. Sobol has a much better way.

18 Generalized Sobol Indices 18 Evaluate f at two points: repeat some components independent draws for the others. Fixing methods Hybrid points For x, z [0, 1] d, y = x u :z u means x j, j u y j = z j, j u. We glue together part of x and part of z to form y = x u :z u. Sobol (1990/3) used the identities: τ 2 u = Cov(f(x), f(x u :z u )) τ 2 u = 1 2 E((f(x) f(x u:z u )) 2

19 Generalized Sobol Indices 19 Identity for τ 2 u f(x)f(x u :z u ) dx dz = f v (x)f v (x u :z u ) dx dz v D (orthogonality) = v u f v (x)f v (x u :z u ) dx dz (line integrals) = µ 2 + v u σ 2 u µ 2 + τ 2 u. τ 2 u = 1 n i=1 Bias adjustment n ( 1 f(x i )f(x i,u :z i, u ) n n i=1 ) 2 f(x i )

20 Generalized Sobol Indices 20 τ 2 u = Even better f(x) ( f(x u :z u ) f(z) ) dx dz τ 2 u = 1 n n f(x i ) ( f(x i,u :z i, u ) f(z i ) ) i=1 This avoids subtracting ˆµ 2. It is unbiased: E ( τ 2 u ) = τ 2 u Kucherenko, Feil, Shah, Mauntz (2011), Mauntz (2002), Saltelli (2002) τ 2 u = 1 n Improved statistical efficiency n f(x i )f(x i,u :z i, u ) i=1 Janon, Klein, Lagnoux, Nodet & Prieur (2012) ( 1 n n i=1 f(x i ) + f(x i,u :z i, u ) 2 Efficient in a class of estimators that does not include the unbiased one above. ) 2

21 Generalized Sobol Indices 21 For τ 2 u 1 2 (f(x) f(x u:z u ) ) 2 dx dz = 1 2 (σ 2 + µ 2 2 ( τ 2 u + µ2) + σ 2 + µ 2) = σ 2 τ 2 u = τ 2 u. Sobol s estimates are like tomography: integrals reveal internal structure.

22 Generalized Sobol Indices 22 τ 2 {j} for the flood model τ 2 /σ 2 Q K s Z v Z m H d C b L B Height H Overflow S Cost C p From n = 100,000 runs Q Maximal annual flow m 3 /s Gumbel(1013, 558) [500, 3000] K s Strickler coefficient m 1/3 /s N (30, 8) [15, ) Z v River downstream level m Triangle(49, 50, 51) Z m River upstream level m Triangle(54, 55, 56) H d Dyke height m U[7, 9] C b Bank level m Triangle(55, 55.5, 56) L Length of river stretch m Triangle(4990, 5000, 5010) B River width m Triangle(295, 300, 305)

23 Generalized Sobol Indices 23 For mean dimension d τ 2 j = j=1 d j=1 v {j} = d σ 2 v = v σ 2 v j=1 1 v {j} = = v v σ 2 v Estimator from Liu & O (2006) Generalizes to v k σv 2 for k 1. v

24 Generalized Sobol Indices 24 Example Kuo, Schwab, Sloan (2012) consider quadrature for f(x) = d j=1 xα j /j!, 0 < α 1. For α = 1 and d = 500 R = 50 replicated estimates of v v σ2 v/σ 2 using n = 10,000 had mean and standard deviation Upshot f(x) is nearly additive, though it is hard to quantify near perfect additivity. (The difficulty seems to be in forming the ratio.)

25 Generalized Sobol Indices 25 For superset importance Υ 2 u v u σ 2 v = 1 2 u ( 2 ( 1) u v f(x v :z v )) dx dz v u Mean of a square of differences better than differences of means of squares. From Liu & O (2006) Generalizes τ 2 u formula from 2 terms to 2 u terms. As a design Use n repeats of a 2 u 1 d u factorial randomly embedded in the unit cube. Does best in comparisons by Fruth, Roustant, Kuhnt (2012)

26 Generalized Sobol Indices 26 Generalized Sobol indices What can be attained via fixing methods? Θ uv = f(x u :z u )f(x v :z v ) dx dz Generalized Sobol index Ω uv Θ uv = tr(ω T Θ) u D v D Θ R 2d 2 d is the Sobol matrix. Ω R 2d 2 d has coefficients. Redundant (but useful) We have a 2 2d dimensional space of estimators for a 2 d dimensional space of estimands: δ µ 2 + δ u σu 2 u >0

27 Generalized Sobol Indices 27 NXOR XOR(u, v) = u v u v NXOR(u, v) = XOR(u, v) c = (u v) (u c v c ) (exclusive OR) (not exclusive OR) Θ uv f(x u :z u )f(x v :z v ) dx dz = µ 2 + τ 2 NXOR(u,v) Θ uv = 1 n f(x i,u :z i, u )f(x i,v :z i, v ) n i=1 Use tr(ω T Θ) often written XNOR

28 Generalized Sobol Indices 28 Special GSIs 1) Mean squares Ω = λλ T ( 2 λ u f(x u :z u )) dx dz Nonnegative u 2) Bilinear (rank one) Ω = λγ T ( )( λ u f(x u :z u ) u v ) γ v f(x v :z v ) dx dz Fast 3) Simple ( u ) λ u f(x u :z u ) f(z) dx dz Only uses one row/col of Θ 4) Contrast Ω u,v = 0 Free of µ 2 u v N.B.: Here a contrast can also be a sum of squares.

29 Generalized Sobol Indices 29 Cost of a GSI C(Ω) counts the # of function evaluations per (x, z) pair. Recipe 1) Count the rows u that are needed for some f(x u :z u ) 2) add the columns (where u appears as the needed v ) 3) subtract any doubly counted subsets We can have tr(ω T 1 Θ) = tr(ω T 2 Θ) but C(Ω 1 ) < C(Ω 2 ).

30 Generalized Sobol Indices 30 Squares For a square (or a sum of squares) tr(ω T Θ) 0. Also E ( tr(ω T Θ) ) = tr(ω T Θ) Therefore tr(ω T Θ) = 0 implies Pr ( tr(ω T Θ) = 0 ) = 1. GSIs with sum of squares estimators τ 2 u and Υ 2 u and No sum of squares exists for τ 2 u Can show that the coefficient of σ 2 D = u λ2 u generally u Ω uu i.e., tr(ω) Same thing happens in ANOVA tables: u σu 2 u when u < d every variance component has a contribution from the measurement error.

31 Generalized Sobol Indices 31 Targeting one variance component σ 2 {1,2,3} = τ 2 {1,2,3} τ 2 {1,2} τ 2 {1,3} τ 2 {2,3} + τ 2 {1} + τ 2 {2} + τ 2 {3} A simple (contrast) GSI f(x) λ u f(x u :z u ), u {1,2,3} λ u = ( 1) 3 u [ λ u ] Cost is = 9 function evaluations.

32 Cost is 6. (no duplication) Generalized Sobol Indices 32 A bilinear GSI For u, v w {1, 2, 3} NXOR(u,v+w c ) Ω = λγ T λ γ Coefficients of τ 2 u are in [(123) (13) (12) + (1)] [(23) (3) (2) + ]

33 Generalized Sobol Indices 33 Simple vs. bilinear for d = 5 f(x) ( f(x 1, x 2, x 3, z 4, z 5 ) f(x 1, x 2, z 3, z 4, z 5 ) f(x 1, z 2, x 3, z 4, z 5 ) f(z 1, x 2, x 3, z 4, z 5 ) + f(x 1, z 2, z 3, z 4, z 5 ) + f(z 1, x 2, z 3, z 4, z 5 ) + f(z 1, z 2, x 3, z 4, z 5 ) f(z) ) Versus ( f(z) f(x1, z 2, z 3, z 4, z 5 ) ) ( f(z1, z 2, z 3, x 4, x 5 ) f(z 1, x 2, z 3, x 4, x 5 ) f(z 1, z 2, x 3, x 4, x 5 ) + f(z 1, x 2, x 3, x 4, x 5 ) ) N.B. The bilinear version is invariant under f f + c

34 Generalized Sobol Indices 34 More generally Simple estimator at cost 2 w + 1 w <d σ 2 w = u w( 1) w v Θ u,d Bilinear for w 1 w and w 2 = w w 1 σ 2 w = u 1 w 1 u 2 w 2 ( 1) u1 + u2 Θ u1,u 2 +w c Bilinear cost is 2 w w 2 2 w /2+1. C bilinear 2 C simple

35 Generalized Sobol Indices 35 Superset importance Let w be a nonempty subset of D for d 1. Let f L 2 [0, 1] d. Choose w 1 w and put w 2 = w w 1. Then Υ 2 w = u 1 w 1 compare σ 2 w = u 1 w 1 ( 1) u1 + u2 Θ wc +u 1,w c +u 2 u 2 w 2 ( 1) u1 + u2 Θ u1,w c +u 2. u 2 w 2 Lower cost than a square estimator but probably much higher variance.

36 Generalized Sobol Indices 36 Bilinear, with O(d) evaluations Suppose λ u = 0 for u {0, 1, d 1, d}. Same for γ v = 0. Then the rule λ u γ v u v f(x u :z u )f(x v :z v ) dx dz takes O(d) computation not O(d 2 ).

37 Generalized Sobol Indices 37 O(d) pairs, with k j For j k, let j represent {j} and j represent {j} etc. All the XORs XOR j k j k D j k j k D j j {j, k} D {j, k} j j j D {j, k} {j, k} j D D j k j k

38 Generalized Sobol Indices 38 All the NXORs NXOR j k j k D D j k j k j j D {j, k} {j, k} j j j {j, k} D {j, k} j D j k j k D For u and v in {0, 1, d 1, d}. We can estimate the corresponding τ 2 NXOR(u,v) with O(d) cost per (x, z) pair. Saltelli (2002) already noticed this (or at least most of it).

39 Generalized Sobol Indices 39 What we can get After some algebra we can get unbiased estimates of u σu 2 u u 2 σu 2 u σu 2 u =1 σu 2 u =2 at cost 2d + 2. (Some parts can be gotten at C = d + 1)

40 Generalized Sobol Indices 40 Initial and final segments Suppose that x 1, x 2 x d are used in that order. E.g. time steps in a Markov chain First j variables {1, 2,..., j}, 1 j d (0, j], j = 0 Last d j variables {j + 1,..., d}, 0 j d 1 (j, d] j = d There are 2d + 1 of these subsets.

41 Generalized Sobol Indices 41 Enumeration NXOR (0,j] (0,k] (j,d] (k,d] D D (j, d] (k, d] (0, j] (0, k] (0,j] (j, d] D (j, k] (j, k] (0, j] (j,d] (0, j] (j, k] D (j, k] (j, d] D (0, j] (0, k] (j, d] (k, d] D WLOG j < k.

42 Generalized Sobol Indices 42 Effect of recent variables First and last elements of u : u = min{j j u} u = max{j j u} Recency weighted variance components d 1 ( ) ΘD,(j,d] Θ D, = ( u 1)σu, 2 j=1 u D d 1 ( ) ΘD,(0,j] Θ D, = (d u )σu. 2 j=1 u D and, Another measure of how fast f() forgets its initial conditions. Weighting by u (d u + 1) also possible.

43 Generalized Sobol Indices 43 Test functions g j = 0 f(x) = d (µ j + τ j g j (x j )) j=1 g 2 j = 1 and gj 4 <. σ 2 u = j u τ 2 j j u µ 2 j g(x) = 12(x 1/2) Min function f(x) = min 1 j d x j τ 2 u = u (d + 1) 2 (2d u + 2) Liu and O. (2006)

44 Generalized Sobol Indices 44 σ 2 {1,2,3} Product function = numerically same estimate for simple or bilinear. Therefore bilinear is better because of lower cost. For min(x) and d = 6 the bilinear estimator was about 5 times as efficient as the simple one based on n = 1,000,000 (x, z) pairs. Υ 2 {1,2,3,4} Product function with d = 8 and µ j = 1 and τ = (4, 4, 3, 3, 2, 2, 1, 1)/4. Square beats bilinear: Measure Value R 2 Square s efficiency Υ 2 {1,2,3,4} Υ 2 {5,6,7,8} Hard to beat a sum of squares when the true effect is small.

45 Generalized Sobol Indices 45 Lower index τ 2 u No sum of squares is available. 1 n 1 n Contrast n f(x i )(f(x i,u :z i, u ) f(z i )) i=1 Simple estimator (bias adjusted) n f(x i )f(x i,u :z i, u ) i=1 ( 1 2n n f(x i ) + f(x i,u :z i, u ) i=1 ) 2 The contrast has an advantage on small τ 2 u. The simple estimator sometimes beats it on large ones.

46 Generalized Sobol Indices 46 GSIs so far Just use 2 inputs, x and z What about 3? x, y, z

47 Generalized Sobol Indices 47 Here it pays to use 3 vectors x, y, x [0, 1] d For small τ 2 u 1 n 1 n 1 n 1 n n f(x i ) ( f(x i,u :y i, u ) f(y) ) i=1 (Mauntz-Saltelli) n ( f(xi ) µ )( f(x i,u :y i, u ) f(y) ) (Oracle centered) i=1 n ( f(xi ) µ )( f(x i,u :y i, u ) µ ) (Double oracle) i=1 n ( f(xi ) f(z i,u :x i, u ) )( f(x i,u :y i, u ) f(y) ) (Use 3 vectors) ( ) i=1 where (x i, y i, z i ) iid U[0, 1] 3d for i = 1,..., n. Simulations: On small effects the new estimator beats both oracles. Double oracle wins on large effects.

48 Generalized Sobol Indices 48 Conclusions Sums of squares are very good. Bilinear estimators λ T Θγ work well, especially when 1 T γ = 1 T λ = 0. Further work 1) Pursue variance inequalities 2) Replace plain MC by Quasi-Monte Carlo and/or 3) Find nice confidence intervals for ratios of means over U -statistics 4) Variance reductions

49 Generalized Sobol Indices 49 Thanks Bertrand Iooss for answering questions about Saint-Venant Sergei Kucherenko for extensive discussions National Science Foundation of the U.S. DMS

50 Generalized Sobol Indices 50 Let η 2 = u δ uσ 2 u. Optimal estimates We would like E (ˆη 2 ) = η 2 and, Var (ˆη 2) cost = minimum. Using variance components theory Unfortunately Var(ˆη 2 ) depends on 4 th moments Fortunately There is a theory of MINimum Quadratic Norm UNbiased Estimates (MINQUE) Unfortunately They do not appear to be available for crossed random effects Fortunately The computed case gives us more options, e.g., quadrature. C. R. Rao (1970s)

VARIANCE COMPONENTS AND GENERALIZED SOBOL INDICES. Copright on this work belongs to SIAM.

VARIANCE COMPONENTS AND GENERALIZED SOBOL INDICES. Copright on this work belongs to SIAM. VARIANCE COMPONENTS AND GENERALIZED SOBOL INDICES ART B. OWEN Abstract. This paper introduces generalized Sobol indices, compares strategies for their estimation, and makes a systematic search for efficient

More information

Sobol Indices: an Introduction and. Some Recent Results

Sobol Indices: an Introduction and. Some Recent Results Sobol indices 1 Sobol Indices: an Introduction and Some Recent Results Art B. Owen Stanford University Sobol indices 2 Ilya Meerovich Sobol At MCM 2001, Salzburg Known for Sobol sequences and Sobol indices

More information

VARIANCE COMPONENTS AND GENERALIZED SOBOL INDICES. Art B. Owen. Department of Statistics STANFORD UNIVERSITY Stanford, California

VARIANCE COMPONENTS AND GENERALIZED SOBOL INDICES. Art B. Owen. Department of Statistics STANFORD UNIVERSITY Stanford, California VARIANCE COMPONENTS AND GENERALIZED SOBOL INDICES By Art B. Owen Technical Report No. 2012-07 April 2012 Department of Statistics STANFORD UNIVERSITY Stanford, California 94305-4065 VARIANCE COMPONENTS

More information

On Shapley value for measuring importance. of dependent inputs

On Shapley value for measuring importance. of dependent inputs Shapley value & variable importance 1 On Shapley value for measuring importance of dependent inputs Art B. Owen, Stanford joint with Clémentine Prieur, Grenoble Based on forthcoming JUQ article (2017).

More information

Better estimation of small Sobol sensitivity indices

Better estimation of small Sobol sensitivity indices Better estimation of small Sobol sensitivity indices ART B. OWEN Stanford University September 2012 c ACM, 2012. This is the author s version of the work. It is posted here by permission of ACM for your

More information

Derivative-based global sensitivity measures for interactions

Derivative-based global sensitivity measures for interactions Derivative-based global sensitivity measures for interactions Olivier Roustant École Nationale Supérieure des Mines de Saint-Étienne Joint work with J. Fruth, B. Iooss and S. Kuhnt SAMO 13 Olivier Roustant

More information

Total interaction index: A variance-based sensitivity index for interaction screening

Total interaction index: A variance-based sensitivity index for interaction screening Total interaction index: A variance-based sensitivity index for interaction screening Jana Fruth, Olivier Roustant, Sonja Kuhnt To cite this version: Jana Fruth, Olivier Roustant, Sonja Kuhnt. Total interaction

More information

Total interaction index: A variance-based sensitivity index for second-order interaction screening

Total interaction index: A variance-based sensitivity index for second-order interaction screening Total interaction index: A variance-based sensitivity index for second-order interaction screening J. Fruth a, O. Roustant b, S. Kuhnt a,c a Faculty of Statistics, TU Dortmund University, Vogelpothsweg

More information

Effective dimension of some weighted pre-sobolev spaces with dominating mixed partial derivatives

Effective dimension of some weighted pre-sobolev spaces with dominating mixed partial derivatives Effective dimension of some weighted pre-sobolev spaces with dominating mixed partial derivatives Art B. Owen Stanford University August 2018 Abstract This paper considers two notions of effective dimension

More information

18Ï È² 7( &: ÄuANOVAp.O`û5 571 Based on this ANOVA model representation, Sobol (1993) proposed global sensitivity index, S i1...i s = D i1...i s /D, w

18Ï È² 7( &: ÄuANOVAp.O`û5 571 Based on this ANOVA model representation, Sobol (1993) proposed global sensitivity index, S i1...i s = D i1...i s /D, w A^VÇÚO 1 Êò 18Ï 2013c12 Chinese Journal of Applied Probability and Statistics Vol.29 No.6 Dec. 2013 Optimal Properties of Orthogonal Arrays Based on ANOVA High-Dimensional Model Representation Chen Xueping

More information

Stratified Random Sampling for Dependent Inputs

Stratified Random Sampling for Dependent Inputs Stratified Random Sampling for Dependent Inputs Anirban Mondal Case Western Reserve University, Cleveland, OH 44106, USA arxiv:1904.00555v1 [stat.me] 1 Apr 2019 Abhijit Mandal Wayne State University, Detroit,

More information

Estimating mean dimensionality of ANOVA. decompositions

Estimating mean dimensionality of ANOVA. decompositions Estimating mean dimensionality of ANOVA decompositions Ruixue Liu 1 and Art B. Owen 2 Department of Statistics Stanford University Stanford CA, 94305 Orig: June 2003. Revised: March 2005 Abstract The analysis

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

Monte Carlo Methods. Geoff Gordon February 9, 2006

Monte Carlo Methods. Geoff Gordon February 9, 2006 Monte Carlo Methods Geoff Gordon ggordon@cs.cmu.edu February 9, 2006 Numerical integration problem 5 4 3 f(x,y) 2 1 1 0 0.5 0 X 0.5 1 1 0.8 0.6 0.4 Y 0.2 0 0.2 0.4 0.6 0.8 1 x X f(x)dx Used for: function

More information

SCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES. Kinjal Basu Art B. Owen

SCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES. Kinjal Basu Art B. Owen SCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES By Kinjal Basu Art B. Owen Technical Report No. 2015-07 March 2015 Department of Statistics STANFORD UNIVERSITY Stanford, California 94305-4065

More information

Sobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis

Sobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis Sobol-Hoeffding decomposition Application to Global SA Computation of the SI Sobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis Olivier Le Maître with Colleague & Friend Omar

More information

A comparison of global sensitivity techniques and sampling method

A comparison of global sensitivity techniques and sampling method 22nd International Congress on Modelling and Simulation, Hobart, Tasmania, Australia, 3 to 8 December 2017 mssanz.org.au/modsim2017 A comparison of global sensitivity techniques and sampling method X.

More information

Multi-fidelity sensitivity analysis

Multi-fidelity sensitivity analysis Multi-fidelity sensitivity analysis Loic Le Gratiet 1,2, Claire Cannamela 2 & Bertrand Iooss 3 1 Université Denis-Diderot, Paris, France 2 CEA, DAM, DIF, F-91297 Arpajon, France 3 EDF R&D, 6 quai Watier,

More information

Wavelets For Computer Graphics

Wavelets For Computer Graphics {f g} := f(x) g(x) dx A collection of linearly independent functions Ψ j spanning W j are called wavelets. i J(x) := 6 x +2 x + x + x Ψ j (x) := Ψ j (2 j x i) i =,..., 2 j Res. Avge. Detail Coef 4 [9 7

More information

Minimizing the Number of Function Evaluations to Estimate Sobol Indices Using Quasi-Monte Carlo

Minimizing the Number of Function Evaluations to Estimate Sobol Indices Using Quasi-Monte Carlo Minimizing the Number of Function Evaluations to Estimate Sobol Indices Using Quasi-Monte Carlo Lluís Antoni Jiménez Rugama Joint work with: Fred J Hickernell (IIT), Clémentine Prieur (Univ Grenoble),

More information

Quasi-Monte Carlo integration over the Euclidean space and applications

Quasi-Monte Carlo integration over the Euclidean space and applications Quasi-Monte Carlo integration over the Euclidean space and applications f.kuo@unsw.edu.au University of New South Wales, Sydney, Australia joint work with James Nichols (UNSW) Journal of Complexity 30

More information

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth

More information

Brief introduction to Markov Chain Monte Carlo

Brief introduction to Markov Chain Monte Carlo Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical

More information

Accuracy, Precision and Efficiency in Sparse Grids

Accuracy, Precision and Efficiency in Sparse Grids John, Information Technology Department, Virginia Tech.... http://people.sc.fsu.edu/ jburkardt/presentations/ sandia 2009.pdf... Computer Science Research Institute, Sandia National Laboratory, 23 July

More information

Analysis of Algorithms I: Perfect Hashing

Analysis of Algorithms I: Perfect Hashing Analysis of Algorithms I: Perfect Hashing Xi Chen Columbia University Goal: Let U = {0, 1,..., p 1} be a huge universe set. Given a static subset V U of n keys (here static means we will never change the

More information

A GENERAL THEORY FOR ORTHOGONAL ARRAY BASED LATIN HYPERCUBE SAMPLING

A GENERAL THEORY FOR ORTHOGONAL ARRAY BASED LATIN HYPERCUBE SAMPLING Statistica Sinica 26 (2016), 761-777 doi:http://dx.doi.org/10.5705/ss.202015.0029 A GENERAL THEORY FOR ORTHOGONAL ARRAY BASED LATIN HYPERCUBE SAMPLING Mingyao Ai, Xiangshun Kong and Kang Li Peking University

More information

Generalized Sobol indices for dependent variables

Generalized Sobol indices for dependent variables Generalized Sobol indices for dependent variables Gaelle Chastaing Fabrice Gamboa Clémentine Prieur July 1st, 2013 1/28 Gaelle Chastaing Sensitivity analysis and dependent variables Contents 1 Context

More information

Tutorial on quasi-monte Carlo methods

Tutorial on quasi-monte Carlo methods Tutorial on quasi-monte Carlo methods Josef Dick School of Mathematics and Statistics, UNSW, Sydney, Australia josef.dick@unsw.edu.au Comparison: MCMC, MC, QMC Roughly speaking: Markov chain Monte Carlo

More information

The square root rule for adaptive importance sampling

The square root rule for adaptive importance sampling The square root rule for adaptive importance sampling Art B. Owen Stanford University Yi Zhou January 2019 Abstract In adaptive importance sampling, and other contexts, we have unbiased and uncorrelated

More information

Sampling from complex probability distributions

Sampling from complex probability distributions Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation

More information

Pascal s Triangle on a Budget. Accuracy, Precision and Efficiency in Sparse Grids

Pascal s Triangle on a Budget. Accuracy, Precision and Efficiency in Sparse Grids Covering : Accuracy, Precision and Efficiency in Sparse Grids https://people.sc.fsu.edu/ jburkardt/presentations/auburn 2009.pdf... John Interdisciplinary Center for Applied Mathematics & Information Technology

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

Mathematics Department Stanford University Math 61CM/DM Inner products

Mathematics Department Stanford University Math 61CM/DM Inner products Mathematics Department Stanford University Math 61CM/DM Inner products Recall the definition of an inner product space; see Appendix A.8 of the textbook. Definition 1 An inner product space V is a vector

More information

A General Overview of Parametric Estimation and Inference Techniques.

A General Overview of Parametric Estimation and Inference Techniques. A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying

More information

Variational sampling approaches to word confusability

Variational sampling approaches to word confusability Variational sampling approaches to word confusability John R. Hershey, Peder A. Olsen and Ramesh A. Gopinath {jrhershe,pederao,rameshg}@us.ibm.com IBM, T. J. Watson Research Center Information Theory and

More information

Spanning and Independence Properties of Finite Frames

Spanning and Independence Properties of Finite Frames Chapter 1 Spanning and Independence Properties of Finite Frames Peter G. Casazza and Darrin Speegle Abstract The fundamental notion of frame theory is redundancy. It is this property which makes frames

More information

A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS

A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS Statistica Sinica 26 (2016), 1117-1128 doi:http://dx.doi.org/10.5705/ss.202015.0240 A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS Xu He and Peter Z. G. Qian Chinese Academy of Sciences

More information

BOUNDARY VALUE PROBLEMS ON A HALF SIERPINSKI GASKET

BOUNDARY VALUE PROBLEMS ON A HALF SIERPINSKI GASKET BOUNDARY VALUE PROBLEMS ON A HALF SIERPINSKI GASKET WEILIN LI AND ROBERT S. STRICHARTZ Abstract. We study boundary value problems for the Laplacian on a domain Ω consisting of the left half of the Sierpinski

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization

CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization Tim Roughgarden February 28, 2017 1 Preamble This lecture fulfills a promise made back in Lecture #1,

More information

P (A G) dp G P (A G)

P (A G) dp G P (A G) First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume

More information

MATH 12 CLASS 23 NOTES, NOV Contents 1. Change of variables: the Jacobian 1

MATH 12 CLASS 23 NOTES, NOV Contents 1. Change of variables: the Jacobian 1 MATH 12 CLASS 23 NOTES, NOV 11 211 Contents 1. Change of variables: the Jacobian 1 1. Change of variables: the Jacobian So far, we have seen three examples of situations where we change variables to help

More information

Gaussian process regression for Sensitivity analysis

Gaussian process regression for Sensitivity analysis Gaussian process regression for Sensitivity analysis GPSS Workshop on UQ, Sheffield, September 2016 Nicolas Durrande, Mines St-Étienne, durrande@emse.fr GPSS workshop on UQ GPs for sensitivity analysis

More information

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with

More information

Stochastic Subgradient Methods

Stochastic Subgradient Methods Stochastic Subgradient Methods Stephen Boyd and Almir Mutapcic Notes for EE364b, Stanford University, Winter 26-7 April 13, 28 1 Noisy unbiased subgradient Suppose f : R n R is a convex function. We say

More information

STA 294: Stochastic Processes & Bayesian Nonparametrics

STA 294: Stochastic Processes & Bayesian Nonparametrics MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Vahid Dehdari and Clayton V. Deutsch Geostatistical modeling involves many variables and many locations.

More information

CHAPTER 3 Further properties of splines and B-splines

CHAPTER 3 Further properties of splines and B-splines CHAPTER 3 Further properties of splines and B-splines In Chapter 2 we established some of the most elementary properties of B-splines. In this chapter our focus is on the question What kind of functions

More information

4th Preparation Sheet - Solutions

4th Preparation Sheet - Solutions Prof. Dr. Rainer Dahlhaus Probability Theory Summer term 017 4th Preparation Sheet - Solutions Remark: Throughout the exercise sheet we use the two equivalent definitions of separability of a metric space

More information

Isomorphisms between pattern classes

Isomorphisms between pattern classes Journal of Combinatorics olume 0, Number 0, 1 8, 0000 Isomorphisms between pattern classes M. H. Albert, M. D. Atkinson and Anders Claesson Isomorphisms φ : A B between pattern classes are considered.

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Eco517 Fall 2004 C. Sims MIDTERM EXAM

Eco517 Fall 2004 C. Sims MIDTERM EXAM Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering

More information

RECYCLING PHYSICAL RANDOM NUMBERS. Art B. Owen. Technical Report No September 2009

RECYCLING PHYSICAL RANDOM NUMBERS. Art B. Owen. Technical Report No September 2009 RECYCLING PHYSICAL RANDOM NUMBERS By Art B. Owen Technical Report No. 2009-11 September 2009 Department of Statistics STANFORD UNIVERSITY Stanford, California 94305-4065 RECYCLING PHYSICAL RANDOM NUMBERS

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Recall: Dot product on R 2 : u v = (u 1, u 2 ) (v 1, v 2 ) = u 1 v 1 + u 2 v 2, u u = u u 2 2 = u 2. Geometric Meaning:

Recall: Dot product on R 2 : u v = (u 1, u 2 ) (v 1, v 2 ) = u 1 v 1 + u 2 v 2, u u = u u 2 2 = u 2. Geometric Meaning: Recall: Dot product on R 2 : u v = (u 1, u 2 ) (v 1, v 2 ) = u 1 v 1 + u 2 v 2, u u = u 2 1 + u 2 2 = u 2. Geometric Meaning: u v = u v cos θ. u θ v 1 Reason: The opposite side is given by u v. u v 2 =

More information

Part III. Quasi Monte Carlo methods 146/349

Part III. Quasi Monte Carlo methods 146/349 Part III Quasi Monte Carlo methods 46/349 Outline Quasi Monte Carlo methods 47/349 Quasi Monte Carlo methods Let y = (y,...,y N ) be a vector of independent uniform random variables in Γ = [0,] N, u(y)

More information

2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure? MA 645-4A (Real Analysis), Dr. Chernov Homework assignment 1 (Due 9/5). Prove that every countable set A is measurable and µ(a) = 0. 2 (Bonus). Let A consist of points (x, y) such that either x or y is

More information

BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1

BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1 BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1 Shaobo Li University of Cincinnati 1 Partially based on Hastie, et al. (2009) ESL, and James, et al. (2013) ISLR Data Mining I Lecture

More information

Tutorials in Optimization. Richard Socher

Tutorials in Optimization. Richard Socher Tutorials in Optimization Richard Socher July 20, 2008 CONTENTS 1 Contents 1 Linear Algebra: Bilinear Form - A Simple Optimization Problem 2 1.1 Definitions........................................ 2 1.2

More information

Asymptotics of minimax stochastic programs

Asymptotics of minimax stochastic programs Asymptotics of minimax stochastic programs Alexander Shapiro Abstract. We discuss in this paper asymptotics of the sample average approximation (SAA) of the optimal value of a minimax stochastic programming

More information

The Hilbert Space of Random Variables

The Hilbert Space of Random Variables The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2

More information

SCALE INVARIANT FOURIER RESTRICTION TO A HYPERBOLIC SURFACE

SCALE INVARIANT FOURIER RESTRICTION TO A HYPERBOLIC SURFACE SCALE INVARIANT FOURIER RESTRICTION TO A HYPERBOLIC SURFACE BETSY STOVALL Abstract. This result sharpens the bilinear to linear deduction of Lee and Vargas for extension estimates on the hyperbolic paraboloid

More information

MATH Linear Algebra

MATH Linear Algebra MATH 4 - Linear Algebra One of the key driving forces for the development of linear algebra was the profound idea (going back to 7th century and the work of Pierre Fermat and Rene Descartes) that geometry

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 1 Bootstrapped Bias and CIs Given a multiple regression model with mean and

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 6: Bias and variance (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 49 Our plan today We saw in last lecture that model scoring methods seem to be trading off two different

More information

LAGRANGE MULTIPLIERS

LAGRANGE MULTIPLIERS LAGRANGE MULTIPLIERS MATH 195, SECTION 59 (VIPUL NAIK) Corresponding material in the book: Section 14.8 What students should definitely get: The Lagrange multiplier condition (one constraint, two constraints

More information

CSE 312: Foundations of Computing II Quiz Section #10: Review Questions for Final Exam (solutions)

CSE 312: Foundations of Computing II Quiz Section #10: Review Questions for Final Exam (solutions) CSE 312: Foundations of Computing II Quiz Section #10: Review Questions for Final Exam (solutions) 1. (Confidence Intervals, CLT) Let X 1,..., X n be iid with unknown mean θ and known variance σ 2. Assume

More information

Functions of Several Variables

Functions of Several Variables Jim Lambers MAT 419/519 Summer Session 2011-12 Lecture 2 Notes These notes correspond to Section 1.2 in the text. Functions of Several Variables We now generalize the results from the previous section,

More information

Polynomial chaos expansions for sensitivity analysis

Polynomial chaos expansions for sensitivity analysis c DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for sensitivity analysis B. Sudret Chair of Risk, Safety & Uncertainty

More information

Analysis II - few selective results

Analysis II - few selective results Analysis II - few selective results Michael Ruzhansky December 15, 2008 1 Analysis on the real line 1.1 Chapter: Functions continuous on a closed interval 1.1.1 Intermediate Value Theorem (IVT) Theorem

More information

Linear Models Review

Linear Models Review Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign

More information

A Few Notes on Fisher Information (WIP)

A Few Notes on Fisher Information (WIP) A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties

More information

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR 1. Definition Existence Theorem 1. Assume that A R m n. Then there exist orthogonal matrices U R m m V R n n, values σ 1 σ 2... σ p 0 with p = min{m, n},

More information

IEOR 165 Lecture 7 1 Bias-Variance Tradeoff

IEOR 165 Lecture 7 1 Bias-Variance Tradeoff IEOR 165 Lecture 7 Bias-Variance Tradeoff 1 Bias-Variance Tradeoff Consider the case of parametric regression with β R, and suppose we would like to analyze the error of the estimate ˆβ in comparison to

More information

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define HILBERT SPACES AND THE RADON-NIKODYM THEOREM STEVEN P. LALLEY 1. DEFINITIONS Definition 1. A real inner product space is a real vector space V together with a symmetric, bilinear, positive-definite mapping,

More information

Discrete Latent Variable Models

Discrete Latent Variable Models Discrete Latent Variable Models Stefano Ermon, Aditya Grover Stanford University Lecture 14 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 14 1 / 29 Summary Major themes in the course

More information

ECE 275A Homework 7 Solutions

ECE 275A Homework 7 Solutions ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten

More information

Poincaré inequalities revisited for dimension reduction

Poincaré inequalities revisited for dimension reduction Poincaré inequalities revisited for dimension reduction Olivier Roustant*, Franck Barthe** and Bertrand Iooss**, *** * Mines Saint-Étienne ** Institut de Mathématiques de Toulouse *** EDF, Chatou March

More information

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the

More information

Evaluating the Performance of Estimators (Section 7.3)

Evaluating the Performance of Estimators (Section 7.3) Evaluating the Performance of Estimators (Section 7.3) Example: Suppose we observe X 1,..., X n iid N(θ, σ 2 0 ), with σ2 0 known, and wish to estimate θ. Two possible estimators are: ˆθ = X sample mean

More information

1 Structures 2. 2 Framework of Riemann surfaces Basic configuration Holomorphic functions... 3

1 Structures 2. 2 Framework of Riemann surfaces Basic configuration Holomorphic functions... 3 Compact course notes Riemann surfaces Fall 2011 Professor: S. Lvovski transcribed by: J. Lazovskis Independent University of Moscow December 23, 2011 Contents 1 Structures 2 2 Framework of Riemann surfaces

More information

Stat 451 Lecture Notes Numerical Integration

Stat 451 Lecture Notes Numerical Integration Stat 451 Lecture Notes 03 12 Numerical Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 5 in Givens & Hoeting, and Chapters 4 & 18 of Lange 2 Updated: February 11, 2016 1 / 29

More information

Penalty Methods for Bivariate Smoothing and Chicago Land Values

Penalty Methods for Bivariate Smoothing and Chicago Land Values Penalty Methods for Bivariate Smoothing and Chicago Land Values Roger Koenker University of Illinois, Urbana-Champaign Ivan Mizera University of Alberta, Edmonton Northwestern University: October 2001

More information

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count

More information

Eigenvectors and Hermitian Operators

Eigenvectors and Hermitian Operators 7 71 Eigenvalues and Eigenvectors Basic Definitions Let L be a linear operator on some given vector space V A scalar λ and a nonzero vector v are referred to, respectively, as an eigenvalue and corresponding

More information

Exercises on chapter 0

Exercises on chapter 0 Exercises on chapter 0 1. A partially ordered set (poset) is a set X together with a relation such that (a) x x for all x X; (b) x y and y x implies that x = y for all x, y X; (c) x y and y z implies that

More information

Lecture 6 Proof for JL Lemma and Linear Dimensionality Reduction

Lecture 6 Proof for JL Lemma and Linear Dimensionality Reduction COMS 4995: Unsupervised Learning (Summer 18) June 7, 018 Lecture 6 Proof for JL Lemma and Linear imensionality Reduction Instructor: Nakul Verma Scribes: Ziyuan Zhong, Kirsten Blancato This lecture gives

More information

Numerical integration and importance sampling

Numerical integration and importance sampling 2 Numerical integration and importance sampling 2.1 Quadrature Consider the numerical evaluation of the integral I(a, b) = b a dx f(x) Rectangle rule: on small interval, construct interpolating function

More information

9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career.

9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career. Introduction to Data and Analysis Wildlife Management is a very quantitative field of study Results from studies will be used throughout this course and throughout your career. Sampling design influences

More information

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection

More information

Gaussian Processes for Computer Experiments

Gaussian Processes for Computer Experiments Gaussian Processes for Computer Experiments Jeremy Oakley School of Mathematics and Statistics, University of Sheffield www.jeremy-oakley.staff.shef.ac.uk 1 / 43 Computer models Computer model represented

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

Math 2 Variable Manipulation Part 7 Absolute Value & Inequalities

Math 2 Variable Manipulation Part 7 Absolute Value & Inequalities Math 2 Variable Manipulation Part 7 Absolute Value & Inequalities 1 MATH 1 REVIEW SOLVING AN ABSOLUTE VALUE EQUATION Absolute value is a measure of distance; how far a number is from zero. In practice,

More information