New results for the functional ANOVA and. Sobol indices
|
|
- Camron Dean
- 5 years ago
- Views:
Transcription
1 Generalized Sobol Indices 1 New results for the functional ANOVA and Sobol indices Art B. Owen Stanford University
2 Generalized Sobol Indices 2 Ilya Meerovich Sobol At MCM 2001, Salzburg Known for Sobol sequences and Sobol indices Every time I read one of his papers, I wish I d read it earlier
3 Generalized Sobol Indices 3 Simplified Saint-Venant flood model Overflow in meters (Lamboni, Iooss, Popelin, Gamboa, 2012) at a dyke S = Z v + H H d C b, ( Q H = BK s (Zm Z v )/L where ) 3/5 (max annual river height) Q Maximal annual flow m 3 /s Gumbel(1013, 558) [500, 3000] K s Strickler coefficient m 1/3 /s N (30, 8) [15, ) Z v River downstream level m Triangle(49, 50, 51) Z m River upstream level m Triangle(54, 55, 56) H d Dyke height m U[7, 9] C b Bank level m Triangle(55, 55.5, 56) L Length of river stretch m Triangle(4990, 5000, 5010) B River width m Triangle(295, 300, 305) Reduced from a Navier-Stokes model
4 Generalized Sobol Indices 4 The cost model C p = 1 S>0m (flood cost) + 1 S 0m ( (1 e 1000m4 /S 4 )) (dyke maintenance) min(h d m 1, 8) (investment cost, from construction) in millions of Euros A discontinuity at S = 0 would be better. Quibble
5 Generalized Sobol Indices 5 Variable importance Which of these variables is most important? How important is any given subset of variables taken together? How should one define/measure variable importance? Similar problems come up in black box models for aerospace, semiconductor manufacturing, climate modeling, etc.
6 Generalized Sobol Indices 6 Why do they care? The ultimate goal may be to optimize something (a mean or variance or max or min) by choosing levels for those variables under your control. Understanding a function is an intermediate goal, not an ultimate goal. Yet variable importance is widely applied. Similar question Why compute R 2? Sobol indices yield many analogues of R 2. Quantitative measures of importance aid discussions.
7 Generalized Sobol Indices 7 Quasi-Monte Carlo QMC sometimes gives accurate estimates for high dimensional integrals. Despite the curse of dimensionality. When it happens, we usually find that the integrand was dominated by low order interactions among the variables. That understanding motivates trying to lower the effective dimension of our integrands. (Analogous to variance reduction.) which never actually said that all high dim problems are insoluble.
8 Generalized Sobol Indices 8 ANOVA Fisher (1923) to Hoeffding (1948) to Sobol (1967) to Efron & Stein (1981) then back to Sobol (1990/3) Recall X ij = X + ( X i X ) + ( X j X ) + (X ij X i X j + X ) Take a grand average X Subtract it out and average over j, getting X i X Similarly get X j X Generally: subtract away all lower order effects and average out the extra variables.
9 Generalized Sobol Indices 9 Extensions of ANOVA From functions on {1, 2,..., I} {1, 2,..., J} to {1, 2,..., I} {1, 2,..., J} {1, 2,..., Z} To functions in L 2 [0, 1] d. To L 2 functions of d arbitrary independent inputs. Also d = works via martingales.
10 Generalized Sobol Indices 10 ANOVA for L 2 [0, 1] d Goes back to Hoeffding (1948) for U -statistics (skillful reading may be required) & Efron & Stein (1981) for jackknife f(x) = f () () + = f () () + d f (j) (x j ) + f (j,k) (x j, x k ) + + f (1,2,...,d) (x 1,..., x d ) j=1 j<k d f (j1,j 2,...,j d )(x j1, x j2,..., x jd ) r=1 1 j 1 <j 2 < <j r d More simply f(x) = u f u (x) Sum over all u D = {1, 2,..., d}
11 Generalized Sobol Indices 11 Notation For u D {1,..., d} u = card(u) u = u c = {1, 2,..., d} u v u strict subset i.e. If u = {j 1, j 2,..., j u } then x u = (x j1,..., x j u ) and dx u = j u dx j
12 Generalized Sobol Indices 12 Recursive definition For u {1,..., d}, f u (x) only depends on x j for j u. Overall mean µ f (x) = f(x) dx (f(x) Main effect j f {j} (x) = f (x) ) dx {j} (f(x) Interaction u f u (x) = f v (x) ) dx u = Dependence v u f(x) dx u v u f v (x) f u (x) is a function of x that happens to only depend on x u f u (x) + f v (x) makes sense because they re on the same domain
13 Generalized Sobol Indices 13 ANOVA properties j u = u v = & 1 f u (x) dx j = 0 0 f u (x)f v (x) dx = 0 f u (x)g v (x) dx = 0 Variances Var(f) (f(x) µ) 2 dx = u {1,...,d} σu 2 = σu(f) 2 fu (x) 2 dx u = 0 u =. σ 2 u
14 Generalized Sobol Indices 14 Sobol s decomposition Let φ 0, φ 1, φ 2... be a complete orthonormal basis of L 2 [0, 1] with φ 0 (x) 1. Sobol (1969) expanded f(x) in a tensor product basis (Haar wavelets). He grouped the terms into 2 d subsets depending on which inputs are active. Sobol has a synthesis not an analysis for this decomposition. Thanks to A. Chouldechova for translation.
15 Generalized Sobol Indices 15 How important is x u? Variable importance Larger σ 2 u means that f u(x) contributes more. We also want to count σ 2 v for v u. Sobol s importance measures τ 2 u = v u σ 2 v v contained in u τ 2 u = v u σ 2 v v touches u, so interactions count Identity: τ 2 u = σ 2 τ 2 u Normalized versions: τ 2 u σ 2 and τ 2 u σ 2
16 Generalized Sobol Indices 16 More derived importance measures Superset importance Υ 2 u = v u σ 2 v Liu & O (2006) Small Υ 2 u means deleting f u and f v for v u (to stay hierarchical) makes little difference. Relevant to Hooker (2004) s simplifications of black box functions. Mean dimension u D σ 2 u σ 2 u Measures dimensionality of f. Liu & O (2006) Higher dimensionality makes for harder numerical handling. Many quadrature problems have mean dimension near 1
17 Generalized Sobol Indices 17 Estimation of τ 2 u and τ 2 u Naive approach for τ 2 u : 1) Sample x i [0, 1] d and get y i = f(x i ) for i = 1,..., n. 2) Statistical machine learning estimate ˆf v (x) for all necessary v. 3) Put ˆσ v 2 = ˆfu (x) 2 dx, u. 4) Sum: τ 2 u = v u ˆσ2 v. This is expensive and has many biases. Sobol has a much better way.
18 Generalized Sobol Indices 18 Evaluate f at two points: repeat some components independent draws for the others. Fixing methods Hybrid points For x, z [0, 1] d, y = x u :z u means x j, j u y j = z j, j u. We glue together part of x and part of z to form y = x u :z u. Sobol (1990/3) used the identities: τ 2 u = Cov(f(x), f(x u :z u )) τ 2 u = 1 2 E((f(x) f(x u:z u )) 2
19 Generalized Sobol Indices 19 Identity for τ 2 u f(x)f(x u :z u ) dx dz = f v (x)f v (x u :z u ) dx dz v D (orthogonality) = v u f v (x)f v (x u :z u ) dx dz (line integrals) = µ 2 + v u σ 2 u µ 2 + τ 2 u. τ 2 u = 1 n i=1 Bias adjustment n ( 1 f(x i )f(x i,u :z i, u ) n n i=1 ) 2 f(x i )
20 Generalized Sobol Indices 20 τ 2 u = Even better f(x) ( f(x u :z u ) f(z) ) dx dz τ 2 u = 1 n n f(x i ) ( f(x i,u :z i, u ) f(z i ) ) i=1 This avoids subtracting ˆµ 2. It is unbiased: E ( τ 2 u ) = τ 2 u Kucherenko, Feil, Shah, Mauntz (2011), Mauntz (2002), Saltelli (2002) τ 2 u = 1 n Improved statistical efficiency n f(x i )f(x i,u :z i, u ) i=1 Janon, Klein, Lagnoux, Nodet & Prieur (2012) ( 1 n n i=1 f(x i ) + f(x i,u :z i, u ) 2 Efficient in a class of estimators that does not include the unbiased one above. ) 2
21 Generalized Sobol Indices 21 For τ 2 u 1 2 (f(x) f(x u:z u ) ) 2 dx dz = 1 2 (σ 2 + µ 2 2 ( τ 2 u + µ2) + σ 2 + µ 2) = σ 2 τ 2 u = τ 2 u. Sobol s estimates are like tomography: integrals reveal internal structure.
22 Generalized Sobol Indices 22 τ 2 {j} for the flood model τ 2 /σ 2 Q K s Z v Z m H d C b L B Height H Overflow S Cost C p From n = 100,000 runs Q Maximal annual flow m 3 /s Gumbel(1013, 558) [500, 3000] K s Strickler coefficient m 1/3 /s N (30, 8) [15, ) Z v River downstream level m Triangle(49, 50, 51) Z m River upstream level m Triangle(54, 55, 56) H d Dyke height m U[7, 9] C b Bank level m Triangle(55, 55.5, 56) L Length of river stretch m Triangle(4990, 5000, 5010) B River width m Triangle(295, 300, 305)
23 Generalized Sobol Indices 23 For mean dimension d τ 2 j = j=1 d j=1 v {j} = d σ 2 v = v σ 2 v j=1 1 v {j} = = v v σ 2 v Estimator from Liu & O (2006) Generalizes to v k σv 2 for k 1. v
24 Generalized Sobol Indices 24 Example Kuo, Schwab, Sloan (2012) consider quadrature for f(x) = d j=1 xα j /j!, 0 < α 1. For α = 1 and d = 500 R = 50 replicated estimates of v v σ2 v/σ 2 using n = 10,000 had mean and standard deviation Upshot f(x) is nearly additive, though it is hard to quantify near perfect additivity. (The difficulty seems to be in forming the ratio.)
25 Generalized Sobol Indices 25 For superset importance Υ 2 u v u σ 2 v = 1 2 u ( 2 ( 1) u v f(x v :z v )) dx dz v u Mean of a square of differences better than differences of means of squares. From Liu & O (2006) Generalizes τ 2 u formula from 2 terms to 2 u terms. As a design Use n repeats of a 2 u 1 d u factorial randomly embedded in the unit cube. Does best in comparisons by Fruth, Roustant, Kuhnt (2012)
26 Generalized Sobol Indices 26 Generalized Sobol indices What can be attained via fixing methods? Θ uv = f(x u :z u )f(x v :z v ) dx dz Generalized Sobol index Ω uv Θ uv = tr(ω T Θ) u D v D Θ R 2d 2 d is the Sobol matrix. Ω R 2d 2 d has coefficients. Redundant (but useful) We have a 2 2d dimensional space of estimators for a 2 d dimensional space of estimands: δ µ 2 + δ u σu 2 u >0
27 Generalized Sobol Indices 27 NXOR XOR(u, v) = u v u v NXOR(u, v) = XOR(u, v) c = (u v) (u c v c ) (exclusive OR) (not exclusive OR) Θ uv f(x u :z u )f(x v :z v ) dx dz = µ 2 + τ 2 NXOR(u,v) Θ uv = 1 n f(x i,u :z i, u )f(x i,v :z i, v ) n i=1 Use tr(ω T Θ) often written XNOR
28 Generalized Sobol Indices 28 Special GSIs 1) Mean squares Ω = λλ T ( 2 λ u f(x u :z u )) dx dz Nonnegative u 2) Bilinear (rank one) Ω = λγ T ( )( λ u f(x u :z u ) u v ) γ v f(x v :z v ) dx dz Fast 3) Simple ( u ) λ u f(x u :z u ) f(z) dx dz Only uses one row/col of Θ 4) Contrast Ω u,v = 0 Free of µ 2 u v N.B.: Here a contrast can also be a sum of squares.
29 Generalized Sobol Indices 29 Cost of a GSI C(Ω) counts the # of function evaluations per (x, z) pair. Recipe 1) Count the rows u that are needed for some f(x u :z u ) 2) add the columns (where u appears as the needed v ) 3) subtract any doubly counted subsets We can have tr(ω T 1 Θ) = tr(ω T 2 Θ) but C(Ω 1 ) < C(Ω 2 ).
30 Generalized Sobol Indices 30 Squares For a square (or a sum of squares) tr(ω T Θ) 0. Also E ( tr(ω T Θ) ) = tr(ω T Θ) Therefore tr(ω T Θ) = 0 implies Pr ( tr(ω T Θ) = 0 ) = 1. GSIs with sum of squares estimators τ 2 u and Υ 2 u and No sum of squares exists for τ 2 u Can show that the coefficient of σ 2 D = u λ2 u generally u Ω uu i.e., tr(ω) Same thing happens in ANOVA tables: u σu 2 u when u < d every variance component has a contribution from the measurement error.
31 Generalized Sobol Indices 31 Targeting one variance component σ 2 {1,2,3} = τ 2 {1,2,3} τ 2 {1,2} τ 2 {1,3} τ 2 {2,3} + τ 2 {1} + τ 2 {2} + τ 2 {3} A simple (contrast) GSI f(x) λ u f(x u :z u ), u {1,2,3} λ u = ( 1) 3 u [ λ u ] Cost is = 9 function evaluations.
32 Cost is 6. (no duplication) Generalized Sobol Indices 32 A bilinear GSI For u, v w {1, 2, 3} NXOR(u,v+w c ) Ω = λγ T λ γ Coefficients of τ 2 u are in [(123) (13) (12) + (1)] [(23) (3) (2) + ]
33 Generalized Sobol Indices 33 Simple vs. bilinear for d = 5 f(x) ( f(x 1, x 2, x 3, z 4, z 5 ) f(x 1, x 2, z 3, z 4, z 5 ) f(x 1, z 2, x 3, z 4, z 5 ) f(z 1, x 2, x 3, z 4, z 5 ) + f(x 1, z 2, z 3, z 4, z 5 ) + f(z 1, x 2, z 3, z 4, z 5 ) + f(z 1, z 2, x 3, z 4, z 5 ) f(z) ) Versus ( f(z) f(x1, z 2, z 3, z 4, z 5 ) ) ( f(z1, z 2, z 3, x 4, x 5 ) f(z 1, x 2, z 3, x 4, x 5 ) f(z 1, z 2, x 3, x 4, x 5 ) + f(z 1, x 2, x 3, x 4, x 5 ) ) N.B. The bilinear version is invariant under f f + c
34 Generalized Sobol Indices 34 More generally Simple estimator at cost 2 w + 1 w <d σ 2 w = u w( 1) w v Θ u,d Bilinear for w 1 w and w 2 = w w 1 σ 2 w = u 1 w 1 u 2 w 2 ( 1) u1 + u2 Θ u1,u 2 +w c Bilinear cost is 2 w w 2 2 w /2+1. C bilinear 2 C simple
35 Generalized Sobol Indices 35 Superset importance Let w be a nonempty subset of D for d 1. Let f L 2 [0, 1] d. Choose w 1 w and put w 2 = w w 1. Then Υ 2 w = u 1 w 1 compare σ 2 w = u 1 w 1 ( 1) u1 + u2 Θ wc +u 1,w c +u 2 u 2 w 2 ( 1) u1 + u2 Θ u1,w c +u 2. u 2 w 2 Lower cost than a square estimator but probably much higher variance.
36 Generalized Sobol Indices 36 Bilinear, with O(d) evaluations Suppose λ u = 0 for u {0, 1, d 1, d}. Same for γ v = 0. Then the rule λ u γ v u v f(x u :z u )f(x v :z v ) dx dz takes O(d) computation not O(d 2 ).
37 Generalized Sobol Indices 37 O(d) pairs, with k j For j k, let j represent {j} and j represent {j} etc. All the XORs XOR j k j k D j k j k D j j {j, k} D {j, k} j j j D {j, k} {j, k} j D D j k j k
38 Generalized Sobol Indices 38 All the NXORs NXOR j k j k D D j k j k j j D {j, k} {j, k} j j j {j, k} D {j, k} j D j k j k D For u and v in {0, 1, d 1, d}. We can estimate the corresponding τ 2 NXOR(u,v) with O(d) cost per (x, z) pair. Saltelli (2002) already noticed this (or at least most of it).
39 Generalized Sobol Indices 39 What we can get After some algebra we can get unbiased estimates of u σu 2 u u 2 σu 2 u σu 2 u =1 σu 2 u =2 at cost 2d + 2. (Some parts can be gotten at C = d + 1)
40 Generalized Sobol Indices 40 Initial and final segments Suppose that x 1, x 2 x d are used in that order. E.g. time steps in a Markov chain First j variables {1, 2,..., j}, 1 j d (0, j], j = 0 Last d j variables {j + 1,..., d}, 0 j d 1 (j, d] j = d There are 2d + 1 of these subsets.
41 Generalized Sobol Indices 41 Enumeration NXOR (0,j] (0,k] (j,d] (k,d] D D (j, d] (k, d] (0, j] (0, k] (0,j] (j, d] D (j, k] (j, k] (0, j] (j,d] (0, j] (j, k] D (j, k] (j, d] D (0, j] (0, k] (j, d] (k, d] D WLOG j < k.
42 Generalized Sobol Indices 42 Effect of recent variables First and last elements of u : u = min{j j u} u = max{j j u} Recency weighted variance components d 1 ( ) ΘD,(j,d] Θ D, = ( u 1)σu, 2 j=1 u D d 1 ( ) ΘD,(0,j] Θ D, = (d u )σu. 2 j=1 u D and, Another measure of how fast f() forgets its initial conditions. Weighting by u (d u + 1) also possible.
43 Generalized Sobol Indices 43 Test functions g j = 0 f(x) = d (µ j + τ j g j (x j )) j=1 g 2 j = 1 and gj 4 <. σ 2 u = j u τ 2 j j u µ 2 j g(x) = 12(x 1/2) Min function f(x) = min 1 j d x j τ 2 u = u (d + 1) 2 (2d u + 2) Liu and O. (2006)
44 Generalized Sobol Indices 44 σ 2 {1,2,3} Product function = numerically same estimate for simple or bilinear. Therefore bilinear is better because of lower cost. For min(x) and d = 6 the bilinear estimator was about 5 times as efficient as the simple one based on n = 1,000,000 (x, z) pairs. Υ 2 {1,2,3,4} Product function with d = 8 and µ j = 1 and τ = (4, 4, 3, 3, 2, 2, 1, 1)/4. Square beats bilinear: Measure Value R 2 Square s efficiency Υ 2 {1,2,3,4} Υ 2 {5,6,7,8} Hard to beat a sum of squares when the true effect is small.
45 Generalized Sobol Indices 45 Lower index τ 2 u No sum of squares is available. 1 n 1 n Contrast n f(x i )(f(x i,u :z i, u ) f(z i )) i=1 Simple estimator (bias adjusted) n f(x i )f(x i,u :z i, u ) i=1 ( 1 2n n f(x i ) + f(x i,u :z i, u ) i=1 ) 2 The contrast has an advantage on small τ 2 u. The simple estimator sometimes beats it on large ones.
46 Generalized Sobol Indices 46 GSIs so far Just use 2 inputs, x and z What about 3? x, y, z
47 Generalized Sobol Indices 47 Here it pays to use 3 vectors x, y, x [0, 1] d For small τ 2 u 1 n 1 n 1 n 1 n n f(x i ) ( f(x i,u :y i, u ) f(y) ) i=1 (Mauntz-Saltelli) n ( f(xi ) µ )( f(x i,u :y i, u ) f(y) ) (Oracle centered) i=1 n ( f(xi ) µ )( f(x i,u :y i, u ) µ ) (Double oracle) i=1 n ( f(xi ) f(z i,u :x i, u ) )( f(x i,u :y i, u ) f(y) ) (Use 3 vectors) ( ) i=1 where (x i, y i, z i ) iid U[0, 1] 3d for i = 1,..., n. Simulations: On small effects the new estimator beats both oracles. Double oracle wins on large effects.
48 Generalized Sobol Indices 48 Conclusions Sums of squares are very good. Bilinear estimators λ T Θγ work well, especially when 1 T γ = 1 T λ = 0. Further work 1) Pursue variance inequalities 2) Replace plain MC by Quasi-Monte Carlo and/or 3) Find nice confidence intervals for ratios of means over U -statistics 4) Variance reductions
49 Generalized Sobol Indices 49 Thanks Bertrand Iooss for answering questions about Saint-Venant Sergei Kucherenko for extensive discussions National Science Foundation of the U.S. DMS
50 Generalized Sobol Indices 50 Let η 2 = u δ uσ 2 u. Optimal estimates We would like E (ˆη 2 ) = η 2 and, Var (ˆη 2) cost = minimum. Using variance components theory Unfortunately Var(ˆη 2 ) depends on 4 th moments Fortunately There is a theory of MINimum Quadratic Norm UNbiased Estimates (MINQUE) Unfortunately They do not appear to be available for crossed random effects Fortunately The computed case gives us more options, e.g., quadrature. C. R. Rao (1970s)
VARIANCE COMPONENTS AND GENERALIZED SOBOL INDICES. Copright on this work belongs to SIAM.
VARIANCE COMPONENTS AND GENERALIZED SOBOL INDICES ART B. OWEN Abstract. This paper introduces generalized Sobol indices, compares strategies for their estimation, and makes a systematic search for efficient
More informationSobol Indices: an Introduction and. Some Recent Results
Sobol indices 1 Sobol Indices: an Introduction and Some Recent Results Art B. Owen Stanford University Sobol indices 2 Ilya Meerovich Sobol At MCM 2001, Salzburg Known for Sobol sequences and Sobol indices
More informationVARIANCE COMPONENTS AND GENERALIZED SOBOL INDICES. Art B. Owen. Department of Statistics STANFORD UNIVERSITY Stanford, California
VARIANCE COMPONENTS AND GENERALIZED SOBOL INDICES By Art B. Owen Technical Report No. 2012-07 April 2012 Department of Statistics STANFORD UNIVERSITY Stanford, California 94305-4065 VARIANCE COMPONENTS
More informationOn Shapley value for measuring importance. of dependent inputs
Shapley value & variable importance 1 On Shapley value for measuring importance of dependent inputs Art B. Owen, Stanford joint with Clémentine Prieur, Grenoble Based on forthcoming JUQ article (2017).
More informationBetter estimation of small Sobol sensitivity indices
Better estimation of small Sobol sensitivity indices ART B. OWEN Stanford University September 2012 c ACM, 2012. This is the author s version of the work. It is posted here by permission of ACM for your
More informationDerivative-based global sensitivity measures for interactions
Derivative-based global sensitivity measures for interactions Olivier Roustant École Nationale Supérieure des Mines de Saint-Étienne Joint work with J. Fruth, B. Iooss and S. Kuhnt SAMO 13 Olivier Roustant
More informationTotal interaction index: A variance-based sensitivity index for interaction screening
Total interaction index: A variance-based sensitivity index for interaction screening Jana Fruth, Olivier Roustant, Sonja Kuhnt To cite this version: Jana Fruth, Olivier Roustant, Sonja Kuhnt. Total interaction
More informationTotal interaction index: A variance-based sensitivity index for second-order interaction screening
Total interaction index: A variance-based sensitivity index for second-order interaction screening J. Fruth a, O. Roustant b, S. Kuhnt a,c a Faculty of Statistics, TU Dortmund University, Vogelpothsweg
More informationEffective dimension of some weighted pre-sobolev spaces with dominating mixed partial derivatives
Effective dimension of some weighted pre-sobolev spaces with dominating mixed partial derivatives Art B. Owen Stanford University August 2018 Abstract This paper considers two notions of effective dimension
More information18Ï È² 7( &: ÄuANOVAp.O`û5 571 Based on this ANOVA model representation, Sobol (1993) proposed global sensitivity index, S i1...i s = D i1...i s /D, w
A^VÇÚO 1 Êò 18Ï 2013c12 Chinese Journal of Applied Probability and Statistics Vol.29 No.6 Dec. 2013 Optimal Properties of Orthogonal Arrays Based on ANOVA High-Dimensional Model Representation Chen Xueping
More informationStratified Random Sampling for Dependent Inputs
Stratified Random Sampling for Dependent Inputs Anirban Mondal Case Western Reserve University, Cleveland, OH 44106, USA arxiv:1904.00555v1 [stat.me] 1 Apr 2019 Abhijit Mandal Wayne State University, Detroit,
More informationEstimating mean dimensionality of ANOVA. decompositions
Estimating mean dimensionality of ANOVA decompositions Ruixue Liu 1 and Art B. Owen 2 Department of Statistics Stanford University Stanford CA, 94305 Orig: June 2003. Revised: March 2005 Abstract The analysis
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationMonte Carlo Methods. Geoff Gordon February 9, 2006
Monte Carlo Methods Geoff Gordon ggordon@cs.cmu.edu February 9, 2006 Numerical integration problem 5 4 3 f(x,y) 2 1 1 0 0.5 0 X 0.5 1 1 0.8 0.6 0.4 Y 0.2 0 0.2 0.4 0.6 0.8 1 x X f(x)dx Used for: function
More informationSCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES. Kinjal Basu Art B. Owen
SCRAMBLED GEOMETRIC NET INTEGRATION OVER GENERAL PRODUCT SPACES By Kinjal Basu Art B. Owen Technical Report No. 2015-07 March 2015 Department of Statistics STANFORD UNIVERSITY Stanford, California 94305-4065
More informationSobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis
Sobol-Hoeffding decomposition Application to Global SA Computation of the SI Sobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis Olivier Le Maître with Colleague & Friend Omar
More informationA comparison of global sensitivity techniques and sampling method
22nd International Congress on Modelling and Simulation, Hobart, Tasmania, Australia, 3 to 8 December 2017 mssanz.org.au/modsim2017 A comparison of global sensitivity techniques and sampling method X.
More informationMulti-fidelity sensitivity analysis
Multi-fidelity sensitivity analysis Loic Le Gratiet 1,2, Claire Cannamela 2 & Bertrand Iooss 3 1 Université Denis-Diderot, Paris, France 2 CEA, DAM, DIF, F-91297 Arpajon, France 3 EDF R&D, 6 quai Watier,
More informationWavelets For Computer Graphics
{f g} := f(x) g(x) dx A collection of linearly independent functions Ψ j spanning W j are called wavelets. i J(x) := 6 x +2 x + x + x Ψ j (x) := Ψ j (2 j x i) i =,..., 2 j Res. Avge. Detail Coef 4 [9 7
More informationMinimizing the Number of Function Evaluations to Estimate Sobol Indices Using Quasi-Monte Carlo
Minimizing the Number of Function Evaluations to Estimate Sobol Indices Using Quasi-Monte Carlo Lluís Antoni Jiménez Rugama Joint work with: Fred J Hickernell (IIT), Clémentine Prieur (Univ Grenoble),
More informationQuasi-Monte Carlo integration over the Euclidean space and applications
Quasi-Monte Carlo integration over the Euclidean space and applications f.kuo@unsw.edu.au University of New South Wales, Sydney, Australia joint work with James Nichols (UNSW) Journal of Complexity 30
More informationParameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn
Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth
More informationBrief introduction to Markov Chain Monte Carlo
Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical
More informationAccuracy, Precision and Efficiency in Sparse Grids
John, Information Technology Department, Virginia Tech.... http://people.sc.fsu.edu/ jburkardt/presentations/ sandia 2009.pdf... Computer Science Research Institute, Sandia National Laboratory, 23 July
More informationAnalysis of Algorithms I: Perfect Hashing
Analysis of Algorithms I: Perfect Hashing Xi Chen Columbia University Goal: Let U = {0, 1,..., p 1} be a huge universe set. Given a static subset V U of n keys (here static means we will never change the
More informationA GENERAL THEORY FOR ORTHOGONAL ARRAY BASED LATIN HYPERCUBE SAMPLING
Statistica Sinica 26 (2016), 761-777 doi:http://dx.doi.org/10.5705/ss.202015.0029 A GENERAL THEORY FOR ORTHOGONAL ARRAY BASED LATIN HYPERCUBE SAMPLING Mingyao Ai, Xiangshun Kong and Kang Li Peking University
More informationGeneralized Sobol indices for dependent variables
Generalized Sobol indices for dependent variables Gaelle Chastaing Fabrice Gamboa Clémentine Prieur July 1st, 2013 1/28 Gaelle Chastaing Sensitivity analysis and dependent variables Contents 1 Context
More informationTutorial on quasi-monte Carlo methods
Tutorial on quasi-monte Carlo methods Josef Dick School of Mathematics and Statistics, UNSW, Sydney, Australia josef.dick@unsw.edu.au Comparison: MCMC, MC, QMC Roughly speaking: Markov chain Monte Carlo
More informationThe square root rule for adaptive importance sampling
The square root rule for adaptive importance sampling Art B. Owen Stanford University Yi Zhou January 2019 Abstract In adaptive importance sampling, and other contexts, we have unbiased and uncorrelated
More informationSampling from complex probability distributions
Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation
More informationPascal s Triangle on a Budget. Accuracy, Precision and Efficiency in Sparse Grids
Covering : Accuracy, Precision and Efficiency in Sparse Grids https://people.sc.fsu.edu/ jburkardt/presentations/auburn 2009.pdf... John Interdisciplinary Center for Applied Mathematics & Information Technology
More informationIf we want to analyze experimental or simulated data we might encounter the following tasks:
Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction
More informationMathematics Department Stanford University Math 61CM/DM Inner products
Mathematics Department Stanford University Math 61CM/DM Inner products Recall the definition of an inner product space; see Appendix A.8 of the textbook. Definition 1 An inner product space V is a vector
More informationA General Overview of Parametric Estimation and Inference Techniques.
A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying
More informationVariational sampling approaches to word confusability
Variational sampling approaches to word confusability John R. Hershey, Peder A. Olsen and Ramesh A. Gopinath {jrhershe,pederao,rameshg}@us.ibm.com IBM, T. J. Watson Research Center Information Theory and
More informationSpanning and Independence Properties of Finite Frames
Chapter 1 Spanning and Independence Properties of Finite Frames Peter G. Casazza and Darrin Speegle Abstract The fundamental notion of frame theory is redundancy. It is this property which makes frames
More informationA CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS
Statistica Sinica 26 (2016), 1117-1128 doi:http://dx.doi.org/10.5705/ss.202015.0240 A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS Xu He and Peter Z. G. Qian Chinese Academy of Sciences
More informationBOUNDARY VALUE PROBLEMS ON A HALF SIERPINSKI GASKET
BOUNDARY VALUE PROBLEMS ON A HALF SIERPINSKI GASKET WEILIN LI AND ROBERT S. STRICHARTZ Abstract. We study boundary value problems for the Laplacian on a domain Ω consisting of the left half of the Sierpinski
More information1 Probabilities. 1.1 Basics 1 PROBABILITIES
1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationCS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization
CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization Tim Roughgarden February 28, 2017 1 Preamble This lecture fulfills a promise made back in Lecture #1,
More informationP (A G) dp G P (A G)
First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume
More informationMATH 12 CLASS 23 NOTES, NOV Contents 1. Change of variables: the Jacobian 1
MATH 12 CLASS 23 NOTES, NOV 11 211 Contents 1. Change of variables: the Jacobian 1 1. Change of variables: the Jacobian So far, we have seen three examples of situations where we change variables to help
More informationGaussian process regression for Sensitivity analysis
Gaussian process regression for Sensitivity analysis GPSS Workshop on UQ, Sheffield, September 2016 Nicolas Durrande, Mines St-Étienne, durrande@emse.fr GPSS workshop on UQ GPs for sensitivity analysis
More informationChapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability
Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with
More informationStochastic Subgradient Methods
Stochastic Subgradient Methods Stephen Boyd and Almir Mutapcic Notes for EE364b, Stanford University, Winter 26-7 April 13, 28 1 Noisy unbiased subgradient Suppose f : R n R is a convex function. We say
More informationSTA 294: Stochastic Processes & Bayesian Nonparametrics
MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a
More informationBrief Review on Estimation Theory
Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on
More information1 Probabilities. 1.1 Basics 1 PROBABILITIES
1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationApplications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices
Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Vahid Dehdari and Clayton V. Deutsch Geostatistical modeling involves many variables and many locations.
More informationCHAPTER 3 Further properties of splines and B-splines
CHAPTER 3 Further properties of splines and B-splines In Chapter 2 we established some of the most elementary properties of B-splines. In this chapter our focus is on the question What kind of functions
More information4th Preparation Sheet - Solutions
Prof. Dr. Rainer Dahlhaus Probability Theory Summer term 017 4th Preparation Sheet - Solutions Remark: Throughout the exercise sheet we use the two equivalent definitions of separability of a metric space
More informationIsomorphisms between pattern classes
Journal of Combinatorics olume 0, Number 0, 1 8, 0000 Isomorphisms between pattern classes M. H. Albert, M. D. Atkinson and Anders Claesson Isomorphisms φ : A B between pattern classes are considered.
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationEco517 Fall 2004 C. Sims MIDTERM EXAM
Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering
More informationRECYCLING PHYSICAL RANDOM NUMBERS. Art B. Owen. Technical Report No September 2009
RECYCLING PHYSICAL RANDOM NUMBERS By Art B. Owen Technical Report No. 2009-11 September 2009 Department of Statistics STANFORD UNIVERSITY Stanford, California 94305-4065 RECYCLING PHYSICAL RANDOM NUMBERS
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationRecall: Dot product on R 2 : u v = (u 1, u 2 ) (v 1, v 2 ) = u 1 v 1 + u 2 v 2, u u = u u 2 2 = u 2. Geometric Meaning:
Recall: Dot product on R 2 : u v = (u 1, u 2 ) (v 1, v 2 ) = u 1 v 1 + u 2 v 2, u u = u 2 1 + u 2 2 = u 2. Geometric Meaning: u v = u v cos θ. u θ v 1 Reason: The opposite side is given by u v. u v 2 =
More informationPart III. Quasi Monte Carlo methods 146/349
Part III Quasi Monte Carlo methods 46/349 Outline Quasi Monte Carlo methods 47/349 Quasi Monte Carlo methods Let y = (y,...,y N ) be a vector of independent uniform random variables in Γ = [0,] N, u(y)
More information2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?
MA 645-4A (Real Analysis), Dr. Chernov Homework assignment 1 (Due 9/5). Prove that every countable set A is measurable and µ(a) = 0. 2 (Bonus). Let A consist of points (x, y) such that either x or y is
More informationBANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1
BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1 Shaobo Li University of Cincinnati 1 Partially based on Hastie, et al. (2009) ESL, and James, et al. (2013) ISLR Data Mining I Lecture
More informationTutorials in Optimization. Richard Socher
Tutorials in Optimization Richard Socher July 20, 2008 CONTENTS 1 Contents 1 Linear Algebra: Bilinear Form - A Simple Optimization Problem 2 1.1 Definitions........................................ 2 1.2
More informationAsymptotics of minimax stochastic programs
Asymptotics of minimax stochastic programs Alexander Shapiro Abstract. We discuss in this paper asymptotics of the sample average approximation (SAA) of the optimal value of a minimax stochastic programming
More informationThe Hilbert Space of Random Variables
The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2
More informationSCALE INVARIANT FOURIER RESTRICTION TO A HYPERBOLIC SURFACE
SCALE INVARIANT FOURIER RESTRICTION TO A HYPERBOLIC SURFACE BETSY STOVALL Abstract. This result sharpens the bilinear to linear deduction of Lee and Vargas for extension estimates on the hyperbolic paraboloid
More informationMATH Linear Algebra
MATH 4 - Linear Algebra One of the key driving forces for the development of linear algebra was the profound idea (going back to 7th century and the work of Pierre Fermat and Rene Descartes) that geometry
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2
MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 1 Bootstrapped Bias and CIs Given a multiple regression model with mean and
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 6: Bias and variance (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 49 Our plan today We saw in last lecture that model scoring methods seem to be trading off two different
More informationLAGRANGE MULTIPLIERS
LAGRANGE MULTIPLIERS MATH 195, SECTION 59 (VIPUL NAIK) Corresponding material in the book: Section 14.8 What students should definitely get: The Lagrange multiplier condition (one constraint, two constraints
More informationCSE 312: Foundations of Computing II Quiz Section #10: Review Questions for Final Exam (solutions)
CSE 312: Foundations of Computing II Quiz Section #10: Review Questions for Final Exam (solutions) 1. (Confidence Intervals, CLT) Let X 1,..., X n be iid with unknown mean θ and known variance σ 2. Assume
More informationFunctions of Several Variables
Jim Lambers MAT 419/519 Summer Session 2011-12 Lecture 2 Notes These notes correspond to Section 1.2 in the text. Functions of Several Variables We now generalize the results from the previous section,
More informationPolynomial chaos expansions for sensitivity analysis
c DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for sensitivity analysis B. Sudret Chair of Risk, Safety & Uncertainty
More informationAnalysis II - few selective results
Analysis II - few selective results Michael Ruzhansky December 15, 2008 1 Analysis on the real line 1.1 Chapter: Functions continuous on a closed interval 1.1.1 Intermediate Value Theorem (IVT) Theorem
More informationLinear Models Review
Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign
More informationA Few Notes on Fisher Information (WIP)
A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties
More informationTHE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR
THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR 1. Definition Existence Theorem 1. Assume that A R m n. Then there exist orthogonal matrices U R m m V R n n, values σ 1 σ 2... σ p 0 with p = min{m, n},
More informationIEOR 165 Lecture 7 1 Bias-Variance Tradeoff
IEOR 165 Lecture 7 Bias-Variance Tradeoff 1 Bias-Variance Tradeoff Consider the case of parametric regression with β R, and suppose we would like to analyze the error of the estimate ˆβ in comparison to
More informationHILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define
HILBERT SPACES AND THE RADON-NIKODYM THEOREM STEVEN P. LALLEY 1. DEFINITIONS Definition 1. A real inner product space is a real vector space V together with a symmetric, bilinear, positive-definite mapping,
More informationDiscrete Latent Variable Models
Discrete Latent Variable Models Stefano Ermon, Aditya Grover Stanford University Lecture 14 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 14 1 / 29 Summary Major themes in the course
More informationECE 275A Homework 7 Solutions
ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator
More informationDiscrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14
CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten
More informationPoincaré inequalities revisited for dimension reduction
Poincaré inequalities revisited for dimension reduction Olivier Roustant*, Franck Barthe** and Bertrand Iooss**, *** * Mines Saint-Étienne ** Institut de Mathématiques de Toulouse *** EDF, Chatou March
More informationEcon 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines
Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the
More informationEvaluating the Performance of Estimators (Section 7.3)
Evaluating the Performance of Estimators (Section 7.3) Example: Suppose we observe X 1,..., X n iid N(θ, σ 2 0 ), with σ2 0 known, and wish to estimate θ. Two possible estimators are: ˆθ = X sample mean
More information1 Structures 2. 2 Framework of Riemann surfaces Basic configuration Holomorphic functions... 3
Compact course notes Riemann surfaces Fall 2011 Professor: S. Lvovski transcribed by: J. Lazovskis Independent University of Moscow December 23, 2011 Contents 1 Structures 2 2 Framework of Riemann surfaces
More informationStat 451 Lecture Notes Numerical Integration
Stat 451 Lecture Notes 03 12 Numerical Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 5 in Givens & Hoeting, and Chapters 4 & 18 of Lange 2 Updated: February 11, 2016 1 / 29
More informationPenalty Methods for Bivariate Smoothing and Chicago Land Values
Penalty Methods for Bivariate Smoothing and Chicago Land Values Roger Koenker University of Illinois, Urbana-Champaign Ivan Mizera University of Alberta, Edmonton Northwestern University: October 2001
More informationLECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel
LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count
More informationEigenvectors and Hermitian Operators
7 71 Eigenvalues and Eigenvectors Basic Definitions Let L be a linear operator on some given vector space V A scalar λ and a nonzero vector v are referred to, respectively, as an eigenvalue and corresponding
More informationExercises on chapter 0
Exercises on chapter 0 1. A partially ordered set (poset) is a set X together with a relation such that (a) x x for all x X; (b) x y and y x implies that x = y for all x, y X; (c) x y and y z implies that
More informationLecture 6 Proof for JL Lemma and Linear Dimensionality Reduction
COMS 4995: Unsupervised Learning (Summer 18) June 7, 018 Lecture 6 Proof for JL Lemma and Linear imensionality Reduction Instructor: Nakul Verma Scribes: Ziyuan Zhong, Kirsten Blancato This lecture gives
More informationNumerical integration and importance sampling
2 Numerical integration and importance sampling 2.1 Quadrature Consider the numerical evaluation of the integral I(a, b) = b a dx f(x) Rectangle rule: on small interval, construct interpolating function
More information9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career.
Introduction to Data and Analysis Wildlife Management is a very quantitative field of study Results from studies will be used throughout this course and throughout your career. Sampling design influences
More informationGI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil
GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection
More informationGaussian Processes for Computer Experiments
Gaussian Processes for Computer Experiments Jeremy Oakley School of Mathematics and Statistics, University of Sheffield www.jeremy-oakley.staff.shef.ac.uk 1 / 43 Computer models Computer model represented
More informationNotes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed
18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,
More informationMath 2 Variable Manipulation Part 7 Absolute Value & Inequalities
Math 2 Variable Manipulation Part 7 Absolute Value & Inequalities 1 MATH 1 REVIEW SOLVING AN ABSOLUTE VALUE EQUATION Absolute value is a measure of distance; how far a number is from zero. In practice,
More information