Zero Variance Markov Chain Monte Carlo for Bayesian Estimators

Size: px
Start display at page:

Download "Zero Variance Markov Chain Monte Carlo for Bayesian Estimators"

Transcription

1 Noname manuscript No. will be inserted by the editor Zero Variance Markov Chain Monte Carlo for Bayesian Estimators Supplementary material Antonietta Mira Reza Solgi Daniele Imparato Received: date / Accepted: date Appendix A: Probit model Mathematical formulation Let y i be Bernoulli r.v. s: y i x i B, p i, p i = Φx T i β, where β Rd is the vector of parameters of the model and Φ is the c.d.f. of a standard normal distribution. The likelihood function is: lβ y, x [ Φx T i β ] y i [ Φx T i β ] y i. As it can be seen by inspection, the likelihood function is invariant under the transformation x i, y i x i, y i. Therefore, for the sake of simplicity, in the rest of the example we assume y i = for any i, so that the likelihood simplifies: lβ y, x Φx T i β. This formula shows that the contribution of x i = 0 is just a constant Φx T i β = Φ0 =, therefore, without loss of generality, we assume for all i, x i 0. A. Mira Swiss Finance Institute, University of Lugano, via Buffi 3 CH-6904 Lugano, Switzerland. antonietta.mira@usi.ch R. Solgi Swiss Finance Institute, University of Lugano, via Buffi 3 CH-6904 Lugano, Switzerland. reza.solgi@usi.ch D. Imparato Department of Economics, University of Insubria, via Monte Generoso 7, 00 Varese, Italy. daniele.imparato@uninsubria.it

2 Antonietta Mira et al. Using flat priors, the posterior of the model is proportional to the likelihood, and the Bayesian estimator of each parameter, β k, is the expected value of f k β = β k under π k =,,, d. Using Schrödinger-type Hamiltonians, H and ψ k β = P k β πβ, as the trial functions, where P k β = d j= a j,kβ j is a first degree polynomial, one gets: f k β = f k β + where, for j =,,..., d, Hψ kβ πβ y, x = f k β + z j = n because of the assumption y i = for any i. x ij φx T i β Φx T i β, d a j,k z j, j= Central limit theorem In the following, it is supposed that P is a linear polynomial. In the Probit model, the ZV-MCMC estimators obey a CLT if z j have finite + δ moment under π, for some δ > 0: [ E π zj +δ] = c E π n x ij φx T i β +δ Φx T i β = c c n x ij φx T i β Φx T i β +δ n Φx T i βdβ <. where c = δ, and c is the normalizing constant of π the target posterior. Define: and therefore: n x ij φx T i K β = β Φx T i β K β = Φx T i β, Kβ = K βk β +δ [ E π zj +δ] = c K βk βdβ. where c = c c. Before studying the convergence of this integral, the following property of the likelihood for the probit model is needed.,

3 Zero Variance Markov Chain Monte Carlo for Bayesian Estimators 3 Proposition Existence and uniqueness of MLE implies that, for any β 0 \ {0}, there exists i such that x T i β 0 < 0. Proof by contradiction. Uniqueness of MLE implies that x T x is full rank, that is, there is no β 0 orthogonal to all observations x i. This can be seen by contradiction: singularity of x T x implies existence of a non-zero β 0 orthogonal to all observations x i. This fact, in turn, implies lβ x, y = lβ + cβ 0 x, y and, therefore, l x, y does not have a unique global maximum. Next, assume there exists some β 0 such that, for any i, x T i β 0 > 0. Then β 0 is a direction of recession for the negative log-likelihood function n ln ΦxT i β that is a proper closed convex function. This implies that this function does not have non-empty bounded minimum set [], which means that the MLE does not exist. Now, rewriting Kβdβ in hyper-spherical coordinates through the bijective transformation ρ, θ,..., θ d := F β, where F is defined as β = ρ cosθ β l = ρ cosθ l l m= sinθ m, for l =,..., d β d = ρ d m= sinθ m, for θ Θ := {0 θ i π, i =,..., d, 0 θ d < π} and ρ > 0, one gets Kβdβ = := + Θ 0 + Θ Θ 0 Aθdθ, d KF ρ, θρ d sin d j θ j dρdθ. j= KF ρ, θρ d dρdθ Observe that the integrand is well defined for any ρ, θ on the domain of integration, so it is enough to study its asymptotic behaviour when ρ goes to infinity, and θ Θ. First, analyze K F n ρ, θ = x ij φ x i ρλ i θ Φ x i ρλ i θ where, for any i, λ i is a suitable function of the angles θ such that λ i [, ], which takes into account the sign of the scalar product in the original coordinates system. For any i, when ρ if λ i < 0, xijφ xi ρλi Φ x i ρλ i O ρ; +δ,

4 4 Antonietta Mira et al. if λ i > 0, xijφ xi ρλi Φ x i ρλ i if λ i = 0, xijφ xi ρλi Φ x i ρλ i Therefore: O φλ i ρ; = x ij π O. n x ij φ x i ρλ i Φ x i ρλ i O ρ and, for any θ Θ: K F ρ, θ O ρ +δ. Now, focus on K F ρ, θ = n Φ x i ρλ i θ; existence of MLE for the probit model implies that, for any θ Θ, there exists some l l n, such that λ l θ < 0, and therefore: K F ρ, θ < Φ x l ρλ l O φλ l ρ ρ. Putting these results together leads to KF ρ, θ = K F ρ, θk F ρ, θ O ρ +δ φλ l θρ so that, for any θ Θ, KF ρ, θρ d O ρ +δ+d φ λ l θρ, ρ +. Therefore, whenever the value θ Θ, its integrand converges to zero rapidly enough when ρ +. This concludes the proof. Note. In [] it is shown that the existence of the posterior under flat priors for probit and logit models is equivalent to the existence and finiteness of MLE. This ensures us that the posterior is well defined in our context. In order to verify the existence of the posterior mean, we can use a simplified version of the proof given above. In other words we should show βj n ΦxT i βdβ < +, that is, K β = β j Oρ. Therefore, a weaker version of the proof given above can be employed. Note. In the proof given above we have used flat priors: although this assumption simplifies the proof, however a very similar proof can be applied for non-flat priors. Assume the prior is π 0 β. Under this assumption the posterior is πβ = π 0 β lβ y, x π 0 β n β and the control variates are: x ijφx T i β Φx T i β ΦxT i z j = d ln π 0β dβ j n. Therefore we need to prove the + δ-th moment of z j under πβ is finite. A sufficient condition for this is the finiteness of + δ-th moments of d ln π 0β dβ j and n x ijφx T i β Φx T i β under πβ. If we assume π 0 β is bounded above, the latter is a trivial consequence of the proof given above for the flat priors. Therefore we only need to prove the finiteness of the integral d ln π 0 β dβ j +δ π 0 β Φx T i βdβ.

5 Zero Variance Markov Chain Monte Carlo for Bayesian Estimators 5 Again if we assume the prior is bounded from above, a sufficient condition for the existence of this integral is the existence of the following integral: +δ d ln π 0 β n dβ j Φx T i βdβ A proof very similar to the one given above will show that this integral is finite for common choices of priors π 0 β such as Normal, Student s T, etc. Appendix B: Logit model Mathematical formulation In the same setting as the probit model, let p i = expxt i β +expx T i β where β Rd is the vector of parameters of the model. The likelihood function is: lβ y, x expx T i β + expx T i β yi + expx T i β yi. 3 By inspection, it is easy to verify that the likelihood function is invariant under the transformation:x i, y i x i, y i. Therefore, for the sake of simplicity, in the sequel we assume y i = 0 for any i, so that the likelihood simplifies as: lβ y, x + expx T i β. The contribution of x i = 0 to the likelihood is just a constant, therefore, without loss of generality, it is assumed that x i 0 for all i. Using flat priors, the posterior distribution is proportional to 3 and the Bayesian estimator of each parameter, β k, is the expected value of f k β = β k under π k =,,, d. Using the same pair of operator H and test function ψ k as in Appendinx A, the control variates are: z j = n expx T i x β ij + expx T for j =,,..., d. i β, Central limit theorem As for the probit model, the ZV-MCMC estimators obey a CLT if the control variates z j have finite + δ moment under π, for some δ > 0 : [ ] n +δ E π z +δ j = c E π expx T i x β ij + expx T i β = c c n +δ expx T i x β n ij + expx T i β + expx T i βdβ

6 6 Antonietta Mira et al. where c = δ, and c is the normalizing constant of π. The finitiness of the integral is, indeed, trivial. Observe that the function e y / + e y is bounded by ; therefore, E π [ z +δ j ] n +δ x ij <. As for probit model, note that it can easily be shown that the posterior means exist under flat priors. Moreover, a very similar proof see Note of Appendix A can be used under a normal or Student s T prior distribution. Appendix C: GARCH model Mathematical formulation We assume that the returns are conditionally normal distributed, rt F t N 0, h t, where h t is a predictable F t measurable process: h t = ω + ω 3 h t + ω r t, where ω > 0, ω 0, and ω 3 0. Let r = r,..., r T be the observed time series. The likelihood function is equal to: l ω, ω, ω 3 r T t= h t exp and using independent truncated Normal priors for the parameters, the posterior is: [ π ω, ω, ω 3 r exp ] ω σ ω + ω σ ω + ω T 3 σ h t exp ω 3 t= Therefore, the control variates when the trial function is a first degree polynomial are: t= r t h t t= r t h t. ln π ω i = ω i σ ω i h t= t h t ω i + r t h t= t h t ω i, i =,, 3, where: h t = ωt 3, ω ω 3 h t = r h t t + ω 3 I t>, ω ω h t h t = h t + ω 3 I t>. ω 3 ω 3

7 Zero Variance Markov Chain Monte Carlo for Bayesian Estimators 7 Central limit theorem In order to prove the CLT for the ZV-MCMC estimator in the Garch model, we need: ln π ln π ln π,, L +δ π. 4 ω ω 3 ω To this end, h t and its partial derivatives should be expressed as a function of h 0 and r: t t h t = ω + ω3 k + ω3h t 0 + ω + ω3 k rt k, k= ln h t ω ln h t ω = ln h t ω 3 = = ω t 3 ω 3 I {t>}, t rt + t h t + k= ω t j 3 rj j=0 ω t j 3 h j j=0 I {t>}, I {t>}. Next, moving to spherical coordinates, the integral 4 can be written as K j ρ; θ, φdρdθdφ := A j θ, φdθdφ, [0,π/] 0 [0,π/] where, for j =,, 3, K j ; θ, φ = W j +δ W, with W = σ ω ρ cos θ sin φ W = σ ω ρ sin θ sin φ W 3 = σ ω ρ cos φ 3 t= h t W = exp and ρ sin θ t= h t t= h t r t h t r t ρ t cos t φ, h ρ cos φ t r t r h t t ht + t t + x t j 3 rj j=0 j=0 x t j 3 h j, ρ σ ω cos θ sin φ + σ ω sin θ sin φ + σ ω 3 cos φ T t= h t t t h t = ρ cos θ sin φ +ρ k cos k φ+h 0 ρ t cos t φ+ρ sin θ sin φ +rt kρ k cos k φ. k= k=, 5 t= r t h t

8 8 Antonietta Mira et al. The aim is to prove that, for any θ, φ [0, π/] and for any j, A j θ, φ is finite. To this end, the convergence of A j for any θ, φ should be discussed. Let us study the proper domain of K j ; θ, φ. Observe that K j ; θ, φ is not defined whenever h t = 0 and, if j = and φ π/, also for ρ = / cos φ. However, the discontinuity of K 3 at this point is removable, so that the domain of K 3 can be extended by continuity also at ρ = / cos φ. Since h t = 0 if and only if ρ = 0, it can be concluded that, for any j and for any θ, φ [0, π/], the proper domain of K j ; θ, φ is domk j ; θ, φ = 0+. By fixing the value of θ and φ, let us study the limits of K j when ρ 0 and ρ +. Observe that, whatever the values of θ and φ are, W j s are rationale functions of ρ. Therefore, for any j, W j +δ cannot grow towards infinity more than polynomially at the boundary of the domain. On the other hand, W goes to zero with an exponential rate both when ρ 0 and ρ +, for any θ and φ. This is sufficient to conclude that, for any θ, φ [0, π/], the integral A j is finite for j =,, 3 and, therefore, condition 4 holds and the ZV estimators for the GARCH model obeys a CLT. Appendix D: unbiasedness In this Appendix, explicit computations are presented, which were omitted in Section 5. Moreover, it is proved that all the ZV-MCMC estimators discussed in Section 8 are unbiased. Following the same notations as in Section 5, equation 9 follows because Hψ := π = Ω Ω Hψ π V ψ π ψ π = V πψ π ψ ndσ + π ψ Ω Ω Ω = V πψ π ψ ndσ + ψ π ndσ Ω Ω Ω = ΩH πψ + [ψ π π ψ] ndσ Ω = [ψ π π ψ] ndσ. Ω Ω ψ π Then, Therefore, Hψ = 0 if ψ π = π ψ on Ω. Now, let ψ = P π. π ψ = π P + P π π,

9 Zero Variance Markov Chain Monte Carlo for Bayesian Estimators 9 so that Hψ π = 0 if P x πx = 0, x Ω, j =,..., d. x j When π has unbounded support, following the previous computations integrating over the bounded set B r and taking the limit for r, one gets Hψ = π lim π P ndσ. 6 r + B r Therefore, unbiasedness in the unbounded case is reached if the limit appearing in the right-hand side of 6 is zero. Now, the unbiasedness of the ZV-MCMC estimators exploited in Section 8 is discussed. To this end, condition 6 should be verified. Let B ρ be a hypersphere of radius ρ and let n := ρβ be its normal versor. Then, for linear P, 6 equals zero if, for any j =,..., d, lim ρ + ρ B ρ πββ j ds = 0. 7 The Probit model is first considered. By using the same notations as in Appendix A, the integral in 7 is proportional to lim K F ρ, θρ d dθ. 8 ρ + ρ Note that, because of, there exist ρ 0 and M such that K F ρ, θρ d Mφλ lθ ρρ d Θ Mφλ lθ ρ 0 ρ d 0 := Gθ ρ ρ 0. Since Gθ L, by the dominated convergence theorem a sufficient condition to get unbiasedness is lim K F ρ, θρ d = 0, 9 ρ + which is true, because of for the Probit model. We now consider the Logit model for which it is easy to prove that Proposition holds. As done for the Probit model, one can write [ ] E π z +δ j K βk βdβ, where K β = K β = n +δ expx T i x β ij + expx T i β, + expx T i β,

10 0 Antonietta Mira et al. By using the hyper-spherical change of variables in, we get [ ] E π z +δ j K F ρ, θk F ρ, θρ d dρ : Θ 0 and, as for the Probit model, we must verify Equation 8. Now analyze K F ρ, θ; for any θ, existence of MLE implies the existence of some l l n, such that λ l θ > 0, and therefore: K F ρ, θ = Therefore, there exist ρ 0, M such that K F ρ, θρ d M exp λ lθ ρρ d + exp x i ρλ i < + exp x l ρλ l O exp ρλ l. 0 M exp λ lθ ρ 0 ρ d 0 := Gθ ρ ρ 0, where Gθ L. These computations allow us to use Equation 9 as a sufficient condition to get unbiasedness, and its proof becomes trivial. Finally consider the GARCH model. In this case, B ρ is the portion of a sphere of radius ρ defined on the positive orthant. Then, the limit lim ρ + ρ [0,π/] W F ρ, θρ d dθ, where W was defined in 5, should be discussed. Again, an application of the dominated convergence theorem leads to the simpler condition lim W F ρ, θρ = 0, ρ + which is true, since W decays with an exponential rate. References. Rockafellar, R.: Convex analysis, pp Princeton University Press 970. Speckman P.L. Lee, J., Sun, D.: Existence of the mle and propriety of posteriors for a general multinomial choice model. Statistica Sinica 9,

Stat 5421 Lecture Notes Proper Conjugate Priors for Exponential Families Charles J. Geyer March 28, 2016

Stat 5421 Lecture Notes Proper Conjugate Priors for Exponential Families Charles J. Geyer March 28, 2016 Stat 5421 Lecture Notes Proper Conjugate Priors for Exponential Families Charles J. Geyer March 28, 2016 1 Theory This section explains the theory of conjugate priors for exponential families of distributions,

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Part III. A Decision-Theoretic Approach and Bayesian testing

Part III. A Decision-Theoretic Approach and Bayesian testing Part III A Decision-Theoretic Approach and Bayesian testing 1 Chapter 10 Bayesian Inference as a Decision Problem The decision-theoretic framework starts with the following situation. We would like to

More information

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

A Very Brief Summary of Bayesian Inference, and Examples

A Very Brief Summary of Bayesian Inference, and Examples A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X

More information

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017 Zig-Zag Monte Carlo Delft University of Technology Joris Bierkens February 7, 2017 Joris Bierkens (TU Delft) Zig-Zag Monte Carlo February 7, 2017 1 / 33 Acknowledgements Collaborators Andrew Duncan Paul

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information

Bayesian Modeling of Conditional Distributions

Bayesian Modeling of Conditional Distributions Bayesian Modeling of Conditional Distributions John Geweke University of Iowa Indiana University Department of Economics February 27, 2007 Outline Motivation Model description Methods of inference Earnings

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis

Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis Stéphanie Allassonnière CIS, JHU July, 15th 28 Context : Computational Anatomy Context and motivations :

More information

Parameter Estimation

Parameter Estimation Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y

More information

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,

More information

HIERARCHICAL BAYESIAN ANALYSIS OF BINARY MATCHED PAIRS DATA

HIERARCHICAL BAYESIAN ANALYSIS OF BINARY MATCHED PAIRS DATA Statistica Sinica 1(2), 647-657 HIERARCHICAL BAYESIAN ANALYSIS OF BINARY MATCHED PAIRS DATA Malay Ghosh, Ming-Hui Chen, Atalanta Ghosh and Alan Agresti University of Florida, Worcester Polytechnic Institute

More information

1. Fisher Information

1. Fisher Information 1. Fisher Information Let f(x θ) be a density function with the property that log f(x θ) is differentiable in θ throughout the open p-dimensional parameter set Θ R p ; then the score statistic (or score

More information

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference 1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE

More information

Fast Maximum Likelihood estimation via Equilibrium Expectation for Large Network Data

Fast Maximum Likelihood estimation via Equilibrium Expectation for Large Network Data Fast Maximum Likelihood estimation via Equilibrium Expectation for Large Network Data Maksym Byshkin 1, Alex Stivala 4,1, Antonietta Mira 1,3, Garry Robins 2, Alessandro Lomi 1,2 1 Università della Svizzera

More information

Math 350 Solutions for Final Exam Page 1. Problem 1. (10 points) (a) Compute the line integral. F ds C. z dx + y dy + x dz C

Math 350 Solutions for Final Exam Page 1. Problem 1. (10 points) (a) Compute the line integral. F ds C. z dx + y dy + x dz C Math 35 Solutions for Final Exam Page Problem. ( points) (a) ompute the line integral F ds for the path c(t) = (t 2, t 3, t) with t and the vector field F (x, y, z) = xi + zj + xk. (b) ompute the line

More information

Separation of Variables in Polar and Spherical Coordinates

Separation of Variables in Polar and Spherical Coordinates Separation of Variables in Polar and Spherical Coordinates Polar Coordinates Suppose we are given the potential on the inside surface of an infinitely long cylindrical cavity, and we want to find the potential

More information

Lecture 7 October 13

Lecture 7 October 13 STATS 300A: Theory of Statistics Fall 2015 Lecture 7 October 13 Lecturer: Lester Mackey Scribe: Jing Miao and Xiuyuan Lu 7.1 Recap So far, we have investigated various criteria for optimal inference. We

More information

Sequential Monte Carlo Methods for Bayesian Computation

Sequential Monte Carlo Methods for Bayesian Computation Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

Likelihood-free MCMC

Likelihood-free MCMC Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

Bohr & Wheeler Fission Theory Calculation 4 March 2009

Bohr & Wheeler Fission Theory Calculation 4 March 2009 Bohr & Wheeler Fission Theory Calculation 4 March 9 () Introduction The goal here is to reproduce the calculation of the limiting Z /A against spontaneous fission Z A lim a S. (.) a C as first done by

More information

Math 1B Final Exam, Solution. Prof. Mina Aganagic Lecture 2, Spring (6 points) Use substitution and integration by parts to find:

Math 1B Final Exam, Solution. Prof. Mina Aganagic Lecture 2, Spring (6 points) Use substitution and integration by parts to find: Math B Final Eam, Solution Prof. Mina Aganagic Lecture 2, Spring 20 The eam is closed book, apart from a sheet of notes 8. Calculators are not allowed. It is your responsibility to write your answers clearly..

More information

Bayesian Multivariate Logistic Regression

Bayesian Multivariate Logistic Regression Bayesian Multivariate Logistic Regression Sean M. O Brien and David B. Dunson Biostatistics Branch National Institute of Environmental Health Sciences Research Triangle Park, NC 1 Goals Brief review of

More information

λ(x + 1)f g (x) > θ 0

λ(x + 1)f g (x) > θ 0 Stat 8111 Final Exam December 16 Eleven students took the exam, the scores were 92, 78, 4 in the 5 s, 1 in the 4 s, 1 in the 3 s and 3 in the 2 s. 1. i) Let X 1, X 2,..., X n be iid each Bernoulli(θ) where

More information

Bayesian inference for multivariate skew-normal and skew-t distributions

Bayesian inference for multivariate skew-normal and skew-t distributions Bayesian inference for multivariate skew-normal and skew-t distributions Brunero Liseo Sapienza Università di Roma Banff, May 2013 Outline Joint research with Antonio Parisi (Roma Tor Vergata) 1. Inferential

More information

STA205 Probability: Week 8 R. Wolpert

STA205 Probability: Week 8 R. Wolpert INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and

More information

Chapter 4: Modelling

Chapter 4: Modelling Chapter 4: Modelling Exchangeability and Invariance Markus Harva 17.10. / Reading Circle on Bayesian Theory Outline 1 Introduction 2 Models via exchangeability 3 Models via invariance 4 Exercise Statistical

More information

HANDBOOK OF APPLICABLE MATHEMATICS

HANDBOOK OF APPLICABLE MATHEMATICS HANDBOOK OF APPLICABLE MATHEMATICS Chief Editor: Walter Ledermann Volume VI: Statistics PART A Edited by Emlyn Lloyd University of Lancaster A Wiley-Interscience Publication JOHN WILEY & SONS Chichester

More information

Monte Carlo in Bayesian Statistics

Monte Carlo in Bayesian Statistics Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview

More information

Lecture 8: The Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm 30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:

More information

Stat 451 Lecture Notes Monte Carlo Integration

Stat 451 Lecture Notes Monte Carlo Integration Stat 451 Lecture Notes 06 12 Monte Carlo Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 23 in Lange, and Chapters 3 4 in Robert & Casella 2 Updated:

More information

David Giles Bayesian Econometrics

David Giles Bayesian Econometrics David Giles Bayesian Econometrics 1. General Background 2. Constructing Prior Distributions 3. Properties of Bayes Estimators and Tests 4. Bayesian Analysis of the Multiple Regression Model 5. Bayesian

More information

Week 2 Notes, Math 865, Tanveer

Week 2 Notes, Math 865, Tanveer Week 2 Notes, Math 865, Tanveer 1. Incompressible constant density equations in different forms Recall we derived the Navier-Stokes equation for incompressible constant density, i.e. homogeneous flows:

More information

Invariant HPD credible sets and MAP estimators

Invariant HPD credible sets and MAP estimators Bayesian Analysis (007), Number 4, pp. 681 69 Invariant HPD credible sets and MAP estimators Pierre Druilhet and Jean-Michel Marin Abstract. MAP estimators and HPD credible sets are often criticized in

More information

Lecture 26: Likelihood ratio tests

Lecture 26: Likelihood ratio tests Lecture 26: Likelihood ratio tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f θ0 (X) > c 0 for

More information

If there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell,

If there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell, Recall The Neyman-Pearson Lemma Neyman-Pearson Lemma: Let Θ = {θ 0, θ }, and let F θ0 (x) be the cdf of the random vector X under hypothesis and F θ (x) be its cdf under hypothesis. Assume that the cdfs

More information

Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies. Abstract

Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies. Abstract Bayesian Estimation of A Distance Functional Weight Matrix Model Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies Abstract This paper considers the distance functional weight

More information

e x3 dx dy. 0 y x 2, 0 x 1.

e x3 dx dy. 0 y x 2, 0 x 1. Problem 1. Evaluate by changing the order of integration y e x3 dx dy. Solution:We change the order of integration over the region y x 1. We find and x e x3 dy dx = y x, x 1. x e x3 dx = 1 x=1 3 ex3 x=

More information

Supporting Information

Supporting Information Supporting Information A: Calculation of radial distribution functions To get an effective propagator in one dimension, we first transform 1) into spherical coordinates: x a = ρ sin θ cos φ, y = ρ sin

More information

Lecture 17: Likelihood ratio and asymptotic tests

Lecture 17: Likelihood ratio and asymptotic tests Lecture 17: Likelihood ratio and asymptotic tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f

More information

Parametric Techniques Lecture 3

Parametric Techniques Lecture 3 Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to

More information

Nuisance parameters and their treatment

Nuisance parameters and their treatment BS2 Statistical Inference, Lecture 2, Hilary Term 2008 April 2, 2008 Ancillarity Inference principles Completeness A statistic A = a(x ) is said to be ancillary if (i) The distribution of A does not depend

More information

MARGINAL MARKOV CHAIN MONTE CARLO METHODS

MARGINAL MARKOV CHAIN MONTE CARLO METHODS Statistica Sinica 20 (2010), 1423-1454 MARGINAL MARKOV CHAIN MONTE CARLO METHODS David A. van Dyk University of California, Irvine Abstract: Marginal Data Augmentation and Parameter-Expanded Data Augmentation

More information

Lecture 14 More on structural estimation

Lecture 14 More on structural estimation Lecture 14 More on structural estimation Economics 8379 George Washington University Instructor: Prof. Ben Williams traditional MLE and GMM MLE requires a full specification of a model for the distribution

More information

TEORIA BAYESIANA Ralph S. Silva

TEORIA BAYESIANA Ralph S. Silva TEORIA BAYESIANA Ralph S. Silva Departamento de Métodos Estatísticos Instituto de Matemática Universidade Federal do Rio de Janeiro Sumário Numerical Integration Polynomial quadrature is intended to approximate

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

Analysis of Polya-Gamma Gibbs sampler for Bayesian logistic analysis of variance

Analysis of Polya-Gamma Gibbs sampler for Bayesian logistic analysis of variance Electronic Journal of Statistics Vol. (207) 326 337 ISSN: 935-7524 DOI: 0.24/7-EJS227 Analysis of Polya-Gamma Gibbs sampler for Bayesian logistic analysis of variance Hee Min Choi Department of Statistics

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications

Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications J.P Bouchaud with: M. Potters, G. Biroli, L. Laloux, M. A. Miceli http://www.cfm.fr Portfolio theory: Basics Portfolio

More information

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix Infinite-State Markov-switching for Dynamic Volatility Models : Web Appendix Arnaud Dufays 1 Centre de Recherche en Economie et Statistique March 19, 2014 1 Comparison of the two MS-GARCH approximations

More information

Econometrics I, Estimation

Econometrics I, Estimation Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the

More information

Physics 342 Lecture 23. Radial Separation. Lecture 23. Physics 342 Quantum Mechanics I

Physics 342 Lecture 23. Radial Separation. Lecture 23. Physics 342 Quantum Mechanics I Physics 342 Lecture 23 Radial Separation Lecture 23 Physics 342 Quantum Mechanics I Friday, March 26th, 2010 We begin our spherical solutions with the simplest possible case zero potential. Aside from

More information

10. Exchangeability and hierarchical models Objective. Recommended reading

10. Exchangeability and hierarchical models Objective. Recommended reading 10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.

More information

Generalized linear models

Generalized linear models 3 Generalisation of linear models Metropolis Hastings algorithms The Probit Model The logit model Loglinear models 178/459 Generalisation of linear models Generalisation of Linear Models Linear models

More information

Power Series. Part 1. J. Gonzalez-Zugasti, University of Massachusetts - Lowell

Power Series. Part 1. J. Gonzalez-Zugasti, University of Massachusetts - Lowell Power Series Part 1 1 Power Series Suppose x is a variable and c k & a are constants. A power series about x = 0 is c k x k A power series about x = a is c k x a k a = center of the power series c k =

More information

Modelling geoadditive survival data

Modelling geoadditive survival data Modelling geoadditive survival data Thomas Kneib & Ludwig Fahrmeir Department of Statistics, Ludwig-Maximilians-University Munich 1. Leukemia survival data 2. Structured hazard regression 3. Mixed model

More information

6.1 Moment Generating and Characteristic Functions

6.1 Moment Generating and Characteristic Functions Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,

More information

Control Variates for Markov Chain Monte Carlo

Control Variates for Markov Chain Monte Carlo Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability

More information

Exercises Chapter 4 Statistical Hypothesis Testing

Exercises Chapter 4 Statistical Hypothesis Testing Exercises Chapter 4 Statistical Hypothesis Testing Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans December 5, 013 Christophe Hurlin (University of Orléans) Advanced Econometrics

More information

Flat and multimodal likelihoods and model lack of fit in curved exponential families

Flat and multimodal likelihoods and model lack of fit in curved exponential families Mathematical Statistics Stockholm University Flat and multimodal likelihoods and model lack of fit in curved exponential families Rolf Sundberg Research Report 2009:1 ISSN 1650-0377 Postal address: Mathematical

More information

New Bayesian methods for model comparison

New Bayesian methods for model comparison Back to the future New Bayesian methods for model comparison Murray Aitkin murray.aitkin@unimelb.edu.au Department of Mathematics and Statistics The University of Melbourne Australia Bayesian Model Comparison

More information

A A + B. ra + A + 1. We now want to solve the Einstein equations in the following cases:

A A + B. ra + A + 1. We now want to solve the Einstein equations in the following cases: Lecture 29: Cosmology Cosmology Reading: Weinberg, Ch A metric tensor appropriate to infalling matter In general (see, eg, Weinberg, Ch ) we may write a spherically symmetric, time-dependent metric in

More information

Data Mining and Analysis: Fundamental Concepts and Algorithms

Data Mining and Analysis: Fundamental Concepts and Algorithms Data Mining and Analysis: Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki Wagner Meira Jr. Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA Department

More information

TESTING ADS/CFT. John H. Schwarz STRINGS 2003

TESTING ADS/CFT. John H. Schwarz STRINGS 2003 TESTING ADS/CFT John H. Schwarz STRINGS 2003 July 6, 2003 1 INTRODUCTION During the past few years 1 Blau et al. constructed a maximally supersymmetric plane-wave background of type IIB string theory as

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

1 Isotropic Covariance Functions

1 Isotropic Covariance Functions 1 Isotropic Covariance Functions Let {Z(s)} be a Gaussian process on, ie, a collection of jointly normal random variables Z(s) associated with n-dimensional locations s The joint distribution of {Z(s)}

More information

1 Directional Derivatives and Differentiability

1 Directional Derivatives and Differentiability Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=

More information

Introduction to Bayesian learning Lecture 2: Bayesian methods for (un)supervised problems

Introduction to Bayesian learning Lecture 2: Bayesian methods for (un)supervised problems Introduction to Bayesian learning Lecture 2: Bayesian methods for (un)supervised problems Anne Sabourin, Ass. Prof., Telecom ParisTech September 2017 1/78 1. Lecture 1 Cont d : Conjugate priors and exponential

More information

Markov Chain Monte Carlo Methods

Markov Chain Monte Carlo Methods Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Technical Vignette 5: Understanding intrinsic Gaussian Markov random field spatial models, including intrinsic conditional autoregressive models

Technical Vignette 5: Understanding intrinsic Gaussian Markov random field spatial models, including intrinsic conditional autoregressive models Technical Vignette 5: Understanding intrinsic Gaussian Markov random field spatial models, including intrinsic conditional autoregressive models Christopher Paciorek, Department of Statistics, University

More information

ECE 275A Homework 7 Solutions

ECE 275A Homework 7 Solutions ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator

More information

Overall Objective Priors

Overall Objective Priors Overall Objective Priors Jim Berger, Jose Bernardo and Dongchu Sun Duke University, University of Valencia and University of Missouri Recent advances in statistical inference: theory and case studies University

More information

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently

More information

Stable Limit Laws for Marginal Probabilities from MCMC Streams: Acceleration of Convergence

Stable Limit Laws for Marginal Probabilities from MCMC Streams: Acceleration of Convergence Stable Limit Laws for Marginal Probabilities from MCMC Streams: Acceleration of Convergence Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham NC 778-5 - Revised April,

More information

Computation of the scattering amplitude in the spheroidal coordinates

Computation of the scattering amplitude in the spheroidal coordinates Computation of the scattering amplitude in the spheroidal coordinates Takuya MINE Kyoto Institute of Technology 12 October 2015 Lab Seminar at Kochi University of Technology Takuya MINE (KIT) Spheroidal

More information

Lecture 5: Spatial probit models. James P. LeSage University of Toledo Department of Economics Toledo, OH

Lecture 5: Spatial probit models. James P. LeSage University of Toledo Department of Economics Toledo, OH Lecture 5: Spatial probit models James P. LeSage University of Toledo Department of Economics Toledo, OH 43606 jlesage@spatial-econometrics.com March 2004 1 A Bayesian spatial probit model with individual

More information

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary

More information

PART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics

PART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics Table of Preface page xi PART I INTRODUCTION 1 1 The meaning of probability 3 1.1 Classical definition of probability 3 1.2 Statistical definition of probability 9 1.3 Bayesian understanding of probability

More information

The Polya-Gamma Gibbs Sampler for Bayesian. Logistic Regression is Uniformly Ergodic

The Polya-Gamma Gibbs Sampler for Bayesian. Logistic Regression is Uniformly Ergodic he Polya-Gamma Gibbs Sampler for Bayesian Logistic Regression is Uniformly Ergodic Hee Min Choi and James P. Hobert Department of Statistics University of Florida August 013 Abstract One of the most widely

More information

Review and continuation from last week Properties of MLEs

Review and continuation from last week Properties of MLEs Review and continuation from last week Properties of MLEs As we have mentioned, MLEs have a nice intuitive property, and as we have seen, they have a certain equivariance property. We will see later that

More information

Appendix A Vector Analysis

Appendix A Vector Analysis Appendix A Vector Analysis A.1 Orthogonal Coordinate Systems A.1.1 Cartesian (Rectangular Coordinate System The unit vectors are denoted by x, ŷ, ẑ in the Cartesian system. By convention, ( x, ŷ, ẑ triplet

More information

STA 294: Stochastic Processes & Bayesian Nonparametrics

STA 294: Stochastic Processes & Bayesian Nonparametrics MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a

More information

Bayesian Estimation with Sparse Grids

Bayesian Estimation with Sparse Grids Bayesian Estimation with Sparse Grids Kenneth L. Judd and Thomas M. Mertens Institute on Computational Economics August 7, 27 / 48 Outline Introduction 2 Sparse grids Construction Integration with sparse

More information

Measuring nuisance parameter effects in Bayesian inference

Measuring nuisance parameter effects in Bayesian inference Measuring nuisance parameter effects in Bayesian inference Alastair Young Imperial College London WHOA-PSI-2017 1 / 31 Acknowledgements: Tom DiCiccio, Cornell University; Daniel Garcia Rasines, Imperial

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Chapter 8 Maximum Likelihood Estimation 8. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function

More information

Visibility estimates in Euclidean and hyperbolic germ-grain models and line tessellations

Visibility estimates in Euclidean and hyperbolic germ-grain models and line tessellations Visibility estimates in Euclidean and hyperbolic germ-grain models and line tessellations Pierre Calka Lille, Stochastic Geometry Workshop, 30 March 2011 Outline Visibility in the vacancy of the Boolean

More information

An exponential family of distributions is a parametric statistical model having densities with respect to some positive measure λ of the form.

An exponential family of distributions is a parametric statistical model having densities with respect to some positive measure λ of the form. Stat 8112 Lecture Notes Asymptotics of Exponential Families Charles J. Geyer January 23, 2013 1 Exponential Families An exponential family of distributions is a parametric statistical model having densities

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School

More information

Bridge estimation of the probability density at a point. July 2000, revised September 2003

Bridge estimation of the probability density at a point. July 2000, revised September 2003 Bridge estimation of the probability density at a point Antonietta Mira Department of Economics University of Insubria Via Ravasi 2 21100 Varese, Italy antonietta.mira@uninsubria.it Geoff Nicholls Department

More information

Chapter 3. Point Estimation. 3.1 Introduction

Chapter 3. Point Estimation. 3.1 Introduction Chapter 3 Point Estimation Let (Ω, A, P θ ), P θ P = {P θ θ Θ}be probability space, X 1, X 2,..., X n : (Ω, A) (IR k, B k ) random variables (X, B X ) sample space γ : Θ IR k measurable function, i.e.

More information

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45 Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS 21 June 2010 9:45 11:45 Answer any FOUR of the questions. University-approved

More information

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis Lecture 3 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,

More information

The square root rule for adaptive importance sampling

The square root rule for adaptive importance sampling The square root rule for adaptive importance sampling Art B. Owen Stanford University Yi Zhou January 2019 Abstract In adaptive importance sampling, and other contexts, we have unbiased and uncorrelated

More information