Zero Variance Markov Chain Monte Carlo for Bayesian Estimators
|
|
- Philomena Sims
- 6 years ago
- Views:
Transcription
1 Noname manuscript No. will be inserted by the editor Zero Variance Markov Chain Monte Carlo for Bayesian Estimators Supplementary material Antonietta Mira Reza Solgi Daniele Imparato Received: date / Accepted: date Appendix A: Probit model Mathematical formulation Let y i be Bernoulli r.v. s: y i x i B, p i, p i = Φx T i β, where β Rd is the vector of parameters of the model and Φ is the c.d.f. of a standard normal distribution. The likelihood function is: lβ y, x [ Φx T i β ] y i [ Φx T i β ] y i. As it can be seen by inspection, the likelihood function is invariant under the transformation x i, y i x i, y i. Therefore, for the sake of simplicity, in the rest of the example we assume y i = for any i, so that the likelihood simplifies: lβ y, x Φx T i β. This formula shows that the contribution of x i = 0 is just a constant Φx T i β = Φ0 =, therefore, without loss of generality, we assume for all i, x i 0. A. Mira Swiss Finance Institute, University of Lugano, via Buffi 3 CH-6904 Lugano, Switzerland. antonietta.mira@usi.ch R. Solgi Swiss Finance Institute, University of Lugano, via Buffi 3 CH-6904 Lugano, Switzerland. reza.solgi@usi.ch D. Imparato Department of Economics, University of Insubria, via Monte Generoso 7, 00 Varese, Italy. daniele.imparato@uninsubria.it
2 Antonietta Mira et al. Using flat priors, the posterior of the model is proportional to the likelihood, and the Bayesian estimator of each parameter, β k, is the expected value of f k β = β k under π k =,,, d. Using Schrödinger-type Hamiltonians, H and ψ k β = P k β πβ, as the trial functions, where P k β = d j= a j,kβ j is a first degree polynomial, one gets: f k β = f k β + where, for j =,,..., d, Hψ kβ πβ y, x = f k β + z j = n because of the assumption y i = for any i. x ij φx T i β Φx T i β, d a j,k z j, j= Central limit theorem In the following, it is supposed that P is a linear polynomial. In the Probit model, the ZV-MCMC estimators obey a CLT if z j have finite + δ moment under π, for some δ > 0: [ E π zj +δ] = c E π n x ij φx T i β +δ Φx T i β = c c n x ij φx T i β Φx T i β +δ n Φx T i βdβ <. where c = δ, and c is the normalizing constant of π the target posterior. Define: and therefore: n x ij φx T i K β = β Φx T i β K β = Φx T i β, Kβ = K βk β +δ [ E π zj +δ] = c K βk βdβ. where c = c c. Before studying the convergence of this integral, the following property of the likelihood for the probit model is needed.,
3 Zero Variance Markov Chain Monte Carlo for Bayesian Estimators 3 Proposition Existence and uniqueness of MLE implies that, for any β 0 \ {0}, there exists i such that x T i β 0 < 0. Proof by contradiction. Uniqueness of MLE implies that x T x is full rank, that is, there is no β 0 orthogonal to all observations x i. This can be seen by contradiction: singularity of x T x implies existence of a non-zero β 0 orthogonal to all observations x i. This fact, in turn, implies lβ x, y = lβ + cβ 0 x, y and, therefore, l x, y does not have a unique global maximum. Next, assume there exists some β 0 such that, for any i, x T i β 0 > 0. Then β 0 is a direction of recession for the negative log-likelihood function n ln ΦxT i β that is a proper closed convex function. This implies that this function does not have non-empty bounded minimum set [], which means that the MLE does not exist. Now, rewriting Kβdβ in hyper-spherical coordinates through the bijective transformation ρ, θ,..., θ d := F β, where F is defined as β = ρ cosθ β l = ρ cosθ l l m= sinθ m, for l =,..., d β d = ρ d m= sinθ m, for θ Θ := {0 θ i π, i =,..., d, 0 θ d < π} and ρ > 0, one gets Kβdβ = := + Θ 0 + Θ Θ 0 Aθdθ, d KF ρ, θρ d sin d j θ j dρdθ. j= KF ρ, θρ d dρdθ Observe that the integrand is well defined for any ρ, θ on the domain of integration, so it is enough to study its asymptotic behaviour when ρ goes to infinity, and θ Θ. First, analyze K F n ρ, θ = x ij φ x i ρλ i θ Φ x i ρλ i θ where, for any i, λ i is a suitable function of the angles θ such that λ i [, ], which takes into account the sign of the scalar product in the original coordinates system. For any i, when ρ if λ i < 0, xijφ xi ρλi Φ x i ρλ i O ρ; +δ,
4 4 Antonietta Mira et al. if λ i > 0, xijφ xi ρλi Φ x i ρλ i if λ i = 0, xijφ xi ρλi Φ x i ρλ i Therefore: O φλ i ρ; = x ij π O. n x ij φ x i ρλ i Φ x i ρλ i O ρ and, for any θ Θ: K F ρ, θ O ρ +δ. Now, focus on K F ρ, θ = n Φ x i ρλ i θ; existence of MLE for the probit model implies that, for any θ Θ, there exists some l l n, such that λ l θ < 0, and therefore: K F ρ, θ < Φ x l ρλ l O φλ l ρ ρ. Putting these results together leads to KF ρ, θ = K F ρ, θk F ρ, θ O ρ +δ φλ l θρ so that, for any θ Θ, KF ρ, θρ d O ρ +δ+d φ λ l θρ, ρ +. Therefore, whenever the value θ Θ, its integrand converges to zero rapidly enough when ρ +. This concludes the proof. Note. In [] it is shown that the existence of the posterior under flat priors for probit and logit models is equivalent to the existence and finiteness of MLE. This ensures us that the posterior is well defined in our context. In order to verify the existence of the posterior mean, we can use a simplified version of the proof given above. In other words we should show βj n ΦxT i βdβ < +, that is, K β = β j Oρ. Therefore, a weaker version of the proof given above can be employed. Note. In the proof given above we have used flat priors: although this assumption simplifies the proof, however a very similar proof can be applied for non-flat priors. Assume the prior is π 0 β. Under this assumption the posterior is πβ = π 0 β lβ y, x π 0 β n β and the control variates are: x ijφx T i β Φx T i β ΦxT i z j = d ln π 0β dβ j n. Therefore we need to prove the + δ-th moment of z j under πβ is finite. A sufficient condition for this is the finiteness of + δ-th moments of d ln π 0β dβ j and n x ijφx T i β Φx T i β under πβ. If we assume π 0 β is bounded above, the latter is a trivial consequence of the proof given above for the flat priors. Therefore we only need to prove the finiteness of the integral d ln π 0 β dβ j +δ π 0 β Φx T i βdβ.
5 Zero Variance Markov Chain Monte Carlo for Bayesian Estimators 5 Again if we assume the prior is bounded from above, a sufficient condition for the existence of this integral is the existence of the following integral: +δ d ln π 0 β n dβ j Φx T i βdβ A proof very similar to the one given above will show that this integral is finite for common choices of priors π 0 β such as Normal, Student s T, etc. Appendix B: Logit model Mathematical formulation In the same setting as the probit model, let p i = expxt i β +expx T i β where β Rd is the vector of parameters of the model. The likelihood function is: lβ y, x expx T i β + expx T i β yi + expx T i β yi. 3 By inspection, it is easy to verify that the likelihood function is invariant under the transformation:x i, y i x i, y i. Therefore, for the sake of simplicity, in the sequel we assume y i = 0 for any i, so that the likelihood simplifies as: lβ y, x + expx T i β. The contribution of x i = 0 to the likelihood is just a constant, therefore, without loss of generality, it is assumed that x i 0 for all i. Using flat priors, the posterior distribution is proportional to 3 and the Bayesian estimator of each parameter, β k, is the expected value of f k β = β k under π k =,,, d. Using the same pair of operator H and test function ψ k as in Appendinx A, the control variates are: z j = n expx T i x β ij + expx T for j =,,..., d. i β, Central limit theorem As for the probit model, the ZV-MCMC estimators obey a CLT if the control variates z j have finite + δ moment under π, for some δ > 0 : [ ] n +δ E π z +δ j = c E π expx T i x β ij + expx T i β = c c n +δ expx T i x β n ij + expx T i β + expx T i βdβ
6 6 Antonietta Mira et al. where c = δ, and c is the normalizing constant of π. The finitiness of the integral is, indeed, trivial. Observe that the function e y / + e y is bounded by ; therefore, E π [ z +δ j ] n +δ x ij <. As for probit model, note that it can easily be shown that the posterior means exist under flat priors. Moreover, a very similar proof see Note of Appendix A can be used under a normal or Student s T prior distribution. Appendix C: GARCH model Mathematical formulation We assume that the returns are conditionally normal distributed, rt F t N 0, h t, where h t is a predictable F t measurable process: h t = ω + ω 3 h t + ω r t, where ω > 0, ω 0, and ω 3 0. Let r = r,..., r T be the observed time series. The likelihood function is equal to: l ω, ω, ω 3 r T t= h t exp and using independent truncated Normal priors for the parameters, the posterior is: [ π ω, ω, ω 3 r exp ] ω σ ω + ω σ ω + ω T 3 σ h t exp ω 3 t= Therefore, the control variates when the trial function is a first degree polynomial are: t= r t h t t= r t h t. ln π ω i = ω i σ ω i h t= t h t ω i + r t h t= t h t ω i, i =,, 3, where: h t = ωt 3, ω ω 3 h t = r h t t + ω 3 I t>, ω ω h t h t = h t + ω 3 I t>. ω 3 ω 3
7 Zero Variance Markov Chain Monte Carlo for Bayesian Estimators 7 Central limit theorem In order to prove the CLT for the ZV-MCMC estimator in the Garch model, we need: ln π ln π ln π,, L +δ π. 4 ω ω 3 ω To this end, h t and its partial derivatives should be expressed as a function of h 0 and r: t t h t = ω + ω3 k + ω3h t 0 + ω + ω3 k rt k, k= ln h t ω ln h t ω = ln h t ω 3 = = ω t 3 ω 3 I {t>}, t rt + t h t + k= ω t j 3 rj j=0 ω t j 3 h j j=0 I {t>}, I {t>}. Next, moving to spherical coordinates, the integral 4 can be written as K j ρ; θ, φdρdθdφ := A j θ, φdθdφ, [0,π/] 0 [0,π/] where, for j =,, 3, K j ; θ, φ = W j +δ W, with W = σ ω ρ cos θ sin φ W = σ ω ρ sin θ sin φ W 3 = σ ω ρ cos φ 3 t= h t W = exp and ρ sin θ t= h t t= h t r t h t r t ρ t cos t φ, h ρ cos φ t r t r h t t ht + t t + x t j 3 rj j=0 j=0 x t j 3 h j, ρ σ ω cos θ sin φ + σ ω sin θ sin φ + σ ω 3 cos φ T t= h t t t h t = ρ cos θ sin φ +ρ k cos k φ+h 0 ρ t cos t φ+ρ sin θ sin φ +rt kρ k cos k φ. k= k=, 5 t= r t h t
8 8 Antonietta Mira et al. The aim is to prove that, for any θ, φ [0, π/] and for any j, A j θ, φ is finite. To this end, the convergence of A j for any θ, φ should be discussed. Let us study the proper domain of K j ; θ, φ. Observe that K j ; θ, φ is not defined whenever h t = 0 and, if j = and φ π/, also for ρ = / cos φ. However, the discontinuity of K 3 at this point is removable, so that the domain of K 3 can be extended by continuity also at ρ = / cos φ. Since h t = 0 if and only if ρ = 0, it can be concluded that, for any j and for any θ, φ [0, π/], the proper domain of K j ; θ, φ is domk j ; θ, φ = 0+. By fixing the value of θ and φ, let us study the limits of K j when ρ 0 and ρ +. Observe that, whatever the values of θ and φ are, W j s are rationale functions of ρ. Therefore, for any j, W j +δ cannot grow towards infinity more than polynomially at the boundary of the domain. On the other hand, W goes to zero with an exponential rate both when ρ 0 and ρ +, for any θ and φ. This is sufficient to conclude that, for any θ, φ [0, π/], the integral A j is finite for j =,, 3 and, therefore, condition 4 holds and the ZV estimators for the GARCH model obeys a CLT. Appendix D: unbiasedness In this Appendix, explicit computations are presented, which were omitted in Section 5. Moreover, it is proved that all the ZV-MCMC estimators discussed in Section 8 are unbiased. Following the same notations as in Section 5, equation 9 follows because Hψ := π = Ω Ω Hψ π V ψ π ψ π = V πψ π ψ ndσ + π ψ Ω Ω Ω = V πψ π ψ ndσ + ψ π ndσ Ω Ω Ω = ΩH πψ + [ψ π π ψ] ndσ Ω = [ψ π π ψ] ndσ. Ω Ω ψ π Then, Therefore, Hψ = 0 if ψ π = π ψ on Ω. Now, let ψ = P π. π ψ = π P + P π π,
9 Zero Variance Markov Chain Monte Carlo for Bayesian Estimators 9 so that Hψ π = 0 if P x πx = 0, x Ω, j =,..., d. x j When π has unbounded support, following the previous computations integrating over the bounded set B r and taking the limit for r, one gets Hψ = π lim π P ndσ. 6 r + B r Therefore, unbiasedness in the unbounded case is reached if the limit appearing in the right-hand side of 6 is zero. Now, the unbiasedness of the ZV-MCMC estimators exploited in Section 8 is discussed. To this end, condition 6 should be verified. Let B ρ be a hypersphere of radius ρ and let n := ρβ be its normal versor. Then, for linear P, 6 equals zero if, for any j =,..., d, lim ρ + ρ B ρ πββ j ds = 0. 7 The Probit model is first considered. By using the same notations as in Appendix A, the integral in 7 is proportional to lim K F ρ, θρ d dθ. 8 ρ + ρ Note that, because of, there exist ρ 0 and M such that K F ρ, θρ d Mφλ lθ ρρ d Θ Mφλ lθ ρ 0 ρ d 0 := Gθ ρ ρ 0. Since Gθ L, by the dominated convergence theorem a sufficient condition to get unbiasedness is lim K F ρ, θρ d = 0, 9 ρ + which is true, because of for the Probit model. We now consider the Logit model for which it is easy to prove that Proposition holds. As done for the Probit model, one can write [ ] E π z +δ j K βk βdβ, where K β = K β = n +δ expx T i x β ij + expx T i β, + expx T i β,
10 0 Antonietta Mira et al. By using the hyper-spherical change of variables in, we get [ ] E π z +δ j K F ρ, θk F ρ, θρ d dρ : Θ 0 and, as for the Probit model, we must verify Equation 8. Now analyze K F ρ, θ; for any θ, existence of MLE implies the existence of some l l n, such that λ l θ > 0, and therefore: K F ρ, θ = Therefore, there exist ρ 0, M such that K F ρ, θρ d M exp λ lθ ρρ d + exp x i ρλ i < + exp x l ρλ l O exp ρλ l. 0 M exp λ lθ ρ 0 ρ d 0 := Gθ ρ ρ 0, where Gθ L. These computations allow us to use Equation 9 as a sufficient condition to get unbiasedness, and its proof becomes trivial. Finally consider the GARCH model. In this case, B ρ is the portion of a sphere of radius ρ defined on the positive orthant. Then, the limit lim ρ + ρ [0,π/] W F ρ, θρ d dθ, where W was defined in 5, should be discussed. Again, an application of the dominated convergence theorem leads to the simpler condition lim W F ρ, θρ = 0, ρ + which is true, since W decays with an exponential rate. References. Rockafellar, R.: Convex analysis, pp Princeton University Press 970. Speckman P.L. Lee, J., Sun, D.: Existence of the mle and propriety of posteriors for a general multinomial choice model. Statistica Sinica 9,
Stat 5421 Lecture Notes Proper Conjugate Priors for Exponential Families Charles J. Geyer March 28, 2016
Stat 5421 Lecture Notes Proper Conjugate Priors for Exponential Families Charles J. Geyer March 28, 2016 1 Theory This section explains the theory of conjugate priors for exponential families of distributions,
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationPart III. A Decision-Theoretic Approach and Bayesian testing
Part III A Decision-Theoretic Approach and Bayesian testing 1 Chapter 10 Bayesian Inference as a Decision Problem The decision-theoretic framework starts with the following situation. We would like to
More informationParameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn
Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationA Very Brief Summary of Bayesian Inference, and Examples
A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X
More informationZig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017
Zig-Zag Monte Carlo Delft University of Technology Joris Bierkens February 7, 2017 Joris Bierkens (TU Delft) Zig-Zag Monte Carlo February 7, 2017 1 / 33 Acknowledgements Collaborators Andrew Duncan Paul
More informationBayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence
Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns
More informationEM Algorithm II. September 11, 2018
EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data
More informationBayesian Modeling of Conditional Distributions
Bayesian Modeling of Conditional Distributions John Geweke University of Iowa Indiana University Department of Economics February 27, 2007 Outline Motivation Model description Methods of inference Earnings
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationGenerative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis
Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis Stéphanie Allassonnière CIS, JHU July, 15th 28 Context : Computational Anatomy Context and motivations :
More informationParameter Estimation
Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y
More informationLasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices
Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,
More informationHIERARCHICAL BAYESIAN ANALYSIS OF BINARY MATCHED PAIRS DATA
Statistica Sinica 1(2), 647-657 HIERARCHICAL BAYESIAN ANALYSIS OF BINARY MATCHED PAIRS DATA Malay Ghosh, Ming-Hui Chen, Atalanta Ghosh and Alan Agresti University of Florida, Worcester Polytechnic Institute
More information1. Fisher Information
1. Fisher Information Let f(x θ) be a density function with the property that log f(x θ) is differentiable in θ throughout the open p-dimensional parameter set Θ R p ; then the score statistic (or score
More informationBayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference
1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE
More informationFast Maximum Likelihood estimation via Equilibrium Expectation for Large Network Data
Fast Maximum Likelihood estimation via Equilibrium Expectation for Large Network Data Maksym Byshkin 1, Alex Stivala 4,1, Antonietta Mira 1,3, Garry Robins 2, Alessandro Lomi 1,2 1 Università della Svizzera
More informationMath 350 Solutions for Final Exam Page 1. Problem 1. (10 points) (a) Compute the line integral. F ds C. z dx + y dy + x dz C
Math 35 Solutions for Final Exam Page Problem. ( points) (a) ompute the line integral F ds for the path c(t) = (t 2, t 3, t) with t and the vector field F (x, y, z) = xi + zj + xk. (b) ompute the line
More informationSeparation of Variables in Polar and Spherical Coordinates
Separation of Variables in Polar and Spherical Coordinates Polar Coordinates Suppose we are given the potential on the inside surface of an infinitely long cylindrical cavity, and we want to find the potential
More informationLecture 7 October 13
STATS 300A: Theory of Statistics Fall 2015 Lecture 7 October 13 Lecturer: Lester Mackey Scribe: Jing Miao and Xiuyuan Lu 7.1 Recap So far, we have investigated various criteria for optimal inference. We
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationNotes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed
18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,
More informationLikelihood-free MCMC
Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of
More informationBohr & Wheeler Fission Theory Calculation 4 March 2009
Bohr & Wheeler Fission Theory Calculation 4 March 9 () Introduction The goal here is to reproduce the calculation of the limiting Z /A against spontaneous fission Z A lim a S. (.) a C as first done by
More informationMath 1B Final Exam, Solution. Prof. Mina Aganagic Lecture 2, Spring (6 points) Use substitution and integration by parts to find:
Math B Final Eam, Solution Prof. Mina Aganagic Lecture 2, Spring 20 The eam is closed book, apart from a sheet of notes 8. Calculators are not allowed. It is your responsibility to write your answers clearly..
More informationBayesian Multivariate Logistic Regression
Bayesian Multivariate Logistic Regression Sean M. O Brien and David B. Dunson Biostatistics Branch National Institute of Environmental Health Sciences Research Triangle Park, NC 1 Goals Brief review of
More informationλ(x + 1)f g (x) > θ 0
Stat 8111 Final Exam December 16 Eleven students took the exam, the scores were 92, 78, 4 in the 5 s, 1 in the 4 s, 1 in the 3 s and 3 in the 2 s. 1. i) Let X 1, X 2,..., X n be iid each Bernoulli(θ) where
More informationBayesian inference for multivariate skew-normal and skew-t distributions
Bayesian inference for multivariate skew-normal and skew-t distributions Brunero Liseo Sapienza Università di Roma Banff, May 2013 Outline Joint research with Antonio Parisi (Roma Tor Vergata) 1. Inferential
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationChapter 4: Modelling
Chapter 4: Modelling Exchangeability and Invariance Markus Harva 17.10. / Reading Circle on Bayesian Theory Outline 1 Introduction 2 Models via exchangeability 3 Models via invariance 4 Exercise Statistical
More informationHANDBOOK OF APPLICABLE MATHEMATICS
HANDBOOK OF APPLICABLE MATHEMATICS Chief Editor: Walter Ledermann Volume VI: Statistics PART A Edited by Emlyn Lloyd University of Lancaster A Wiley-Interscience Publication JOHN WILEY & SONS Chichester
More informationMonte Carlo in Bayesian Statistics
Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview
More informationLecture 8: The Metropolis-Hastings Algorithm
30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:
More informationStat 451 Lecture Notes Monte Carlo Integration
Stat 451 Lecture Notes 06 12 Monte Carlo Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 23 in Lange, and Chapters 3 4 in Robert & Casella 2 Updated:
More informationDavid Giles Bayesian Econometrics
David Giles Bayesian Econometrics 1. General Background 2. Constructing Prior Distributions 3. Properties of Bayes Estimators and Tests 4. Bayesian Analysis of the Multiple Regression Model 5. Bayesian
More informationWeek 2 Notes, Math 865, Tanveer
Week 2 Notes, Math 865, Tanveer 1. Incompressible constant density equations in different forms Recall we derived the Navier-Stokes equation for incompressible constant density, i.e. homogeneous flows:
More informationInvariant HPD credible sets and MAP estimators
Bayesian Analysis (007), Number 4, pp. 681 69 Invariant HPD credible sets and MAP estimators Pierre Druilhet and Jean-Michel Marin Abstract. MAP estimators and HPD credible sets are often criticized in
More informationLecture 26: Likelihood ratio tests
Lecture 26: Likelihood ratio tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f θ0 (X) > c 0 for
More informationIf there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell,
Recall The Neyman-Pearson Lemma Neyman-Pearson Lemma: Let Θ = {θ 0, θ }, and let F θ0 (x) be the cdf of the random vector X under hypothesis and F θ (x) be its cdf under hypothesis. Assume that the cdfs
More informationKazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies. Abstract
Bayesian Estimation of A Distance Functional Weight Matrix Model Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies Abstract This paper considers the distance functional weight
More informatione x3 dx dy. 0 y x 2, 0 x 1.
Problem 1. Evaluate by changing the order of integration y e x3 dx dy. Solution:We change the order of integration over the region y x 1. We find and x e x3 dy dx = y x, x 1. x e x3 dx = 1 x=1 3 ex3 x=
More informationSupporting Information
Supporting Information A: Calculation of radial distribution functions To get an effective propagator in one dimension, we first transform 1) into spherical coordinates: x a = ρ sin θ cos φ, y = ρ sin
More informationLecture 17: Likelihood ratio and asymptotic tests
Lecture 17: Likelihood ratio and asymptotic tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f
More informationParametric Techniques Lecture 3
Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to
More informationNuisance parameters and their treatment
BS2 Statistical Inference, Lecture 2, Hilary Term 2008 April 2, 2008 Ancillarity Inference principles Completeness A statistic A = a(x ) is said to be ancillary if (i) The distribution of A does not depend
More informationMARGINAL MARKOV CHAIN MONTE CARLO METHODS
Statistica Sinica 20 (2010), 1423-1454 MARGINAL MARKOV CHAIN MONTE CARLO METHODS David A. van Dyk University of California, Irvine Abstract: Marginal Data Augmentation and Parameter-Expanded Data Augmentation
More informationLecture 14 More on structural estimation
Lecture 14 More on structural estimation Economics 8379 George Washington University Instructor: Prof. Ben Williams traditional MLE and GMM MLE requires a full specification of a model for the distribution
More informationTEORIA BAYESIANA Ralph S. Silva
TEORIA BAYESIANA Ralph S. Silva Departamento de Métodos Estatísticos Instituto de Matemática Universidade Federal do Rio de Janeiro Sumário Numerical Integration Polynomial quadrature is intended to approximate
More informationσ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =
Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,
More informationAnalysis of Polya-Gamma Gibbs sampler for Bayesian logistic analysis of variance
Electronic Journal of Statistics Vol. (207) 326 337 ISSN: 935-7524 DOI: 0.24/7-EJS227 Analysis of Polya-Gamma Gibbs sampler for Bayesian logistic analysis of variance Hee Min Choi Department of Statistics
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More informationRandom Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications
Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications J.P Bouchaud with: M. Potters, G. Biroli, L. Laloux, M. A. Miceli http://www.cfm.fr Portfolio theory: Basics Portfolio
More informationInfinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix
Infinite-State Markov-switching for Dynamic Volatility Models : Web Appendix Arnaud Dufays 1 Centre de Recherche en Economie et Statistique March 19, 2014 1 Comparison of the two MS-GARCH approximations
More informationEconometrics I, Estimation
Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the
More informationPhysics 342 Lecture 23. Radial Separation. Lecture 23. Physics 342 Quantum Mechanics I
Physics 342 Lecture 23 Radial Separation Lecture 23 Physics 342 Quantum Mechanics I Friday, March 26th, 2010 We begin our spherical solutions with the simplest possible case zero potential. Aside from
More information10. Exchangeability and hierarchical models Objective. Recommended reading
10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.
More informationGeneralized linear models
3 Generalisation of linear models Metropolis Hastings algorithms The Probit Model The logit model Loglinear models 178/459 Generalisation of linear models Generalisation of Linear Models Linear models
More informationPower Series. Part 1. J. Gonzalez-Zugasti, University of Massachusetts - Lowell
Power Series Part 1 1 Power Series Suppose x is a variable and c k & a are constants. A power series about x = 0 is c k x k A power series about x = a is c k x a k a = center of the power series c k =
More informationModelling geoadditive survival data
Modelling geoadditive survival data Thomas Kneib & Ludwig Fahrmeir Department of Statistics, Ludwig-Maximilians-University Munich 1. Leukemia survival data 2. Structured hazard regression 3. Mixed model
More information6.1 Moment Generating and Characteristic Functions
Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,
More informationControl Variates for Markov Chain Monte Carlo
Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability
More informationExercises Chapter 4 Statistical Hypothesis Testing
Exercises Chapter 4 Statistical Hypothesis Testing Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans December 5, 013 Christophe Hurlin (University of Orléans) Advanced Econometrics
More informationFlat and multimodal likelihoods and model lack of fit in curved exponential families
Mathematical Statistics Stockholm University Flat and multimodal likelihoods and model lack of fit in curved exponential families Rolf Sundberg Research Report 2009:1 ISSN 1650-0377 Postal address: Mathematical
More informationNew Bayesian methods for model comparison
Back to the future New Bayesian methods for model comparison Murray Aitkin murray.aitkin@unimelb.edu.au Department of Mathematics and Statistics The University of Melbourne Australia Bayesian Model Comparison
More informationA A + B. ra + A + 1. We now want to solve the Einstein equations in the following cases:
Lecture 29: Cosmology Cosmology Reading: Weinberg, Ch A metric tensor appropriate to infalling matter In general (see, eg, Weinberg, Ch ) we may write a spherically symmetric, time-dependent metric in
More informationData Mining and Analysis: Fundamental Concepts and Algorithms
Data Mining and Analysis: Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki Wagner Meira Jr. Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA Department
More informationTESTING ADS/CFT. John H. Schwarz STRINGS 2003
TESTING ADS/CFT John H. Schwarz STRINGS 2003 July 6, 2003 1 INTRODUCTION During the past few years 1 Blau et al. constructed a maximally supersymmetric plane-wave background of type IIB string theory as
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More information1 Isotropic Covariance Functions
1 Isotropic Covariance Functions Let {Z(s)} be a Gaussian process on, ie, a collection of jointly normal random variables Z(s) associated with n-dimensional locations s The joint distribution of {Z(s)}
More information1 Directional Derivatives and Differentiability
Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=
More informationIntroduction to Bayesian learning Lecture 2: Bayesian methods for (un)supervised problems
Introduction to Bayesian learning Lecture 2: Bayesian methods for (un)supervised problems Anne Sabourin, Ass. Prof., Telecom ParisTech September 2017 1/78 1. Lecture 1 Cont d : Conjugate priors and exponential
More informationMarkov Chain Monte Carlo Methods
Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)
More informationBayesian linear regression
Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationTechnical Vignette 5: Understanding intrinsic Gaussian Markov random field spatial models, including intrinsic conditional autoregressive models
Technical Vignette 5: Understanding intrinsic Gaussian Markov random field spatial models, including intrinsic conditional autoregressive models Christopher Paciorek, Department of Statistics, University
More informationECE 275A Homework 7 Solutions
ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator
More informationOverall Objective Priors
Overall Objective Priors Jim Berger, Jose Bernardo and Dongchu Sun Duke University, University of Valencia and University of Missouri Recent advances in statistical inference: theory and case studies University
More informationPh.D. Qualifying Exam Friday Saturday, January 3 4, 2014
Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently
More informationStable Limit Laws for Marginal Probabilities from MCMC Streams: Acceleration of Convergence
Stable Limit Laws for Marginal Probabilities from MCMC Streams: Acceleration of Convergence Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham NC 778-5 - Revised April,
More informationComputation of the scattering amplitude in the spheroidal coordinates
Computation of the scattering amplitude in the spheroidal coordinates Takuya MINE Kyoto Institute of Technology 12 October 2015 Lab Seminar at Kochi University of Technology Takuya MINE (KIT) Spheroidal
More informationLecture 5: Spatial probit models. James P. LeSage University of Toledo Department of Economics Toledo, OH
Lecture 5: Spatial probit models James P. LeSage University of Toledo Department of Economics Toledo, OH 43606 jlesage@spatial-econometrics.com March 2004 1 A Bayesian spatial probit model with individual
More informationLatent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent
Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary
More informationPART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics
Table of Preface page xi PART I INTRODUCTION 1 1 The meaning of probability 3 1.1 Classical definition of probability 3 1.2 Statistical definition of probability 9 1.3 Bayesian understanding of probability
More informationThe Polya-Gamma Gibbs Sampler for Bayesian. Logistic Regression is Uniformly Ergodic
he Polya-Gamma Gibbs Sampler for Bayesian Logistic Regression is Uniformly Ergodic Hee Min Choi and James P. Hobert Department of Statistics University of Florida August 013 Abstract One of the most widely
More informationReview and continuation from last week Properties of MLEs
Review and continuation from last week Properties of MLEs As we have mentioned, MLEs have a nice intuitive property, and as we have seen, they have a certain equivariance property. We will see later that
More informationAppendix A Vector Analysis
Appendix A Vector Analysis A.1 Orthogonal Coordinate Systems A.1.1 Cartesian (Rectangular Coordinate System The unit vectors are denoted by x, ŷ, ẑ in the Cartesian system. By convention, ( x, ŷ, ẑ triplet
More informationSTA 294: Stochastic Processes & Bayesian Nonparametrics
MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a
More informationBayesian Estimation with Sparse Grids
Bayesian Estimation with Sparse Grids Kenneth L. Judd and Thomas M. Mertens Institute on Computational Economics August 7, 27 / 48 Outline Introduction 2 Sparse grids Construction Integration with sparse
More informationMeasuring nuisance parameter effects in Bayesian inference
Measuring nuisance parameter effects in Bayesian inference Alastair Young Imperial College London WHOA-PSI-2017 1 / 31 Acknowledgements: Tom DiCiccio, Cornell University; Daniel Garcia Rasines, Imperial
More informationMaximum Likelihood Estimation
Chapter 8 Maximum Likelihood Estimation 8. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function
More informationVisibility estimates in Euclidean and hyperbolic germ-grain models and line tessellations
Visibility estimates in Euclidean and hyperbolic germ-grain models and line tessellations Pierre Calka Lille, Stochastic Geometry Workshop, 30 March 2011 Outline Visibility in the vacancy of the Boolean
More informationAn exponential family of distributions is a parametric statistical model having densities with respect to some positive measure λ of the form.
Stat 8112 Lecture Notes Asymptotics of Exponential Families Charles J. Geyer January 23, 2013 1 Exponential Families An exponential family of distributions is a parametric statistical model having densities
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School
More informationBridge estimation of the probability density at a point. July 2000, revised September 2003
Bridge estimation of the probability density at a point Antonietta Mira Department of Economics University of Insubria Via Ravasi 2 21100 Varese, Italy antonietta.mira@uninsubria.it Geoff Nicholls Department
More informationChapter 3. Point Estimation. 3.1 Introduction
Chapter 3 Point Estimation Let (Ω, A, P θ ), P θ P = {P θ θ Θ}be probability space, X 1, X 2,..., X n : (Ω, A) (IR k, B k ) random variables (X, B X ) sample space γ : Θ IR k measurable function, i.e.
More informationTwo hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45
Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS 21 June 2010 9:45 11:45 Answer any FOUR of the questions. University-approved
More informationLecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis
Lecture 3 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,
More informationThe square root rule for adaptive importance sampling
The square root rule for adaptive importance sampling Art B. Owen Stanford University Yi Zhou January 2019 Abstract In adaptive importance sampling, and other contexts, we have unbiased and uncorrelated
More information