Bayesian Empirical Likelihood: a survey
|
|
- Jeremy Clarke
- 5 years ago
- Views:
Transcription
1 Bayesian Empirical Likelihood 1 Bayesian Empirical Likelihood: a survey Art B. Owen Stanford University
2 Bayesian Empirical Likelihood 2 These are the slides I presented at BNP 11 in Paris, a very fine meeting. I have added a few interstitial slides like this one to include comments I made while presenting the slides and to give some background thoughts in this area. The session included a delightful presentation by Nils Lid Hjort on some not yet published theorems where he connects Bayesian empirical likelihood to Dirichlet process Bayesian nonparametrics. I was pleased to see our session got a positive review on Xian s Og (written by our session chair, Christian Robert). Also, the conference included two very strong posters by students working on Bayesian empirical likelihood. Frank Liu (Monash) has a nice way to robustify BETEL. Laura Turbatu (Geneva) took a close look at high order asymptotics to compare frequentist coverage of Bayes with various kinds of nonparametric priors.
3 Bayesian Empirical Likelihood 3 The goal Science driven prior Data driven likelihood
4 Bayesian Empirical Likelihood 4 This goal is not everybody s goal. But this goal means you don t have to make up a likelihood which could bring unwanted and consequential restrictions. If the data deliver the likelihood and the prior is chosen flat or by previously known science, then what is there left for the statistician to do?
5 Bayesian Empirical Likelihood 5 The goal Science driven prior Data driven likelihood The ask IID data and an estimand defined by estimating equations. X 1, X 2,..., X n iid F m(x, θ) df (x) = 0, θ R p
6 Bayesian Empirical Likelihood 6 We have IID X and an estimating equation. If you picture the plate diagram, it is one of the simplest possible ones. It covers an important set of problems, but Bayes includes much else as well.
7 Bayesian Empirical Likelihood 7 Outline 1) Empirical Likelihood 2) Bayes EL Lazar (2003) 3) Exponential tilting, and ETEL Schennach (2005,2007) 4) ABC and EL Mengersen, Pudio, Robert (2013) 5) BEL with Hamiltonian MCMC Chaudhuri et al (2017) 6) Some other recent works Notes Apologies about necessarily omitted papers. Motivation: bioinformatics.
8 Bayesian Empirical Likelihood 8 In the bioinformatics motivation, the data are counts. They are assuredly not Poisson. So a common response is to model them as negative binomial. Of course they are not that either. This is work in progress. SAMSI is having a year on QMC. The original slides touted the opening workshop. However the application deadline will almost surely have passed by the time you re reading this.
9 Bayesian Empirical Likelihood 9 Empirical likelihood For X i iid F observed Xi = x i R d, define n L(F ) = F ({x i }). i=1 There are many other NPMLEs. The nonparametric MLE (NPMLE) is F arg max F E.g., Kaplan-Meier for right censored survival, L(F ) = 1 n Log concave densities: Walther, Dumbgen, Rufibach, Cule, Samworth, Stewart n i=1 δ xi
10 Bayesian Empirical Likelihood 10 Basic notions Model: x i iid F True value: F = F0 NPMLE: F Estimand T (F ) E.g., population mean, median, regression coefficient etc. True value: θ 0 = T (F 0 ) NPMLE: ˆθ = T ( F )
11 Bayesian Empirical Likelihood 11 Empirical likelihood ratios Point est. Interval / Test Parametric ˆθ = arg maxθ R(θ) 2 log R(θ 0 ) χ 2 Non-parametric T ( F ) The purpose of empirical likelihood ratios is to fill in the, getting likelihood ratio confidence intervals and tests without parametric assumptions. I.e., a nonparametric Wilks theorem O (2000) Chapman & Hall, Monograph
12 Bayesian Empirical Likelihood 12 Let w i = w i (F ) = F ({x i }) Then w i 0 and n i=1 w i 1. Empirical likelihood ratio R(F ) = R(F ) R( F ) = n i=1 n w i 1/n = i=1 (nw i ) Literal EL confidence regions {T (F ) R(F ) c} Profile likelihood R(θ) = max{r(f ) T (F ) = θ} Conf set = {θ R(θ) c}. We still have to choose c
13 Bayesian Empirical Likelihood 13 Usual EL It is usual to force i w i = i F ({x i}) = 1. R(F ) c already implies 1 F ({x 1,..., x n }) = O(1/n) Supported on x 1,..., x n It is a multinomial likelihood There are other choices
14 Bayesian Empirical Likelihood 14 Empirical likelihood for the mean T (F ) = x df (x) { n R(µ) = max i=1 Profile likelihood nw i n i=1 w i x i = µ, w i 0, Confidence set {µ R(µ) c} n i=1 } w i = 1 To pick c 2 log(r(µ 0 )) d χ 2 (k), ( c = exp 1 ) 2 χ2,1 α (k) k = rank(var(x))
15 Bayesian Empirical Likelihood 15 Dipper, Cinclus cinclus Eats larvae of Mayflies, Stoneflies, Caddis flies, other Photo By Mark Medcalf. From Wikipedia. Lic under CC by
16 Bayesian Empirical Likelihood 16 Dipper diet means Stonefly larvae Other invertebrates Caddis fly larvae Mayfly larvae Stonefly larvae Other invertebrates Caddis fly larvae Top row shows EL; bottom Hotelling s T 2 ellipses Data from Iles (1993) 22 sites in Wales. Hall (1990) quantifies correctness of shape Mayfly larvae
17 Bayesian Empirical Likelihood 17 Computing EL for the mean Easy by convex duality. The dual is self-concordant = truncated Newton assured of convergence. O (2013) Canadian Journal of Statistics R code online
18 Bayesian Empirical Likelihood 18 Estimating equations θ solves E(m(x, θ)) = 0 Examples m(x, θ) x θ Estimand θ E(x) 1 x<θ 0.5 median(x) (x θ) 1 z A E(x z A) x(y x T β) regression θ log f(x, θ) MLE
19 Bayesian Empirical Likelihood 19 Note that some of these estimands can be well defined without making any assumptions about the likelihood. For instance, we could well want a median without wanting to assume a double exponential distribution or any other whose MLE is the median. Similarly, least squares estimates are interpretable as sample based counterparts to population least squares quantities. Other cases, like logistic regression, make it harder to interpret what you get if the motivating parametric model does not hold. That said, you might still prefer reliable to unreliable uncertainty quantification for such an estimand.
20 Bayesian Empirical Likelihood 20 EL properties High power Kitamura (2001), Lazar & Mykland (1998) Comparable to a true parametric model (when there is one) Bartlett correctable DiCiccio, Hall & Romano (1991) Coverage errors O(1/n) O(1/n 2 ) Challenges Convex hull issue, fixable by pseudo-data, Chen & Variyath (2008), Tsao & Wu (2013/14), Emerson & O (2009) Computing max{r(θ 1,..., θ p ) θ j = θ j0 } can be hard O (2000) used expensive NPSOL solver Possibility Sampling might work more easily than optimization
21 Bayesian Empirical Likelihood 21 Overdetermined estimating equations θ R p, m(x, θ) R q, q > p 1 n q equations in p < q unknowns: m(x i, θ) = 0 n i=1 Popular in econometrics E.g., generalized method of moments (GMM) Hansen (1982) E.g., regression through (0, 0) E(y xβ) = E(x(y xβ)) = 0 Misspecification Sometimes no θ R p gives E(m(x, θ)) = 0 R q
22 Bayesian Empirical Likelihood 22 EL for overdetermined Get n more multinomial parameters w 1,..., w n NPMLE max θ,w n i=1 w i subject to n w i m(x i, θ) = 0, i=1 n w i = 1, w i > 0 i=1 Then solve i ŵim(x i, θ) = 0 for ˆθ Newey & Smith (2004): same variance & less bias than GMM
23 Bayesian Empirical Likelihood 23 Bayesian EL p(θ x) p(θ)r(θ x) Lazar (2003) justifies by a Bayesian CLT, like Boos & Monahan (1992) Also Hjort (Today!) Is that Bayesian enough?
24 Bayesian Empirical Likelihood 24 Cressie-Read likelihoods There are numerous alternative nonparametric likelihoods. Chang & Mukerjee (2008) give conditions for them to have a very accurate BCLT for very general priors. EL is uniquely able to do so. Some others will work with a flat prior. Poster by Turbatu (yesterday) Also
25 Bayesian Empirical Likelihood 25 Least favorable families Pr EL (x i ) = w i = 1 n p-dimensional θ = p-dimensional λ(θ). There is a least favorable p-dimensional family for θ. Any other family makes the inference artificially easy λ(θ) T m(x i, θ) EL family is asymptotically the LFF DiCiccio & Romano (1990) The connection p(θ) induces a prior on the LFF, and R(θ x) =. likelihood on LFF = p(θ) R(θ x) =. posterior on LFF = posterior on θ
26 Bayesian Empirical Likelihood 26 To my knowledge nobody has tracked down exactly how the approximations work out for the least favorable family motivation of Bayesian EL.
27 Bayesian Empirical Likelihood 27 Schennach (2005) Exponential tilting (ET) To get an empirical probability model, for θ Θ R p, partition Θ = N l=1 Θ l p(θ) is multinomial on N disjoint cells. Now let N for universal approximation. Ultimately n p(θ x) p(θ) wi (θ), s.t. i=1 w i maximize n wi x i = θ, where n w i log(w i ) i=1 i=1 i w i = 1
28 Bayesian Empirical Likelihood 28 Exponential tilting EL Schennach (2007) Form parametric family via exponential tilting w i (θ) = e λ(θ)t m(x i,θ) n j=1 eλ(θ)t m(x j,θ), where 0 = n e λ(θ)t m(x i,θ) m(x i, θ) i=1 Likelihood L(θ) = n nwi (θ) i=1 ETEL ˆθ ETEL = arg max θ L(θ) Works better on misspecified overdetermined models, i.e., no θ R p makes E(m(x, θ)) = 0 R q Reason: at least there is an estimand.
29 Bayesian Empirical Likelihood 29 If we know that no θ can make E(m(x, θ)) = 0 then I m not sure why we still want an estimand. It seems like we should instead change the choice of m.
30 Bayesian Empirical Likelihood 30 BETEL Chib, Shin, Simoni (2016,2017) Use a slack parameter V k. V k 0 for non-truthfulness of moment condition k. At most m p nonzero V k. Metropolized Independence Sampling from a t distribution. Asymptotic model selection Use log marginal likelihood. If there is a true model then it will be selected over any false one. Among true models: most equations holding wins. poster by Liu (yesterday) Robust BETEL
31 Bayesian Empirical Likelihood 31 EL with ABC Mengersen, Pudlo & Robert (2013) Sample θ i p(θ) Compute w i = R(θ i x) Normalize the weights Basic EL-ABC Advanced EL-ABC Use adaptive multiple importance sampling (AMIS) Algorithm has 5 loops Max depth 3
32 Bayesian Empirical Likelihood 32 Hamiltonian MCMC Chaudhuri, Mondal & Yin (2017) Using p(θ)r(θ x) they report that: Gibbs is not applicable Random walk Metropolis has problems with irregular support of R(θ x) Use gradients, Hamiltonian MCMC. Gradient gets steep just where you need it. Their solution
33 Bayesian Empirical Likelihood 33 Vexler, Tao & Hutson (2014) For x i R Posterior moments x(n) x (1) x(n) x (1) θ k R(θ x)p(θ) dθ R(θ x)p(θ) dθ Nonparametric James-Stein For x i = (x i1,..., x ik ) R K, get R j (θ j ) = R(θ j x 1j,..., x nj ) ˆθ j = θj R j (θ)ϕ( θj θ 0 σ ) dθ j Rj (θ)ϕ((θ j θ 0 )/σ ) dθ j σ 2 = arg max σ 2 K j=1 ( 1 log 2πσ R j (θ j )e θ2 j /(2σ2) dθ j )
34 Bayesian Empirical Likelihood 34 Quantiles Vexler, Yu & Lazar (2017) EL Bayes factors for quantile θ Q α = Q 0 vs Q α Q 0 They overcome a technical challenge handling H A R(θ x)p(θ) dθ because R has jump discontinuities. Also two sample Bayes factors. If scientific or regulalatory interest is in a specific quantile then the user does not have to pick a whole parametric family.
35 Bayesian Empirical Likelihood 35 Quantile regression Pr(Y α + βx x) = τ, e.g., τ = 0.9 Estimating equations 0 = E ( 1 y α+βx τ ) 0 = E ( x ( 1 y α+βx τ )) Lancaster & Jun (2010) BETEL for quantiles similar to Jeffrey s prior Yang & He (2012) BCLT accounting for non smoothness of estimating equations Shrinking priors pool β for multiple τ Both use MCMC
36 Bayesian Empirical Likelihood 36 Summary We can do Bayes without requiring a parametric family for the observations. It is the other nonparametric Bayes.
37 Bayesian Empirical Likelihood 37 Thanks Luke Bornn, invitation Sanjay Chaudhuri, many references Scientific and organizing committees Especially Judith Rousseau NSF DMS , DMS
ICSA Applied Statistics Symposium 1. Balanced adjusted empirical likelihood
ICSA Applied Statistics Symposium 1 Balanced adjusted empirical likelihood Art B. Owen Stanford University Sarah Emerson Oregon State University ICSA Applied Statistics Symposium 2 Empirical likelihood
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationHigh Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data
High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data Song Xi CHEN Guanghua School of Management and Center for Statistical Science, Peking University Department
More informationJackknife Empirical Likelihood for the Variance in the Linear Regression Model
Georgia State University ScholarWorks @ Georgia State University Mathematics Theses Department of Mathematics and Statistics Summer 7-25-2013 Jackknife Empirical Likelihood for the Variance in the Linear
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More information11 Survival Analysis and Empirical Likelihood
11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with
More informationNew Bayesian methods for model comparison
Back to the future New Bayesian methods for model comparison Murray Aitkin murray.aitkin@unimelb.edu.au Department of Mathematics and Statistics The University of Melbourne Australia Bayesian Model Comparison
More informationThe Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.
Christian P. Robert The Bayesian Choice From Decision-Theoretic Foundations to Computational Implementation Second Edition With 23 Illustrations ^Springer" Contents Preface to the Second Edition Preface
More informationStatistical Tools and Techniques for Solar Astronomers
Statistical Tools and Techniques for Solar Astronomers Alexander W Blocker Nathan Stein SolarStat 2012 Outline Outline 1 Introduction & Objectives 2 Statistical issues with astronomical data 3 Example:
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationParameter Estimation
Parameter Estimation Chapters 13-15 Stat 477 - Loss Models Chapters 13-15 (Stat 477) Parameter Estimation Brian Hartman - BYU 1 / 23 Methods for parameter estimation Methods for parameter estimation Methods
More informationCOS513 LECTURE 8 STATISTICAL CONCEPTS
COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions
More informationEmpirical Likelihood
Empirical Likelihood Patrick Breheny September 20 Patrick Breheny STA 621: Nonparametric Statistics 1/15 Introduction Empirical likelihood We will discuss one final approach to constructing confidence
More informationAFT Models and Empirical Likelihood
AFT Models and Empirical Likelihood Mai Zhou Department of Statistics, University of Kentucky Collaborators: Gang Li (UCLA); A. Bathke; M. Kim (Kentucky) Accelerated Failure Time (AFT) models: Y = log(t
More informationA Very Brief Summary of Bayesian Inference, and Examples
A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationEconometrics I, Estimation
Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationEmpirical Likelihood Methods for Sample Survey Data: An Overview
AUSTRIAN JOURNAL OF STATISTICS Volume 35 (2006), Number 2&3, 191 196 Empirical Likelihood Methods for Sample Survey Data: An Overview J. N. K. Rao Carleton University, Ottawa, Canada Abstract: The use
More informationNoninformative Priors for the Ratio of the Scale Parameters in the Inverted Exponential Distributions
Communications for Statistical Applications and Methods 03, Vol. 0, No. 5, 387 394 DOI: http://dx.doi.org/0.535/csam.03.0.5.387 Noninformative Priors for the Ratio of the Scale Parameters in the Inverted
More informationProbability and Estimation. Alan Moses
Probability and Estimation Alan Moses Random variables and probability A random variable is like a variable in algebra (e.g., y=e x ), but where at least part of the variability is taken to be stochastic.
More informationStat 5102 Final Exam May 14, 2015
Stat 5102 Final Exam May 14, 2015 Name Student ID The exam is closed book and closed notes. You may use three 8 1 11 2 sheets of paper with formulas, etc. You may also use the handouts on brand name distributions
More informationCLASS NOTES Models, Algorithms and Data: Introduction to computing 2018
CLASS NOTES Models, Algorithms and Data: Introduction to computing 208 Petros Koumoutsakos, Jens Honore Walther (Last update: June, 208) IMPORTANT DISCLAIMERS. REFERENCES: Much of the material (ideas,
More informationIntroduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models
Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models Matthew S. Johnson New York ASA Chapter Workshop CUNY Graduate Center New York, NY hspace1in December 17, 2009 December
More informationPart III. A Decision-Theoretic Approach and Bayesian testing
Part III A Decision-Theoretic Approach and Bayesian testing 1 Chapter 10 Bayesian Inference as a Decision Problem The decision-theoretic framework starts with the following situation. We would like to
More informationQuantile methods. Class Notes Manuel Arellano December 1, Let F (r) =Pr(Y r). Forτ (0, 1), theτth population quantile of Y is defined to be
Quantile methods Class Notes Manuel Arellano December 1, 2009 1 Unconditional quantiles Let F (r) =Pr(Y r). Forτ (0, 1), theτth population quantile of Y is defined to be Q τ (Y ) q τ F 1 (τ) =inf{r : F
More informationand Comparison with NPMLE
NONPARAMETRIC BAYES ESTIMATOR OF SURVIVAL FUNCTIONS FOR DOUBLY/INTERVAL CENSORED DATA and Comparison with NPMLE Mai Zhou Department of Statistics, University of Kentucky, Lexington, KY 40506 USA http://ms.uky.edu/
More informationModel Selection, Estimation, and Bootstrap Smoothing. Bradley Efron Stanford University
Model Selection, Estimation, and Bootstrap Smoothing Bradley Efron Stanford University Estimation After Model Selection Usually: (a) look at data (b) choose model (linear, quad, cubic...?) (c) fit estimates
More informationStatistical Estimation
Statistical Estimation Use data and a model. The plug-in estimators are based on the simple principle of applying the defining functional to the ECDF. Other methods of estimation: minimize residuals from
More informationParametric Techniques Lecture 3
Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to
More informationMaximum likelihood estimation of a log-concave density based on censored data
Maximum likelihood estimation of a log-concave density based on censored data Dominic Schuhmacher Institute of Mathematical Statistics and Actuarial Science University of Bern Joint work with Lutz Dümbgen
More informationBayesian Analysis of Risk for Data Mining Based on Empirical Likelihood
1 / 29 Bayesian Analysis of Risk for Data Mining Based on Empirical Likelihood Yuan Liao Wenxin Jiang Northwestern University Presented at: Department of Statistics and Biostatistics Rutgers University
More informationCarl N. Morris. University of Texas
EMPIRICAL BAYES: A FREQUENCY-BAYES COMPROMISE Carl N. Morris University of Texas Empirical Bayes research has expanded significantly since the ground-breaking paper (1956) of Herbert Robbins, and its province
More informationSTAT 461/561- Assignments, Year 2015
STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and
More informationProbabilistic Graphical Models
Parameter Estimation December 14, 2015 Overview 1 Motivation 2 3 4 What did we have so far? 1 Representations: how do we model the problem? (directed/undirected). 2 Inference: given a model and partially
More informationInvariant HPD credible sets and MAP estimators
Bayesian Analysis (007), Number 4, pp. 681 69 Invariant HPD credible sets and MAP estimators Pierre Druilhet and Jean-Michel Marin Abstract. MAP estimators and HPD credible sets are often criticized in
More informationPattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite
More informationStatistica Sinica Preprint No: SS R1
Statistica Sinica Preprint No: SS-2017-0059.R1 Title Peter Hall's Contribution to Empirical Likelihood Manuscript ID SS-2017-0059.R1 URL http://www.stat.sinica.edu.tw/statistica/ DOI 10.5705/ss.202017.0059
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationMultistate Modeling and Applications
Multistate Modeling and Applications Yang Yang Department of Statistics University of Michigan, Ann Arbor IBM Research Graduate Student Workshop: Statistics for a Smarter Planet Yang Yang (UM, Ann Arbor)
More informationLinear Models A linear model is defined by the expression
Linear Models A linear model is defined by the expression x = F β + ɛ. where x = (x 1, x 2,..., x n ) is vector of size n usually known as the response vector. β = (β 1, β 2,..., β p ) is the transpose
More informationSMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES
Statistica Sinica 19 (2009), 71-81 SMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES Song Xi Chen 1,2 and Chiu Min Wong 3 1 Iowa State University, 2 Peking University and
More informationConstrained estimation for binary and survival data
Constrained estimation for binary and survival data Jeremy M. G. Taylor Yong Seok Park John D. Kalbfleisch Biostatistics, University of Michigan May, 2010 () Constrained estimation May, 2010 1 / 43 Outline
More informationA note on multiple imputation for general purpose estimation
A note on multiple imputation for general purpose estimation Shu Yang Jae Kwang Kim SSC meeting June 16, 2015 Shu Yang, Jae Kwang Kim Multiple Imputation June 16, 2015 1 / 32 Introduction Basic Setup Assume
More informationSpring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM
University of Illinois Department of Economics Spring 2017 Econ 574 Roger Koenker Lecture 14 GEE-GMM Throughout the course we have emphasized methods of estimation and inference based on the principle
More informationVarious types of likelihood
Various types of likelihood 1. likelihood, marginal likelihood, conditional likelihood, profile likelihood, adjusted profile likelihood 2. semi-parametric likelihood, partial likelihood 3. empirical likelihood,
More informationProblem Set 3: Bootstrap, Quantile Regression and MCMC Methods. MIT , Fall Due: Wednesday, 07 November 2007, 5:00 PM
Problem Set 3: Bootstrap, Quantile Regression and MCMC Methods MIT 14.385, Fall 2007 Due: Wednesday, 07 November 2007, 5:00 PM 1 Applied Problems Instructions: The page indications given below give you
More informationIntroduction to Applied Bayesian Modeling. ICPSR Day 4
Introduction to Applied Bayesian Modeling ICPSR Day 4 Simple Priors Remember Bayes Law: Where P(A) is the prior probability of A Simple prior Recall the test for disease example where we specified the
More informationBayesian Inference. Chapter 2: Conjugate models
Bayesian Inference Chapter 2: Conjugate models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in
More informationA Bayesian perspective on GMM and IV
A Bayesian perspective on GMM and IV Christopher A. Sims Princeton University sims@princeton.edu November 26, 2013 What is a Bayesian perspective? A Bayesian perspective on scientific reporting views all
More informationBayesian and frequentist inference
Bayesian and frequentist inference Nancy Reid March 26, 2007 Don Fraser, Ana-Maria Staicu Overview Methods of inference Asymptotic theory Approximate posteriors matching priors Examples Logistic regression
More informationCan we do statistical inference in a non-asymptotic way? 1
Can we do statistical inference in a non-asymptotic way? 1 Guang Cheng 2 Statistics@Purdue www.science.purdue.edu/bigdata/ ONR Review Meeting@Duke Oct 11, 2017 1 Acknowledge NSF, ONR and Simons Foundation.
More informationPrimer on statistics:
Primer on statistics: MLE, Confidence Intervals, and Hypothesis Testing ryan.reece@gmail.com http://rreece.github.io/ Insight Data Science - AI Fellows Workshop Feb 16, 018 Outline 1. Maximum likelihood
More informationi=1 h n (ˆθ n ) = 0. (2)
Stat 8112 Lecture Notes Unbiased Estimating Equations Charles J. Geyer April 29, 2012 1 Introduction In this handout we generalize the notion of maximum likelihood estimation to solution of unbiased estimating
More informationSome Curiosities Arising in Objective Bayesian Analysis
. Some Curiosities Arising in Objective Bayesian Analysis Jim Berger Duke University Statistical and Applied Mathematical Institute Yale University May 15, 2009 1 Three vignettes related to John s work
More informationModule 22: Bayesian Methods Lecture 9 A: Default prior selection
Module 22: Bayesian Methods Lecture 9 A: Default prior selection Peter Hoff Departments of Statistics and Biostatistics University of Washington Outline Jeffreys prior Unit information priors Empirical
More informationPractice Exam 1. (A) (B) (C) (D) (E) You are given the following data on loss sizes:
Practice Exam 1 1. Losses for an insurance coverage have the following cumulative distribution function: F(0) = 0 F(1,000) = 0.2 F(5,000) = 0.4 F(10,000) = 0.9 F(100,000) = 1 with linear interpolation
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationBayesian estimation of the discrepancy with misspecified parametric models
Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012
More informationLecture 23 Maximum Likelihood Estimation and Bayesian Inference
Lecture 23 Maximum Likelihood Estimation and Bayesian Inference Thais Paiva STA 111 - Summer 2013 Term II August 7, 2013 1 / 31 Thais Paiva STA 111 - Summer 2013 Term II Lecture 23, 08/07/2013 Lecture
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationBayesian methods in economics and finance
1/26 Bayesian methods in economics and finance Linear regression: Bayesian model selection and sparsity priors Linear Regression 2/26 Linear regression Model for relationship between (several) independent
More informationAssociation studies and regression
Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative
More informationBayesian Inference and the Parametric Bootstrap. Bradley Efron Stanford University
Bayesian Inference and the Parametric Bootstrap Bradley Efron Stanford University Importance Sampling for Bayes Posterior Distribution Newton and Raftery (1994 JRSS-B) Nonparametric Bootstrap: good choice
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School
More informationDefault priors and model parametrization
1 / 16 Default priors and model parametrization Nancy Reid O-Bayes09, June 6, 2009 Don Fraser, Elisabeta Marras, Grace Yun-Yi 2 / 16 Well-calibrated priors model f (y; θ), F(y; θ); log-likelihood l(θ)
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationContents. Part I: Fundamentals of Bayesian Inference 1
Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian
More informationMarkov Networks.
Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function
More informationParameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn!
Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Questions?! C. Porciani! Estimation & forecasting! 2! Cosmological parameters! A branch of modern cosmological research focuses
More informationCSC321 Lecture 18: Learning Probabilistic Models
CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling
More informationLecture 3: More on regularization. Bayesian vs maximum likelihood learning
Lecture 3: More on regularization. Bayesian vs maximum likelihood learning L2 and L1 regularization for linear estimators A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting
More informationTail bound inequalities and empirical likelihood for the mean
Tail bound inequalities and empirical likelihood for the mean Sandra Vucane 1 1 University of Latvia, Riga 29 th of September, 2011 Sandra Vucane (LU) Tail bound inequalities and EL for the mean 29.09.2011
More informationParametric Techniques
Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationA COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky
A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky Empirical likelihood with right censored data were studied by Thomas and Grunkmier (1975), Li (1995),
More informationIntroduction to Machine Learning. Lecture 2
Introduction to Machine Learning Lecturer: Eran Halperin Lecture 2 Fall Semester Scribe: Yishay Mansour Some of the material was not presented in class (and is marked with a side line) and is given for
More informationBootstrap and Parametric Inference: Successes and Challenges
Bootstrap and Parametric Inference: Successes and Challenges G. Alastair Young Department of Mathematics Imperial College London Newton Institute, January 2008 Overview Overview Review key aspects of frequentist
More informationBayesian inference for multivariate extreme value distributions
Bayesian inference for multivariate extreme value distributions Sebastian Engelke Clément Dombry, Marco Oesting Toronto, Fields Institute, May 4th, 2016 Main motivation For a parametric model Z F θ of
More informationFundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner
Fundamentals CS 281A: Statistical Learning Theory Yangqing Jia Based on tutorial slides by Lester Mackey and Ariel Kleiner August, 2011 Outline 1 Probability 2 Statistics 3 Linear Algebra 4 Optimization
More informationBayesian Inference: Posterior Intervals
Bayesian Inference: Posterior Intervals Simple values like the posterior mean E[θ X] and posterior variance var[θ X] can be useful in learning about θ. Quantiles of π(θ X) (especially the posterior median)
More informationSTAT215: Solutions for Homework 2
STAT25: Solutions for Homework 2 Due: Wednesday, Feb 4. (0 pt) Suppose we take one observation, X, from the discrete distribution, x 2 0 2 Pr(X x θ) ( θ)/4 θ/2 /2 (3 θ)/2 θ/4, 0 θ Find an unbiased estimator
More informationModelling geoadditive survival data
Modelling geoadditive survival data Thomas Kneib & Ludwig Fahrmeir Department of Statistics, Ludwig-Maximilians-University Munich 1. Leukemia survival data 2. Structured hazard regression 3. Mixed model
More informationBayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017
Bayesian inference Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 10, 2017 1 / 22 Outline for today A genetic example Bayes theorem Examples Priors Posterior summaries
More informationBayesian Inference: Concept and Practice
Inference: Concept and Practice fundamentals Johan A. Elkink School of Politics & International Relations University College Dublin 5 June 2017 1 2 3 Bayes theorem In order to estimate the parameters of
More informationReview and continuation from last week Properties of MLEs
Review and continuation from last week Properties of MLEs As we have mentioned, MLEs have a nice intuitive property, and as we have seen, they have a certain equivariance property. We will see later that
More informationIntroduction to Bayesian Methods
Introduction to Bayesian Methods Jessi Cisewski Department of Statistics Yale University Sagan Summer Workshop 2016 Our goal: introduction to Bayesian methods Likelihoods Priors: conjugate priors, non-informative
More informationAsymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands
Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Elizabeth C. Mannshardt-Shamseldin Advisor: Richard L. Smith Duke University Department
More informationGeneral Bayesian Inference I
General Bayesian Inference I Outline: Basic concepts, One-parameter models, Noninformative priors. Reading: Chapters 10 and 11 in Kay-I. (Occasional) Simplified Notation. When there is no potential for
More information7. Estimation and hypothesis testing. Objective. Recommended reading
7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing
More informationPreliminaries The bootstrap Bias reduction Hypothesis tests Regression Confidence intervals Time series Final remark. Bootstrap inference
1 / 172 Bootstrap inference Francisco Cribari-Neto Departamento de Estatística Universidade Federal de Pernambuco Recife / PE, Brazil email: cribari@gmail.com October 2014 2 / 172 Unpaid advertisement
More informationRobustness to Parametric Assumptions in Missing Data Models
Robustness to Parametric Assumptions in Missing Data Models Bryan Graham NYU Keisuke Hirano University of Arizona April 2011 Motivation Motivation We consider the classic missing data problem. In practice
More informationModel Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao
Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley
More informationCovariate Balancing Propensity Score for General Treatment Regimes
Covariate Balancing Propensity Score for General Treatment Regimes Kosuke Imai Princeton University October 14, 2014 Talk at the Department of Psychiatry, Columbia University Joint work with Christian
More informationApproximate Bayesian computation for spatial extremes via open-faced sandwich adjustment
Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment Ben Shaby SAMSI August 3, 2010 Ben Shaby (SAMSI) OFS adjustment August 3, 2010 1 / 29 Outline 1 Introduction 2 Spatial
More informationIEOR 165 Lecture 7 1 Bias-Variance Tradeoff
IEOR 165 Lecture 7 Bias-Variance Tradeoff 1 Bias-Variance Tradeoff Consider the case of parametric regression with β R, and suppose we would like to analyze the error of the estimate ˆβ in comparison to
More information