Bayesian Statistics. Debdeep Pati Florida State University. April 3, 2017
|
|
- Calvin Johnson
- 5 years ago
- Views:
Transcription
1 Bayesian Statistics Debdeep Pati Florida State University April 3, 2017
2 Finite mixture model The finite mixture of normals can be equivalently expressed as y i N(µ Si ; τ 1 S i ), S i k π h δ h h=1 δ h = probability measure concentrated at the integer h, S i {1, 2,..., k} indexes the mixture component for subject i, i = 1,..., n. A prior on π = (π 1, π 2,..., π k ) and (µ h, τ h ), h = 1,..., k is given by π Dir(α 1, α 2,..., α k ) (µ h, τ h ) N(µ h ; µ 0, κτ 1 h )Ga(τ h; a τ ; b τ ), h = 1,..., k
3 Posterior Computation in finite mixture models Update S i from its multinomial conditional posterior with Pr(S i = h ) = π h N(y i ; µ h, τ 1 h ) k l=1 π hn(y i ; µ h, τ 1 h = 1,..., k ), h Let n h = #{S i = h, i = 1,..., n} and ȳ h = 1 n h i:s i =h y i and update (µ h, τ 1 h ) from its conditional posterior (µ h, τ 1 h ) = N(µ h, ˆµ h, ˆκ h τ 1 h )Ga(τ h, â τh, ˆb τh ) where ˆκ h = (κ 1 + n h ) 1, ˆµ h = ˆκ(κ 1 µ 0 + n h ȳ h ), â τh = a τ + n h ˆb τh = b τ + 1 { (y i ȳ h ) 2 + n } h (ȳ h µ 0 ) κn h i:s i =h Update π as (π ) = Dir(a 1 + n 1,..., a k + n k ). 2,
4 Some comments Gibbs sampler is trivial to implement Discarding a burn-in,monitor f (y) = k h=1 π hn(y; µ h, τ 1 for a large number of iterations & a dense grid of y values h ) Bayes estimate of f (y) under squared error loss averages the samples Can also obtain 95% pointwise intervals for unknown density
5 Choosing the Dirichlet hyperparameters The choice of hyperparameters in the mixture model can have an important impact Focus initially on the choice of (a 1,..., a k ) in the Dirichlet prior A common choice is a 1 = = a k = 1,which seems non-informative. However, this is actually a poor choice in many cases, as it favors assigning roughly equal weights to the different components Ideally, we could choose k as an upper bound and choose hyperparameters, which favor a small number of components with relatively large weights
6 Finite Approximation to Dirichlet Process Ishwaran and Zarepour (2002), Dirichlet prior sieves in finite normal mixtures, Statistica Sinica, 12, ) propose a finite approximation to the DP In particular, they propose letting π Dir(α/k,..., α/k) iid Assuming also that θ h = (µ h, τ h ) P 0, they show that lim k h=1 k π h δ θh DP(αP 0 ) In addition, the posterior for the density is L 1 consistent if log k/n 0.
7 Implications We can implement a finite mixture model analysis with a carefully chosen prior & sufficiently large k to obtain an accurate approximation to the DP mixture (DPM) model: f (y) = K(y; θ)dp(θ), P DP(αP 0 ) Here, K(y; θ) is a kernel parameterized by θ - e.g., K(y; θ) = N(y; µ, τ 1 ) with θ = (µ, τ 1 ) for normal mixtures P is now an unknown mixing measure Hence, we no longer use the DP as a prior directly for the distribution of the data but instead use it for the mixture distribution
8 Dirichlet Process Mixtures - comments The discreteness of the Dirichlet process is not a problem when it is used for a mixture distribution instead of directly for the data distribution In fact, in this setting the discreteness is appealing in leading to a simple representation of the mixture distribution that leads to clustering of the observations as a side effect Focusing on the finite approximation, P DP k (α, P 0 ), let f (y) = N(y; µ, τ)dp(µ, τ) = k h=1 π h N(y; µ h, τ 1 h ) This induces a prior on f.
9 Dirichlet Process Mixtures For density estimation,consider the DP mixture (DPM)model y i µ i, τ i N(µ i, τ 1 i ), θ i = (µ i, τ i ) P, P DP(αP 0 )( ) Not immediate clear how to conduct posterior computation One strategy relies on marginalizing out P to obtain ( ) α i 1 1 (θ i θ 1,..., θ i 1 ) P 0 + α i + 1 α + 1 δ θ j j=1 DP prediction rule or Polya urn scheme (Blackwell & MacQueen, 73)
10 Avoiding Marginalization By marginalizing out the RPM P, we give up the ability to conduct inferences on P By having approaches that avoid marginalization, we open the door to generalizations of DPMs Stick-breaking representation (Sethuraman, 94), θ i P = i.i.d i.i.d V h (1 V l )δ Θh, V h beta(1, α), θ h P 0 h=1 l<h
11 Samples from Dirichlet process with precision α
12 Implications of Stick-Breaking For small α, most of the probability is allocated to the first few components, favoring few latent classes Expected number of occupied components α log n Weights π h decrease stochastically towards near zero rapidly in the index h Suggests truncation approximation (Muliere & Tardella, 98), P = N V h (1 V l )δ Θh h=1 l<h with V N = 1 so that weights sum to one
13 Blocked Gibbs Sampler (Ishwaran & James, 01) 1. Update S i {1,..., N} by multinomial sampling with P(S i = h ) = π hn(y i ; Θ h ) N l=1 π ln(y i ; Θ l ), h = 1,..., N 2. Update stick-breaking weight V h, h = 1,..., N 1, from ( Beta 1 + n h, α + N l=h+1 n l ). 3. Update Θ h, h = 1,..., N, exactly as in the finite mixture model.
14 Comments on Blocked Gibbs N acts as an upper bound on the number of mixture components in the sample By choosing a large value, the approximation error should be small Possible to monitor this error during the MCMC Approximate inferences on functionals of P are possible Slice (Walker, 07) & retrospective sampling (Papaspiliopoulos & Roberts, 08) approaches avoid truncation - exact block Gibbs (Papaspiliopoulos, 08) combine these approaches
15 Choosing the DP precision parameter The DP precision parameter α plays a key role in controlling the prior on the number of clusters A number of strategies have been proposed in the literature - 1. Fix α at a small number to favor allocation to few clusters relative to the sample size - a commonly used default value is α = Assign a hyperprior (typically gamma) to α - refer to technical report by West (92) & recent article by Dorazio (09, JSPI, 139, ) 3. Estimate α via empirical Bayes (Liu 96; McAulliffe et al. 06)
Nonparametric Bayes Uncertainty Quantification
Nonparametric Bayes Uncertainty Quantification David Dunson Department of Statistical Science, Duke University Funded from NIH R01-ES017240, R01-ES017436 & ONR Review of Bayes Intro to Nonparametric Bayes
More informationSTAT Advanced Bayesian Inference
1 / 32 STAT 625 - Advanced Bayesian Inference Meng Li Department of Statistics Jan 23, 218 The Dirichlet distribution 2 / 32 θ Dirichlet(a 1,...,a k ) with density p(θ 1,θ 2,...,θ k ) = k j=1 Γ(a j) Γ(
More informationLecture 16: Mixtures of Generalized Linear Models
Lecture 16: Mixtures of Generalized Linear Models October 26, 2006 Setting Outline Often, a single GLM may be insufficiently flexible to characterize the data Setting Often, a single GLM may be insufficiently
More informationBayesian non-parametric model to longitudinally predict churn
Bayesian non-parametric model to longitudinally predict churn Bruno Scarpa Università di Padova Conference of European Statistics Stakeholders Methodologists, Producers and Users of European Statistics
More informationBayesian Nonparametrics: Dirichlet Process
Bayesian Nonparametrics: Dirichlet Process Yee Whye Teh Gatsby Computational Neuroscience Unit, UCL http://www.gatsby.ucl.ac.uk/~ywteh/teaching/npbayes2012 Dirichlet Process Cornerstone of modern Bayesian
More informationBayesian nonparametrics
Bayesian nonparametrics 1 Some preliminaries 1.1 de Finetti s theorem We will start our discussion with this foundational theorem. We will assume throughout all variables are defined on the probability
More informationA Nonparametric Model for Stationary Time Series
A Nonparametric Model for Stationary Time Series Isadora Antoniano-Villalobos Bocconi University, Milan, Italy. isadora.antoniano@unibocconi.it Stephen G. Walker University of Texas at Austin, USA. s.g.walker@math.utexas.edu
More informationSlice Sampling Mixture Models
Slice Sampling Mixture Models Maria Kalli, Jim E. Griffin & Stephen G. Walker Centre for Health Services Studies, University of Kent Institute of Mathematics, Statistics & Actuarial Science, University
More informationOutline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution
Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model
More informationChapter 3. Dirichlet Process
Chapter 3 Dirichlet Process 3.1. The Dirichlet Process Prior 3.1.1. Definition The Dirichlet process (DP) is arguably the most popular BNP model for random probability measures (RPM), and plays a central
More informationLecture 16-17: Bayesian Nonparametrics I. STAT 6474 Instructor: Hongxiao Zhu
Lecture 16-17: Bayesian Nonparametrics I STAT 6474 Instructor: Hongxiao Zhu Plan for today Why Bayesian Nonparametrics? Dirichlet Distribution and Dirichlet Processes. 2 Parameter and Patterns Reference:
More informationA Simple Proof of the Stick-Breaking Construction of the Dirichlet Process
A Simple Proof of the Stick-Breaking Construction of the Dirichlet Process John Paisley Department of Computer Science Princeton University, Princeton, NJ jpaisley@princeton.edu Abstract We give a simple
More informationNonparametric Bayes regression and classification through mixtures of product kernels
Nonparametric Bayes regression and classification through mixtures of product kernels David B. Dunson & Abhishek Bhattacharya Department of Statistical Science Box 90251, Duke University Durham, NC 27708-0251,
More informationGibbs Sampling for (Coupled) Infinite Mixture Models in the Stick Breaking Representation
Gibbs Sampling for (Coupled) Infinite Mixture Models in the Stick Breaking Representation Ian Porteous, Alex Ihler, Padhraic Smyth, Max Welling Department of Computer Science UC Irvine, Irvine CA 92697-3425
More informationBayesian Nonparametric Autoregressive Models via Latent Variable Representation
Bayesian Nonparametric Autoregressive Models via Latent Variable Representation Maria De Iorio Yale-NUS College Dept of Statistical Science, University College London Collaborators: Lifeng Ye (UCL, London,
More informationBayesian Nonparametrics
Bayesian Nonparametrics Lorenzo Rosasco 9.520 Class 18 April 11, 2011 About this class Goal To give an overview of some of the basic concepts in Bayesian Nonparametrics. In particular, to discuss Dirichelet
More informationAn adaptive truncation method for inference in Bayesian nonparametric models
An adaptive truncation method for inference in Bayesian nonparametric models arxiv:1308.045v [stat.co] 1 May 014 J.E. Griffin School of Mathematics, Statistics and Actuarial Science, University of Kent,
More informationLecture 3a: Dirichlet processes
Lecture 3a: Dirichlet processes Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London c.archambeau@cs.ucl.ac.uk Advanced Topics
More informationOn Simulations form the Two-Parameter. Poisson-Dirichlet Process and the Normalized. Inverse-Gaussian Process
On Simulations form the Two-Parameter arxiv:1209.5359v1 [stat.co] 24 Sep 2012 Poisson-Dirichlet Process and the Normalized Inverse-Gaussian Process Luai Al Labadi and Mahmoud Zarepour May 8, 2018 ABSTRACT
More informationKernel Stick-Breaking Processes
Kernel Stick-Breaking Processes David B. Dunson 1 and Ju-Hyun Park 1,2 1 Biostatistics Branch, National Institute of Environmental Health Sciences U.S. National Institute of Health P.O. Box 12233, RTP,
More informationA comparative review of variable selection techniques for covariate dependent Dirichlet process mixture models
A comparative review of variable selection techniques for covariate dependent Dirichlet process mixture models William Barcella 1, Maria De Iorio 1 and Gianluca Baio 1 1 Department of Statistical Science,
More informationNon-Parametric Bayes
Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian
More informationNon-parametric Clustering with Dirichlet Processes
Non-parametric Clustering with Dirichlet Processes Timothy Burns SUNY at Buffalo Mar. 31 2009 T. Burns (SUNY at Buffalo) Non-parametric Clustering with Dirichlet Processes Mar. 31 2009 1 / 24 Introduction
More informationBayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 : Supplemental Material
Bayesian Mixture Modeling of Significant P Values: A Meta-Analytic Method to Estimate the Degree of Contamination from H 0 : Supplemental Material Quentin Frederik Gronau 1, Monique Duizer 1, Marjan Bakker
More informationVariational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures
17th Europ. Conf. on Machine Learning, Berlin, Germany, 2006. Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures Shipeng Yu 1,2, Kai Yu 2, Volker Tresp 2, and Hans-Peter
More informationMultivariate kernel partition processes
Multivariate kernel partition processes David B. Dunson Department of Statistical Science Box 95, 8 Old Chemistry Building Duke University Durham, NC 778-5 dunson@stat.duke.edu SUMMARY This article considers
More informationBayesian Nonparametric Regression through Mixture Models
Bayesian Nonparametric Regression through Mixture Models Sara Wade Bocconi University Advisor: Sonia Petrone October 7, 2013 Outline 1 Introduction 2 Enriched Dirichlet Process 3 EDP Mixtures for Regression
More informationDirichlet Process. Yee Whye Teh, University College London
Dirichlet Process Yee Whye Teh, University College London Related keywords: Bayesian nonparametrics, stochastic processes, clustering, infinite mixture model, Blackwell-MacQueen urn scheme, Chinese restaurant
More informationNONPARAMETRIC BAYESIAN DENSITY ESTIMATION ON A CIRCLE
NONPARAMETRIC BAYESIAN DENSITY ESTIMATION ON A CIRCLE MARK FISHER Incomplete. Abstract. This note describes nonparametric Bayesian density estimation for circular data using a Dirichlet Process Mixture
More informationGibbs Sampling in Linear Models #2
Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling
More informationNonparametric Bayes Inference on Manifolds with Applications
Nonparametric Bayes Inference on Manifolds with Applications Abhishek Bhattacharya Indian Statistical Institute Based on the book Nonparametric Statistics On Manifolds With Applications To Shape Spaces
More informationPackage msbp. December 13, 2018
Type Package Package msbp December 13, 2018 Title Multiscale Bernstein Polynomials for Densities Version 1.4 Date 2018-12-13 Author Antonio Canale Maintainer Antonio Canale Performs
More informationNonparametric Mixed Membership Models
5 Nonparametric Mixed Membership Models Daniel Heinz Department of Mathematics and Statistics, Loyola University of Maryland, Baltimore, MD 21210, USA CONTENTS 5.1 Introduction................................................................................
More informationNonparametric Bayes Modeling
Nonparametric Bayes Modeling Lecture 6: Advanced Applications of DPMs David Dunson Department of Statistical Science, Duke University Tuesday February 2, 2010 Motivation Functional data analysis Variable
More informationDirichlet Enhanced Latent Semantic Analysis
Dirichlet Enhanced Latent Semantic Analysis Kai Yu Siemens Corporate Technology D-81730 Munich, Germany Kai.Yu@siemens.com Shipeng Yu Institute for Computer Science University of Munich D-80538 Munich,
More informationParticle Learning for General Mixtures
Particle Learning for General Mixtures Hedibert Freitas Lopes 1 Booth School of Business University of Chicago Dipartimento di Scienze delle Decisioni Università Bocconi, Milano 1 Joint work with Nicholas
More informationFoundations of Nonparametric Bayesian Methods
1 / 27 Foundations of Nonparametric Bayesian Methods Part II: Models on the Simplex Peter Orbanz http://mlg.eng.cam.ac.uk/porbanz/npb-tutorial.html 2 / 27 Tutorial Overview Part I: Basics Part II: Models
More informationHybrid Dirichlet processes for functional data
Hybrid Dirichlet processes for functional data Sonia Petrone Università Bocconi, Milano Joint work with Michele Guindani - U.T. MD Anderson Cancer Center, Houston and Alan Gelfand - Duke University, USA
More informationBayesian Nonparametric Inference Methods for Mean Residual Life Functions
Bayesian Nonparametric Inference Methods for Mean Residual Life Functions Valerie Poynor Department of Applied Mathematics and Statistics, University of California, Santa Cruz April 28, 212 1/3 Outline
More informationLecturer: David Blei Lecture #3 Scribes: Jordan Boyd-Graber and Francisco Pereira October 1, 2007
COS 597C: Bayesian Nonparametrics Lecturer: David Blei Lecture # Scribes: Jordan Boyd-Graber and Francisco Pereira October, 7 Gibbs Sampling with a DP First, let s recapitulate the model that we re using.
More informationDirichlet Processes and other non-parametric Bayesian models
Dirichlet Processes and other non-parametric Bayesian models Zoubin Ghahramani http://learning.eng.cam.ac.uk/zoubin/ zoubin@cs.cmu.edu Statistical Machine Learning CMU 10-702 / 36-702 Spring 2008 Model
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationColouring and breaking sticks, pairwise coincidence losses, and clustering expression profiles
Colouring and breaking sticks, pairwise coincidence losses, and clustering expression profiles Peter Green and John Lau University of Bristol P.J.Green@bristol.ac.uk Isaac Newton Institute, 11 December
More informationMatrix-Variate Dirichlet Process Mixture Models
Zhihua Zhang Guang Dai Michael I. Jordan College of Comp. Sci. and Tech. Zhejiang University Zhejiang 317, China zhzhang@cs.zju.edu.cn College of Comp. Sci. and Tech. Zhejiang University Zhejiang 317,
More informationCMPS 242: Project Report
CMPS 242: Project Report RadhaKrishna Vuppala Univ. of California, Santa Cruz vrk@soe.ucsc.edu Abstract The classification procedures impose certain models on the data and when the assumption match the
More information2 Ishwaran and James: Gibbs sampling stick-breaking priors Our second method, the blocked Gibbs sampler, works in greater generality in that it can be
Hemant ISHWARAN 1 and Lancelot F. JAMES 2 Gibbs Sampling Methods for Stick-Breaking Priors A rich and exible class of random probability measures, which we call stick-breaking priors, can be constructed
More informationBAYESIAN DENSITY REGRESSION AND PREDICTOR-DEPENDENT CLUSTERING
BAYESIAN DENSITY REGRESSION AND PREDICTOR-DEPENDENT CLUSTERING by Ju-Hyun Park A dissertation submitted to the faculty of the University of North Carolina at Chapel Hill in partial fulfillment of the requirements
More informationIntroduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf
1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 2013-14 We know that X ~ B(n,p), but we do not know p. We get a random sample
More informationInfinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix
Infinite-State Markov-switching for Dynamic Volatility Models : Web Appendix Arnaud Dufays 1 Centre de Recherche en Economie et Statistique March 19, 2014 1 Comparison of the two MS-GARCH approximations
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationPrésentation des travaux en statistiques bayésiennes
Présentation des travaux en statistiques bayésiennes Membres impliqués à Lille 3 Ophélie Guin Aurore Lavigne Thèmes de recherche Modélisation hiérarchique temporelle, spatiale, et spatio-temporelle Classification
More informationA Fully Nonparametric Modeling Approach to. BNP Binary Regression
A Fully Nonparametric Modeling Approach to Binary Regression Maria Department of Applied Mathematics and Statistics University of California, Santa Cruz SBIES, April 27-28, 2012 Outline 1 2 3 Simulation
More informationDavid B. Dahl. Department of Statistics, and Department of Biostatistics & Medical Informatics University of Wisconsin Madison
AN IMPROVED MERGE-SPLIT SAMPLER FOR CONJUGATE DIRICHLET PROCESS MIXTURE MODELS David B. Dahl dbdahl@stat.wisc.edu Department of Statistics, and Department of Biostatistics & Medical Informatics University
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationarxiv: v1 [stat.me] 3 Mar 2013
Anjishnu Banerjee Jared Murray David B. Dunson Statistical Science, Duke University arxiv:133.449v1 [stat.me] 3 Mar 213 Abstract There is increasing interest in broad application areas in defining flexible
More informationBayesian isotonic density regression
Bayesian isotonic density regression Lianming Wang and David B. Dunson Biostatistics Branch, MD A3-3 National Institute of Environmental Health Sciences U.S. National Institutes of Health P.O. Box 33,
More informationCS281B / Stat 241B : Statistical Learning Theory Lecture: #22 on 19 Apr Dirichlet Process I
X i Ν CS281B / Stat 241B : Statistical Learning Theory Lecture: #22 on 19 Apr 2004 Dirichlet Process I Lecturer: Prof. Michael Jordan Scribe: Daniel Schonberg dschonbe@eecs.berkeley.edu 22.1 Dirichlet
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationNon-parametric Bayesian Methods
Non-parametric Bayesian Methods Uncertainty in Artificial Intelligence Tutorial July 25 Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK Center for Automated Learning
More informationDefining Predictive Probability Functions for Species Sampling Models
Defining Predictive Probability Functions for Species Sampling Models Jaeyong Lee Department of Statistics, Seoul National University leejyc@gmail.com Fernando A. Quintana Departamento de Estadísica, Pontificia
More informationIntroduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf
1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Lior Wolf 2014-15 We know that X ~ B(n,p), but we do not know p. We get a random sample from X, a
More informationA Stick-Breaking Construction of the Beta Process
John Paisley 1 jwp4@ee.duke.edu Aimee Zaas 2 aimee.zaas@duke.edu Christopher W. Woods 2 woods004@mc.duke.edu Geoffrey S. Ginsburg 2 ginsb005@duke.edu Lawrence Carin 1 lcarin@ee.duke.edu 1 Department of
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationHidden Markov Models With Stick-Breaking Priors John Paisley, Student Member, IEEE, and Lawrence Carin, Fellow, IEEE
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 57, NO. 10, OCTOBER 2009 3905 Hidden Markov Models With Stick-Breaking Priors John Paisley, Student Member, IEEE, and Lawrence Carin, Fellow, IEEE Abstract
More informationA Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness
A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness A. Linero and M. Daniels UF, UT-Austin SRC 2014, Galveston, TX 1 Background 2 Working model
More informationBayesian Sparse Correlated Factor Analysis
Bayesian Sparse Correlated Factor Analysis 1 Abstract In this paper, we propose a new sparse correlated factor model under a Bayesian framework that intended to model transcription factor regulation in
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationStochastic Processes, Kernel Regression, Infinite Mixture Models
Stochastic Processes, Kernel Regression, Infinite Mixture Models Gabriel Huang (TA for Simon Lacoste-Julien) IFT 6269 : Probabilistic Graphical Models - Fall 2018 Stochastic Process = Random Function 2
More informationOn the posterior structure of NRMI
On the posterior structure of NRMI Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with L.F. James and A. Lijoi Isaac Newton Institute, BNR Programme, 8th August 2007 Outline
More informationMCMC: Markov Chain Monte Carlo
I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov
More informationBayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework
HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for
More informationBayesian semiparametric modeling and inference with mixtures of symmetric distributions
Bayesian semiparametric modeling and inference with mixtures of symmetric distributions Athanasios Kottas 1 and Gilbert W. Fellingham 2 1 Department of Applied Mathematics and Statistics, University of
More informationDefault priors for density estimation with mixture models
Bayesian Analysis ) 5, Number, pp. 45 64 Default priors for density estimation with mixture models J.E. Griffin Abstract. The infinite mixture of normals model has become a popular method for density estimation
More informationResearch Article Spiked Dirichlet Process Priors for Gaussian Process Models
Hindawi Publishing Corporation Journal of Probability and Statistics Volume 200, Article ID 20489, 4 pages doi:0.55/200/20489 Research Article Spiked Dirichlet Process Priors for Gaussian Process Models
More informationNonparametric Bayesian models through probit stick-breaking processes
Bayesian Analysis (2011) 6, Number 1, pp. 145 178 Nonparametric Bayesian models through probit stick-breaking processes Abel Rodríguez and David B. Dunson Abstract. We describe a novel class of Bayesian
More informationA Nonparametric Bayesian Model for Multivariate Ordinal Data
A Nonparametric Bayesian Model for Multivariate Ordinal Data Athanasios Kottas, University of California at Santa Cruz Peter Müller, The University of Texas M. D. Anderson Cancer Center Fernando A. Quintana,
More informationA Process over all Stationary Covariance Kernels
A Process over all Stationary Covariance Kernels Andrew Gordon Wilson June 9, 0 Abstract I define a process over all stationary covariance kernels. I show how one might be able to perform inference that
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationA Nonparametric Approach Using Dirichlet Process for Hierarchical Generalized Linear Mixed Models
Journal of Data Science 8(2010), 43-59 A Nonparametric Approach Using Dirichlet Process for Hierarchical Generalized Linear Mixed Models Jing Wang Louisiana State University Abstract: In this paper, we
More informationModelling Regime Switching and Structural Breaks with an Infinite Dimension Markov Switching Model
Modelling Regime Switching and Structural Breaks with an Infinite Dimension Markov Switching Model Yong Song University of Toronto tommy.song@utoronto.ca October, 00 Abstract This paper proposes an infinite
More informationThe Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.
Christian P. Robert The Bayesian Choice From Decision-Theoretic Foundations to Computational Implementation Second Edition With 23 Illustrations ^Springer" Contents Preface to the Second Edition Preface
More informationProbabilistic Graphical Models
School of Computer Science Probabilistic Graphical Models Infinite Feature Models: The Indian Buffet Process Eric Xing Lecture 21, April 2, 214 Acknowledgement: slides first drafted by Sinead Williamson
More informationBayesian Nonparametrics: Models Based on the Dirichlet Process
Bayesian Nonparametrics: Models Based on the Dirichlet Process Alessandro Panella Department of Computer Science University of Illinois at Chicago Machine Learning Seminar Series February 18, 2013 Alessandro
More informationBayesian linear regression
Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding
More informationNonparametric Bayes tensor factorizations for big data
Nonparametric Bayes tensor factorizations for big data David Dunson Department of Statistical Science, Duke University Funded from NIH R01-ES017240, R01-ES017436 & DARPA N66001-09-C-2082 Motivation Conditional
More informationBayesian Inference for Dirichlet-Multinomials
Bayesian Inference for Dirichlet-Multinomials Mark Johnson Macquarie University Sydney, Australia MLSS Summer School 1 / 50 Random variables and distributed according to notation A probability distribution
More informationDirichlet Processes: Tutorial and Practical Course
Dirichlet Processes: Tutorial and Practical Course (updated) Yee Whye Teh Gatsby Computational Neuroscience Unit University College London August 2007 / MLSS Yee Whye Teh (Gatsby) DP August 2007 / MLSS
More informationModelling stochastic order in the analysis of receiver operating characteristic data: Bayesian non-parametric approaches
Appl. Statist. (2008) 57, Part 2, pp. 207 225 Modelling stochastic order in the analysis of receiver operating characteristic data: Bayesian non-parametric approaches Timothy E. Hanson, University of Minnesota,
More informationMotivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University
Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined
More informationFlexible Regression Modeling using Bayesian Nonparametric Mixtures
Flexible Regression Modeling using Bayesian Nonparametric Mixtures Athanasios Kottas Department of Applied Mathematics and Statistics University of California, Santa Cruz Department of Statistics Brigham
More informationSteven L. Scott. Presented by Ahmet Engin Ural
Steven L. Scott Presented by Ahmet Engin Ural Overview of HMM Evaluating likelihoods The Likelihood Recursion The Forward-Backward Recursion Sampling HMM DG and FB samplers Autocovariance of samplers Some
More informationPart 3: Applications of Dirichlet processes. 3.2 First examples: applying DPs to density estimation and cluster
CS547Q Statistical Modeling with Stochastic Processes Winter 2011 Part 3: Applications of Dirichlet processes Lecturer: Alexandre Bouchard-Côté Scribe(s): Seagle Liu, Chao Xiong Disclaimer: These notes
More informationA marginal sampler for σ-stable Poisson-Kingman mixture models
A marginal sampler for σ-stable Poisson-Kingman mixture models joint work with Yee Whye Teh and Stefano Favaro María Lomelí Gatsby Unit, University College London Talk at the BNP 10 Raleigh, North Carolina
More informationFitting Narrow Emission Lines in X-ray Spectra
Outline Fitting Narrow Emission Lines in X-ray Spectra Taeyoung Park Department of Statistics, University of Pittsburgh October 11, 2007 Outline of Presentation Outline This talk has three components:
More informationFast Bayesian Inference in Dirichlet Process Mixture Models
Fast Bayesian Inference in Dirichlet Process Mixture Models Lianming Wang 1 and David B. Dunson Biostatistics Branch, MD A3-03 National Institute of Environmental Health Sciences P.O. Box 12233, RTP, NC
More informationMCMC Sampling for Bayesian Inference using L1-type Priors
MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling
More informationDependent mixture models: clustering and borrowing information
ISSN 2279-9362 Dependent mixture models: clustering and borrowing information Antonio Lijoi Bernardo Nipoti Igor Pruenster No. 32 June 213 www.carloalberto.org/research/working-papers 213 by Antonio Lijoi,
More informationNonparametric Bayesian Survival Analysis using Mixtures of Weibull Distributions
Nonparametric Bayesian Survival Analysis using Mixtures of Weibull Distributions ATHANASIOS KOTTAS University of California, Santa Cruz ABSTRACT. Bayesian nonparametric methods have been applied to survival
More information