Bayesian Nonparametrics: some contributions to construction and properties of prior distributions

Size: px
Start display at page:

Download "Bayesian Nonparametrics: some contributions to construction and properties of prior distributions"

Transcription

1 Bayesian Nonparametrics: some contributions to construction and properties of prior distributions Annalisa Cerquetti Collegio Nuovo, University of Pavia, Italy Interview Day, CETL Lectureship in Statistics, PSC & Dept of Mathematics and Statistics Lancaster University, April 29, 2009

2 Outline Some background on Bayesian Nonparametrics Exchangeability and de Finetti s theorem What we need for BNP How to construct a.s. discrete priors : resorting to Pitman & Gnedin work A note on priors obtained from exponentially tilted PK models On a Gibbs characterization of N-GG priors Generalized Chinese Restaurant rules for GIBBS partitions On a class of priors based on subordination of Stable processes.

3 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background on BNP In BNP no assumption is made on the law P on (S, S ) governing (X n ) n 1 w.r.t. belonging to a parametric family, we just assume exchangeability i.e. P (X 1 B 1,..., X n B n ) = P (X π(1) B 1,..., X π(n) B n ) the order of the observed values is assumed to be irrelevant to the inference P is invariant to the family of finite permutations

4 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background: de Finetti s representation theorem exchangeable p.m.s are mixtures of product p.m.s P (B) = P P(B i )q(dp) there exists a random probability measure P on (S, S) such that X i P i.i.d P q (the prior) on (P, P) is the law of the a.s. limit of the empirical distributions: 1 n i n δ Xi ( ) P( ) q i=1

5 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background: What we need for BNP inference Propose Priors i.e. random probability measures (a.s. discrete) by specifying finite dimensional distributions by means of Infinite sum representations by Normalization of Completely Random Measures Calculate Posteriors Without conjugacy: derive Predictive distributions for generalized Pólya urn Gibbs sampler. An example: Dirichlet Process Prior (first BNP prior proposed)

6 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background: Dirichlet process prior Finite dimensional distributions (Ferguson, 1973), P(B 1 ),..., P(B k ) Dir(θH(B 1 ),..., θh(b k )) for each measurable partition (B 1,..., B k ) of S, H a non-atomic p.m. on S θh( ) a positive measure on S, with θh(s) <.

7 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background: Dirichlet process prior Infinite sum representations (Kingman, 1975) P( ) := P i δ Xi ( ) i=1 P( ) has the Ferguson-Dirichlet (θh) distribution if (P i ) = (J i / i J i) have the law of the normalized ranked lengths of the points of a Poisson prox with mean intensity ρ(x) = θx 1 e x (Gamma) (X i ) iid H( ), independent of (P i ),

8 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background: Dirichlet process prior Normalization of a Completely Random Measure (Ferguson, 1973; Kingman, 1975) P( ) := µ( ) T has the Ferguson-Dirichlet (θh) distribution if µ( ) Ga(θH( ), 1), (Gamma CRM) T = i J i = µ(s) Ga(θ, 1) a.s. positive Lévy density ρ θ (x) = θx 1 e x (infinite total mass),

9 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Motivation BNP has dramatically re-emerged during the last ten years. Many contributions, very active community, a regular series of Workshops... BNP theory actually lies at the intersection among: Bayesian statistical inference, Random partitions theory, Lévy processes theory, Combinatorics... It happens that some new contributions turn out to be reformulations of results known in collateral topics. Clarifying these intersections/connections is important for future developments

10 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Pitman & Gnedin work on Poisson-Kingman (ρ α, γ) models A large class of models for random ranked atoms (P i ) is introduced in Pitman (2003), first generalizing Kingman (1975) (P i ) = (J i / i J i) Poisson-Kingman (ρ α ) on P 1 for ρ α the α-stable Lévy density, α (0, 1) then deriving the class of mixed PK(ρ α, γ) models PK(ρ α, γ) := 0 PK(ρ α t)γ(dt) PK(ρ α, γ) are models for exchangeable random partitions

11 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Pitman & Gnedin work on Poisson-Kingman (ρ α, γ) models (X 1,..., X n,... ) from a.s. discrete P induces a random partition Π = {A 1,..., A k,... } of N by i l X i = X l for each restriction to [n], for n j = A j, Pr(Π n = {A 1,..., A k }) = p(n 1,..., n k ) p (symmetric) is the exchangeable partition probability function (EPPF) which completely describes Π.

12 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Pitman & Gnedin work on Poisson-Kingman (ρ α, γ) models Gnedin & Pitman (2005) show that EPPFs induced by PK(ρ α, γ) models, for α (0, 1), are characterized by having Gibbs form i.e. p(n 1,..., n k ) = V n,k k (1 α) nj 1. Consistency conditions imply the non-negative weights V = (V n,k ) must satisfy the recursion j=1 V n,k = (n αk)v n+1,k + V n+1,k+1

13 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions A note on RPM derived from exponentially tilted PK models [Stat & Prob Lett 07] In Lijoi et al. (JASA 2005), by mimicking the finite dimensional distributions construction of the Dirichlet process priors with the Inverse Gaussian density playing the role of the Gamma density, Normalized Inverse Gaussian priors are proposed. the Inverse Gaussian is an Infinite Divisible (ID) r.v. arising from the exponential family generated by 1/2-stable r.v. (exponential tilting)

14 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions A note on RPM derived from exponentially tilted PK models [Stat & Prob Lett 07] I have shown that: since the same operation on α-stable r.v. for α (0, 1) produces Generalized Gamma distributions (Brix, 1999) we can derive Normalized Generalized Gamma priors by Infinite sum representations for (P i ) PK(ρ GG = ρ exp α ) and all the relevant quantities for BNP can be easily derived from Pitman s results: EPPF induced by N-GG priors sequential rules to describe predictive distributions distribution of the number of blocks of the N-GG partition.

15 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions On a Gibbs characterization of N-GG [Stat & Prob Lett, 08] In Lijoi et al. (Stat Sinica 2008), an analytic procedure is proposed to prove the following characterization: N-GG priors for α (0, 1) are the unique class of priors inducing EPPFs in Gibbs form and also admitting a construction by normalization of Completely random measures. I have shown that it suffices to merge two already known characterization results: Pitman (2003), for exponentially tilted PK models Gnedin & Pitman (2005), for Gibbs partitions of type α to prove a slightly more general characterization, circumventing the need for an analytical derivation.

16 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Generalized Chinese Restaurant rules for GIBBS partitions [arxiv: ] In Lijoi et al. (AoAP 2008) a BNP analysis under Gibbs priors with applications in species sampling problems is proposed. Given a sample from P with (P i ) PK(ρ α, γ) they obtain distributional posterior results for an additional sample of size m w.r.t. number of new distinct species number of new observations belonging to new species random partition induced by the additional sample

17 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Generalized Chinese Restaurant rules for GIBBS partitions [arxiv: ] I have derived a generalized (group sequential) version of the Pitman s sequential construction of EPPFs, (known as Chinese Restaurant Process) as the theoretical framework in which such analysis should and can be effectively embedded. According to Prof. Pitman this generalization: provides some clarification of Lijoi et al. s work, connects it better to the Gnedin & Pitman development, provides a better platform for future work.

18 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Generalized Chinese Restaurant rules for GIBBS partitions [arxiv: ] Given an infinite EPPF, p(n) = V n,k k j=1 (1 α) n j 1, assume that an unlimited number of group of customers arrives sequentially in a restaurant for n 1, given (n 1,..., n k ), the placement of the first group of n customers in a (n 1,..., n k ) configuration in k tables, the new group of m customers is: all seated at the old k tables in configuration (m 1,..., m k ), with probability p(m n) = V k n+m,k (n j α) mj, (1) V n,k j=1

19 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Generalized Chinese Restaurant rules for GIBBS partitions [arxiv: ] all seated at k new tables in configuration (s 1,..., s k ) with probability p(s n) = V k n+m,k+k (1 α) sj 1, (2) V n,k s < m are seated at k new tables in configuration (s 1,..., s k ) and the remaining m s customers at the k old tables in configuration (m 1,..., m k ) with probability j=1 p(m, s n) = V n+m,k+k V n,k k k (n j α) mj (1 α) sj 1. (3) j=1 j=1 These rules, by means of some combinatorial calculus and properties of generalized Stirling numbers, are enough to embed Lijoi et al. BNP analysis of Gibbs partitions.

20 On priors derived by subordination of Stable processes Ongoing work: On priors derived by convolution mixtures of Stable r.v. Each BNP prior correspond to an ID r.v. (with specific properties) Until now BNP researchers focus on Exponential Tilting of r.v.s (N-IG, N-GG) Finite Convolutions of ID r.v.s (Generalized Dirichlet Priors) A natural conjecture is to investigate the possibility to exploit the third method to obtain ID from ID, i.e. Convolution mixtures

21 On priors derived by subordination of Stable processes Ongoing work: On priors derived by convolution mixtures of Stable r.v. In Lévy process theory this corresponds to Bochner s subordination: the random time change of a subordinator by a subordinator independent of the original one. (e.g. Sato, 2000) I am investigating convolution mixtures of Stable r.v. with Generalized Gamma Convolutions (Bondesson, 1992). By exploiting recent results on BNP for normalized CRM, (James et al. 2009) it should be possible to derive a complete posterior and predictive BNP analysis.

22 Selected references Cerquetti, A. (2007) A note on Bayesian nonparametric priors derived from exponentially tilted Poisson-Kingman models. Stat. Prob. Lett.,77, Cerquetti, A. (2008) On a Gibbs characterization of normalized generalized Gamma processes. Stat. Prob. Letters, 78, Cerquetti, A. (2009) Generalized Chinese restaurant construction of exchangeable Gibbs partitions and related results. arxiv: v1 [math.pr] (Submitted) Ferguson, T. S. (1973) A Bayesian analysis of some nonparametric problems. Ann. Stat., 1, 2, Gnedin, S. and Pitman, J. (2005) Exchangeable Gibbs partitions and Stirling triangles arxiv:math.pr/ Kingman, J.F.C. (1975) Random discrete distributions. JRSS B, 37, Lijoi, A., Mena, R. and Prünster, I. (2005) Hierarchical mixture modeling with normalized Inverse-Gaussian priors. JASA, 100, Lijoi, A., Pruenster, I. and Walker, S.G. (2008) Investigating nonparametric priors with Gibbs structure. Stat. Sinica, 18, Lijoi, A. Pruenster, I. and Walker, S.G. (2008) Bayesian nonparametric estimators derived from conditional Gibbs structures. Ann. Appl. Probab. 18, Pitman, J. (2003) Poisson-Kingman partitions. IMS, LN 40, Institute of Mathematical Statistics, Hayward, California. (arxiv:math.pr/ ).

23 For those of you interested... 7th Workshop on Bayesian Nonparametrics Collegio Carlo Alberto, Moncalieri, Torino (Italy) June 21 to 25, 2009 Registration deadline: 30th April

Dependent hierarchical processes for multi armed bandits

Dependent hierarchical processes for multi armed bandits Dependent hierarchical processes for multi armed bandits Federico Camerlenghi University of Bologna, BIDSA & Collegio Carlo Alberto First Italian meeting on Probability and Mathematical Statistics, Torino

More information

On some distributional properties of Gibbs-type priors

On some distributional properties of Gibbs-type priors On some distributional properties of Gibbs-type priors Igor Prünster University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics Workshop ICERM, 21st September 2012 Joint work with: P. De Blasi,

More information

Discussion of On simulation and properties of the stable law by L. Devroye and L. James

Discussion of On simulation and properties of the stable law by L. Devroye and L. James Stat Methods Appl (2014) 23:371 377 DOI 10.1007/s10260-014-0269-4 Discussion of On simulation and properties of the stable law by L. Devroye and L. James Antonio Lijoi Igor Prünster Accepted: 16 May 2014

More information

A marginal sampler for σ-stable Poisson-Kingman mixture models

A marginal sampler for σ-stable Poisson-Kingman mixture models A marginal sampler for σ-stable Poisson-Kingman mixture models joint work with Yee Whye Teh and Stefano Favaro María Lomelí Gatsby Unit, University College London Talk at the BNP 10 Raleigh, North Carolina

More information

Slice sampling σ stable Poisson Kingman mixture models

Slice sampling σ stable Poisson Kingman mixture models ISSN 2279-9362 Slice sampling σ stable Poisson Kingman mixture models Stefano Favaro S.G. Walker No. 324 December 2013 www.carloalberto.org/research/working-papers 2013 by Stefano Favaro and S.G. Walker.

More information

On the posterior structure of NRMI

On the posterior structure of NRMI On the posterior structure of NRMI Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with L.F. James and A. Lijoi Isaac Newton Institute, BNR Programme, 8th August 2007 Outline

More information

Lecture 16-17: Bayesian Nonparametrics I. STAT 6474 Instructor: Hongxiao Zhu

Lecture 16-17: Bayesian Nonparametrics I. STAT 6474 Instructor: Hongxiao Zhu Lecture 16-17: Bayesian Nonparametrics I STAT 6474 Instructor: Hongxiao Zhu Plan for today Why Bayesian Nonparametrics? Dirichlet Distribution and Dirichlet Processes. 2 Parameter and Patterns Reference:

More information

Bayesian nonparametric models for bipartite graphs

Bayesian nonparametric models for bipartite graphs Bayesian nonparametric models for bipartite graphs François Caron Department of Statistics, Oxford Statistics Colloquium, Harvard University November 11, 2013 F. Caron 1 / 27 Bipartite networks Readers/Customers

More information

Bayesian nonparametric latent feature models

Bayesian nonparametric latent feature models Bayesian nonparametric latent feature models Indian Buffet process, beta process, and related models François Caron Department of Statistics, Oxford Applied Bayesian Statistics Summer School Como, Italy

More information

Two Tales About Bayesian Nonparametric Modeling

Two Tales About Bayesian Nonparametric Modeling Two Tales About Bayesian Nonparametric Modeling Pierpaolo De Blasi Stefano Favaro Antonio Lijoi Ramsés H. Mena Igor Prünster Abstract Gibbs type priors represent a natural generalization of the Dirichlet

More information

Defining Predictive Probability Functions for Species Sampling Models

Defining Predictive Probability Functions for Species Sampling Models Defining Predictive Probability Functions for Species Sampling Models Jaeyong Lee Department of Statistics, Seoul National University leejyc@gmail.com Fernando A. Quintana Departamento de Estadísica, Pontificia

More information

Truncation error of a superposed gamma process in a decreasing order representation

Truncation error of a superposed gamma process in a decreasing order representation Truncation error of a superposed gamma process in a decreasing order representation B julyan.arbel@inria.fr Í www.julyanarbel.com Inria, Mistis, Grenoble, France Joint work with Igor Pru nster (Bocconi

More information

CS281B / Stat 241B : Statistical Learning Theory Lecture: #22 on 19 Apr Dirichlet Process I

CS281B / Stat 241B : Statistical Learning Theory Lecture: #22 on 19 Apr Dirichlet Process I X i Ν CS281B / Stat 241B : Statistical Learning Theory Lecture: #22 on 19 Apr 2004 Dirichlet Process I Lecturer: Prof. Michael Jordan Scribe: Daniel Schonberg dschonbe@eecs.berkeley.edu 22.1 Dirichlet

More information

The two-parameter generalization of Ewens random partition structure

The two-parameter generalization of Ewens random partition structure The two-parameter generalization of Ewens random partition structure Jim Pitman Technical Report No. 345 Department of Statistics U.C. Berkeley CA 94720 March 25, 1992 Reprinted with an appendix and updated

More information

arxiv: v1 [math.st] 28 Feb 2015

arxiv: v1 [math.st] 28 Feb 2015 Are Gibbs type priors the most natural generalization of the Dirichlet process? P. De Blasi 1, S. Favaro 1, A. Lijoi 2, R.H. Mena 3, I. Prünster 1 and M. Ruggiero 1 1 Università degli Studi di Torino and

More information

Bayesian Nonparametric Models for Ranking Data

Bayesian Nonparametric Models for Ranking Data Bayesian Nonparametric Models for Ranking Data François Caron 1, Yee Whye Teh 1 and Brendan Murphy 2 1 Dept of Statistics, University of Oxford, UK 2 School of Mathematical Sciences, University College

More information

Dependent Random Measures and Prediction

Dependent Random Measures and Prediction Dependent Random Measures and Prediction Igor Prünster University of Torino & Collegio Carlo Alberto 10th Bayesian Nonparametric Conference Raleigh, June 26, 2015 Joint wor with: Federico Camerlenghi,

More information

Bayesian Nonparametrics: Dirichlet Process

Bayesian Nonparametrics: Dirichlet Process Bayesian Nonparametrics: Dirichlet Process Yee Whye Teh Gatsby Computational Neuroscience Unit, UCL http://www.gatsby.ucl.ac.uk/~ywteh/teaching/npbayes2012 Dirichlet Process Cornerstone of modern Bayesian

More information

Dependent mixture models: clustering and borrowing information

Dependent mixture models: clustering and borrowing information ISSN 2279-9362 Dependent mixture models: clustering and borrowing information Antonio Lijoi Bernardo Nipoti Igor Pruenster No. 32 June 213 www.carloalberto.org/research/working-papers 213 by Antonio Lijoi,

More information

Bayesian nonparametrics

Bayesian nonparametrics Bayesian nonparametrics 1 Some preliminaries 1.1 de Finetti s theorem We will start our discussion with this foundational theorem. We will assume throughout all variables are defined on the probability

More information

On the Truncation Error of a Superposed Gamma Process

On the Truncation Error of a Superposed Gamma Process On the Truncation Error of a Superposed Gamma Process Julyan Arbel and Igor Prünster Abstract Completely random measures (CRMs) form a key ingredient of a wealth of stochastic models, in particular in

More information

Asymptotics for posterior hazards

Asymptotics for posterior hazards Asymptotics for posterior hazards Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with P. Di Biasi and G. Peccati Workshop on Limit Theorems and Applications Paris, 16th January

More information

Bayesian Nonparametrics

Bayesian Nonparametrics Bayesian Nonparametrics Peter Orbanz Columbia University PARAMETERS AND PATTERNS Parameters P(X θ) = Probability[data pattern] 3 2 1 0 1 2 3 5 0 5 Inference idea data = underlying pattern + independent

More information

Foundations of Nonparametric Bayesian Methods

Foundations of Nonparametric Bayesian Methods 1 / 27 Foundations of Nonparametric Bayesian Methods Part II: Models on the Simplex Peter Orbanz http://mlg.eng.cam.ac.uk/porbanz/npb-tutorial.html 2 / 27 Tutorial Overview Part I: Basics Part II: Models

More information

A Brief Overview of Nonparametric Bayesian Models

A Brief Overview of Nonparametric Bayesian Models A Brief Overview of Nonparametric Bayesian Models Eurandom Zoubin Ghahramani Department of Engineering University of Cambridge, UK zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin Also at Machine

More information

On Consistency of Nonparametric Normal Mixtures for Bayesian Density Estimation

On Consistency of Nonparametric Normal Mixtures for Bayesian Density Estimation On Consistency of Nonparametric Normal Mixtures for Bayesian Density Estimation Antonio LIJOI, Igor PRÜNSTER, andstepheng.walker The past decade has seen a remarkable development in the area of Bayesian

More information

Asymptotics for posterior hazards

Asymptotics for posterior hazards Asymptotics for posterior hazards Pierpaolo De Blasi University of Turin 10th August 2007, BNR Workshop, Isaac Newton Intitute, Cambridge, UK Joint work with Giovanni Peccati (Université Paris VI) and

More information

Priors for Random Count Matrices with Random or Fixed Row Sums

Priors for Random Count Matrices with Random or Fixed Row Sums Priors for Random Count Matrices with Random or Fixed Row Sums Mingyuan Zhou Joint work with Oscar Madrid and James Scott IROM Department, McCombs School of Business Department of Statistics and Data Sciences

More information

Truncation error of a superposed gamma process in a decreasing order representation

Truncation error of a superposed gamma process in a decreasing order representation Truncation error of a superposed gamma process in a decreasing order representation Julyan Arbel Inria Grenoble, Université Grenoble Alpes julyan.arbel@inria.fr Igor Prünster Bocconi University, Milan

More information

Bayesian nonparametric models of sparse and exchangeable random graphs

Bayesian nonparametric models of sparse and exchangeable random graphs Bayesian nonparametric models of sparse and exchangeable random graphs F. Caron & E. Fox Technical Report Discussion led by Esther Salazar Duke University May 16, 2014 (Reading group) May 16, 2014 1 /

More information

Exchangeable random partitions and random discrete probability measures: a brief tour guided by the Dirichlet Process

Exchangeable random partitions and random discrete probability measures: a brief tour guided by the Dirichlet Process Exchangeable random partitions and random discrete probability measures: a brief tour guided by the Dirichlet Process Notes for Oxford Statistics Grad Lecture Benjamin Bloem-Reddy benjamin.bloem-reddy@stats.ox.ac.uk

More information

Controlling the reinforcement in Bayesian non-parametric mixture models

Controlling the reinforcement in Bayesian non-parametric mixture models J. R. Statist. Soc. B (2007) 69, Part 4, pp. 715 740 Controlling the reinforcement in Bayesian non-parametric mixture models Antonio Lijoi, Università degli Studi di Pavia, Italy Ramsés H. Mena Universidad

More information

Department of Statistics. University of California. Berkeley, CA May 1998

Department of Statistics. University of California. Berkeley, CA May 1998 Prediction rules for exchangeable sequences related to species sampling 1 by Ben Hansen and Jim Pitman Technical Report No. 520 Department of Statistics University of California 367 Evans Hall # 3860 Berkeley,

More information

arxiv: v2 [math.st] 27 May 2014

arxiv: v2 [math.st] 27 May 2014 Full Bayesian inference with hazard mixture models Julyan Arbel a, Antonio Lijoi b,a, Bernardo Nipoti c,a, arxiv:145.6628v2 [math.st] 27 May 214 a Collegio Carlo Alberto, via Real Collegio, 3, 124 Moncalieri,

More information

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model

More information

NONPARAMETRIC BAYESIAN INFERENCE ON PLANAR SHAPES

NONPARAMETRIC BAYESIAN INFERENCE ON PLANAR SHAPES NONPARAMETRIC BAYESIAN INFERENCE ON PLANAR SHAPES Author: Abhishek Bhattacharya Coauthor: David Dunson Department of Statistical Science, Duke University 7 th Workshop on Bayesian Nonparametrics Collegio

More information

Bayesian Nonparametrics

Bayesian Nonparametrics Bayesian Nonparametrics Lorenzo Rosasco 9.520 Class 18 April 11, 2011 About this class Goal To give an overview of some of the basic concepts in Bayesian Nonparametrics. In particular, to discuss Dirichelet

More information

Hierarchical Mixture Modeling With Normalized Inverse-Gaussian Priors

Hierarchical Mixture Modeling With Normalized Inverse-Gaussian Priors Hierarchical Mixture Modeling With Normalized Inverse-Gaussian Priors Antonio LIJOI, RamsésH.MENA, and Igor PRÜNSTER In recent years the Dirichlet process prior has experienced a great success in the context

More information

Exchangeable random partitions for statistical and economic modelling

Exchangeable random partitions for statistical and economic modelling Exchangeable random partitions for statistical and economic modelling Antonio Lijoi 1, Pietro Muliere 2, Igor Prünster 3 and Filippo Taddei 4 1 Dipartimento Economia Politica e Metodi Quantitatavi, Università

More information

Defining Predictive Probability Functions for Species Sampling Models

Defining Predictive Probability Functions for Species Sampling Models Defining Predictive Probability Functions for Species Sampling Models Jaeyong Lee Department of Statistics, Seoul National University leejyc@gmail.com Fernando A. Quintana Departamento de Estadísica, Pontificia

More information

Dirichlet Process. Yee Whye Teh, University College London

Dirichlet Process. Yee Whye Teh, University College London Dirichlet Process Yee Whye Teh, University College London Related keywords: Bayesian nonparametrics, stochastic processes, clustering, infinite mixture model, Blackwell-MacQueen urn scheme, Chinese restaurant

More information

An ergodic theorem for partially exchangeable random partitions

An ergodic theorem for partially exchangeable random partitions Electron. Commun. Probab. 22 (2017), no. 64, 1 10. DOI: 10.1214/17-ECP95 ISSN: 1083-589X ELECTRONIC COMMUNICATIONS in PROBABILITY An ergodic theorem for partially exchangeable random partitions Jim Pitman

More information

Bayesian estimation of the discrepancy with misspecified parametric models

Bayesian estimation of the discrepancy with misspecified parametric models Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012

More information

Spatial Normalized Gamma Process

Spatial Normalized Gamma Process Spatial Normalized Gamma Process Vinayak Rao Yee Whye Teh Presented at NIPS 2009 Discussion and Slides by Eric Wang June 23, 2010 Outline Introduction Motivation The Gamma Process Spatial Normalized Gamma

More information

On Simulations form the Two-Parameter. Poisson-Dirichlet Process and the Normalized. Inverse-Gaussian Process

On Simulations form the Two-Parameter. Poisson-Dirichlet Process and the Normalized. Inverse-Gaussian Process On Simulations form the Two-Parameter arxiv:1209.5359v1 [stat.co] 24 Sep 2012 Poisson-Dirichlet Process and the Normalized Inverse-Gaussian Process Luai Al Labadi and Mahmoud Zarepour May 8, 2018 ABSTRACT

More information

Full Bayesian inference with hazard mixture models

Full Bayesian inference with hazard mixture models ISSN 2279-9362 Full Bayesian inference with hazard mixture models Julyan Arbel Antonio Lijoi Bernardo Nipoti No. 381 December 214 www.carloalberto.org/research/working-papers 214 by Julyan Arbel, Antonio

More information

Distance-Based Probability Distribution for Set Partitions with Applications to Bayesian Nonparametrics

Distance-Based Probability Distribution for Set Partitions with Applications to Bayesian Nonparametrics Distance-Based Probability Distribution for Set Partitions with Applications to Bayesian Nonparametrics David B. Dahl August 5, 2008 Abstract Integration of several types of data is a burgeoning field.

More information

Bayesian Nonparametrics for Speech and Signal Processing

Bayesian Nonparametrics for Speech and Signal Processing Bayesian Nonparametrics for Speech and Signal Processing Michael I. Jordan University of California, Berkeley June 28, 2011 Acknowledgments: Emily Fox, Erik Sudderth, Yee Whye Teh, and Romain Thibaux Computer

More information

Non-Parametric Bayes

Non-Parametric Bayes Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian

More information

Non-parametric Clustering with Dirichlet Processes

Non-parametric Clustering with Dirichlet Processes Non-parametric Clustering with Dirichlet Processes Timothy Burns SUNY at Buffalo Mar. 31 2009 T. Burns (SUNY at Buffalo) Non-parametric Clustering with Dirichlet Processes Mar. 31 2009 1 / 24 Introduction

More information

Bayesian Nonparametric Models on Decomposable Graphs

Bayesian Nonparametric Models on Decomposable Graphs Bayesian Nonparametric Models on Decomposable Graphs François Caron INRIA Bordeaux Sud Ouest Institut de Mathématiques de Bordeaux University of Bordeaux, France francois.caron@inria.fr Arnaud Doucet Departments

More information

Bayesian Nonparametric Models

Bayesian Nonparametric Models Bayesian Nonparametric Models David M. Blei Columbia University December 15, 2015 Introduction We have been looking at models that posit latent structure in high dimensional data. We use the posterior

More information

Feature Allocations, Probability Functions, and Paintboxes

Feature Allocations, Probability Functions, and Paintboxes Feature Allocations, Probability Functions, and Paintboxes Tamara Broderick, Jim Pitman, Michael I. Jordan Abstract The problem of inferring a clustering of a data set has been the subject of much research

More information

Abstract These lecture notes are largely based on Jim Pitman s textbook Combinatorial Stochastic Processes, Bertoin s textbook Random fragmentation

Abstract These lecture notes are largely based on Jim Pitman s textbook Combinatorial Stochastic Processes, Bertoin s textbook Random fragmentation Abstract These lecture notes are largely based on Jim Pitman s textbook Combinatorial Stochastic Processes, Bertoin s textbook Random fragmentation and coagulation processes, and various papers. It is

More information

CSCI 5822 Probabilistic Model of Human and Machine Learning. Mike Mozer University of Colorado

CSCI 5822 Probabilistic Model of Human and Machine Learning. Mike Mozer University of Colorado CSCI 5822 Probabilistic Model of Human and Machine Learning Mike Mozer University of Colorado Topics Language modeling Hierarchical processes Pitman-Yor processes Based on work of Teh (2006), A hierarchical

More information

Bayesian Nonparametric Meta-Analysis Model George Karabatsos University of Illinois-Chicago (UIC)

Bayesian Nonparametric Meta-Analysis Model George Karabatsos University of Illinois-Chicago (UIC) Bayesian Nonparametric Meta-Analysis Model George Karabatsos University of Illinois-Chicago (UIC) Collaborators: Elizabeth Talbott, UIC. Stephen Walker, UT-Austin. August 9, 5, 4:5-4:45pm JSM 5 Meeting,

More information

Exchangeable Hoeffding decompositions over finite sets: a combinatorial characterization and counterexamples

Exchangeable Hoeffding decompositions over finite sets: a combinatorial characterization and counterexamples Exchangeable Hoeffding decompositions over finite sets: a combinatorial characterization and counterexamples Omar El-Dakkak 1, Giovanni Peccati 2, Igor Prünster 3 1 Université Paris-Ouest, France E-mail:

More information

arxiv: v1 [stat.ml] 20 Nov 2012

arxiv: v1 [stat.ml] 20 Nov 2012 A survey of non-exchangeable priors for Bayesian nonparametric models arxiv:1211.4798v1 [stat.ml] 20 Nov 2012 Nicholas J. Foti 1 and Sinead Williamson 2 1 Department of Computer Science, Dartmouth College

More information

REVERSIBLE MARKOV STRUCTURES ON DIVISIBLE SET PAR- TITIONS

REVERSIBLE MARKOV STRUCTURES ON DIVISIBLE SET PAR- TITIONS Applied Probability Trust (29 September 2014) REVERSIBLE MARKOV STRUCTURES ON DIVISIBLE SET PAR- TITIONS HARRY CRANE, Rutgers University PETER MCCULLAGH, University of Chicago Abstract We study k-divisible

More information

Infinite latent feature models and the Indian Buffet Process

Infinite latent feature models and the Indian Buffet Process p.1 Infinite latent feature models and the Indian Buffet Process Tom Griffiths Cognitive and Linguistic Sciences Brown University Joint work with Zoubin Ghahramani p.2 Beyond latent classes Unsupervised

More information

Bayesian nonparametric latent feature models

Bayesian nonparametric latent feature models Bayesian nonparametric latent feature models François Caron UBC October 2, 2007 / MLRG François Caron (UBC) Bayes. nonparametric latent feature models October 2, 2007 / MLRG 1 / 29 Overview 1 Introduction

More information

Probabilistic Graphical Models

Probabilistic Graphical Models School of Computer Science Probabilistic Graphical Models Infinite Feature Models: The Indian Buffet Process Eric Xing Lecture 21, April 2, 214 Acknowledgement: slides first drafted by Sinead Williamson

More information

Exchangeability. Peter Orbanz. Columbia University

Exchangeability. Peter Orbanz. Columbia University Exchangeability Peter Orbanz Columbia University PARAMETERS AND PATTERNS Parameters P(X θ) = Probability[data pattern] 3 2 1 0 1 2 3 5 0 5 Inference idea data = underlying pattern + independent noise Peter

More information

Nonparametric Bayesian Methods: Models, Algorithms, and Applications (Day 5)

Nonparametric Bayesian Methods: Models, Algorithms, and Applications (Day 5) Nonparametric Bayesian Methods: Models, Algorithms, and Applications (Day 5) Tamara Broderick ITT Career Development Assistant Professor Electrical Engineering & Computer Science MIT Bayes Foundations

More information

Hybrid Dirichlet processes for functional data

Hybrid Dirichlet processes for functional data Hybrid Dirichlet processes for functional data Sonia Petrone Università Bocconi, Milano Joint work with Michele Guindani - U.T. MD Anderson Cancer Center, Houston and Alan Gelfand - Duke University, USA

More information

Unit-rate Poisson representations of completely random measures

Unit-rate Poisson representations of completely random measures Unit-rate Poisson representations of completely random measures Peter Orbanz and Sinead Williamson Abstract: Constructive definitions of discrete random measures, which specify a sampling procedure for

More information

Bayesian Nonparametrics: Models Based on the Dirichlet Process

Bayesian Nonparametrics: Models Based on the Dirichlet Process Bayesian Nonparametrics: Models Based on the Dirichlet Process Alessandro Panella Department of Computer Science University of Illinois at Chicago Machine Learning Seminar Series February 18, 2013 Alessandro

More information

Normalized kernel-weighted random measures

Normalized kernel-weighted random measures Normalized kernel-weighted random measures Jim Griffin University of Kent 1 August 27 Outline 1 Introduction 2 Ornstein-Uhlenbeck DP 3 Generalisations Bayesian Density Regression We observe data (x 1,

More information

An Infinite Product of Sparse Chinese Restaurant Processes

An Infinite Product of Sparse Chinese Restaurant Processes An Infinite Product of Sparse Chinese Restaurant Processes Yarin Gal Tomoharu Iwata Zoubin Ghahramani yg279@cam.ac.uk CRP quick recap The Chinese restaurant process (CRP) Distribution over partitions of

More information

New Dirichlet Mean Identities

New Dirichlet Mean Identities Hong Kong University of Science and Technology Isaac Newton Institute, August 10, 2007 Origins CIFARELLI, D. M. and REGAZZINI, E. (1979). Considerazioni generali sull impostazione bayesiana di problemi

More information

Bayesian Statistics. Debdeep Pati Florida State University. April 3, 2017

Bayesian Statistics. Debdeep Pati Florida State University. April 3, 2017 Bayesian Statistics Debdeep Pati Florida State University April 3, 2017 Finite mixture model The finite mixture of normals can be equivalently expressed as y i N(µ Si ; τ 1 S i ), S i k π h δ h h=1 δ h

More information

Bayesian Nonparametric Mixture, Admixture, and Language Models

Bayesian Nonparametric Mixture, Admixture, and Language Models Bayesian Nonparametric Mixture, Admixture, and Language Models Yee Whye Teh University of Oxford Nov 2015 Overview Bayesian nonparametrics and random probability measures Mixture models and clustering

More information

Gentle Introduction to Infinite Gaussian Mixture Modeling

Gentle Introduction to Infinite Gaussian Mixture Modeling Gentle Introduction to Infinite Gaussian Mixture Modeling with an application in neuroscience By Frank Wood Rasmussen, NIPS 1999 Neuroscience Application: Spike Sorting Important in neuroscience and for

More information

arxiv: v2 [math.pr] 26 Aug 2017

arxiv: v2 [math.pr] 26 Aug 2017 Ordered and size-biased frequencies in GEM and Gibbs models for species sampling Jim Pitman Yuri Yakubovich arxiv:1704.04732v2 [math.pr] 26 Aug 2017 March 14, 2018 Abstract We describe the distribution

More information

Flexible Bayesian Nonparametric Priors and Bayesian Computational Methods

Flexible Bayesian Nonparametric Priors and Bayesian Computational Methods Universidad Carlos III de Madrid Repositorio institucional e-archivo Tesis http://e-archivo.uc3m.es Tesis Doctorales 2016-02-29 Flexible Bayesian Nonparametric Priors and Bayesian Computational Methods

More information

The Indian Buffet Process: An Introduction and Review

The Indian Buffet Process: An Introduction and Review Journal of Machine Learning Research 12 (2011) 1185-1224 Submitted 3/10; Revised 3/11; Published 4/11 The Indian Buffet Process: An Introduction and Review Thomas L. Griffiths Department of Psychology

More information

Lecture 3a: Dirichlet processes

Lecture 3a: Dirichlet processes Lecture 3a: Dirichlet processes Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London c.archambeau@cs.ucl.ac.uk Advanced Topics

More information

STAT Advanced Bayesian Inference

STAT Advanced Bayesian Inference 1 / 32 STAT 625 - Advanced Bayesian Inference Meng Li Department of Statistics Jan 23, 218 The Dirichlet distribution 2 / 32 θ Dirichlet(a 1,...,a k ) with density p(θ 1,θ 2,...,θ k ) = k j=1 Γ(a j) Γ(

More information

Hierarchical Bayesian Languge Model Based on Pitman-Yor Processes. Yee Whye Teh

Hierarchical Bayesian Languge Model Based on Pitman-Yor Processes. Yee Whye Teh Hierarchical Bayesian Languge Model Based on Pitman-Yor Processes Yee Whye Teh Probabilistic model of language n-gram model Utility i-1 P(word i word i-n+1 ) Typically, trigram model (n=3) e.g., speech,

More information

Some Developments of the Normalized Random Measures with Independent Increments

Some Developments of the Normalized Random Measures with Independent Increments Sankhyā : The Indian Journal of Statistics 2006, Volume 68, Part 3, pp. 46-487 c 2006, Indian Statistical Institute Some Developments of the Normalized Random Measures with Independent Increments Laura

More information

An Alternative Prior Process for Nonparametric Bayesian Clustering

An Alternative Prior Process for Nonparametric Bayesian Clustering An Alternative Prior Process for Nonparametric Bayesian Clustering Hanna Wallach (UMass Amherst) Shane Jensen (UPenn) Lee Dicker (Harvard) Katherine Heller (Cambridge) Nonparametric Bayesian Clustering

More information

Clustering using Mixture Models

Clustering using Mixture Models Clustering using Mixture Models The full posterior of the Gaussian Mixture Model is p(x, Z, µ,, ) =p(x Z, µ, )p(z )p( )p(µ, ) data likelihood (Gaussian) correspondence prob. (Multinomial) mixture prior

More information

arxiv: v1 [stat.me] 22 Feb 2015

arxiv: v1 [stat.me] 22 Feb 2015 MIXTURE MODELS WITH A PRIOR ON THE NUMBER OF COMPONENTS JEFFREY W. MILLER AND MATTHEW T. HARRISON arxiv:1502.06241v1 [stat.me] 22 Feb 2015 Abstract. A natural Bayesian approach for mixture models with

More information

Chapter 4: Modelling

Chapter 4: Modelling Chapter 4: Modelling Exchangeability and Invariance Markus Harva 17.10. / Reading Circle on Bayesian Theory Outline 1 Introduction 2 Models via exchangeability 3 Models via invariance 4 Exercise Statistical

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 4 Problem: Density Estimation We have observed data, y 1,..., y n, drawn independently from some unknown

More information

Bayesian nonparametric models for bipartite graphs

Bayesian nonparametric models for bipartite graphs Bayesian nonparametric models for bipartite graphs François Caron INRIA IMB - University of Bordeaux Talence, France Francois.Caron@inria.fr Abstract We develop a novel Bayesian nonparametric model for

More information

Infinite Latent Feature Models and the Indian Buffet Process

Infinite Latent Feature Models and the Indian Buffet Process Infinite Latent Feature Models and the Indian Buffet Process Thomas L. Griffiths Cognitive and Linguistic Sciences Brown University, Providence RI 292 tom griffiths@brown.edu Zoubin Ghahramani Gatsby Computational

More information

Bayesian non parametric approaches: an introduction

Bayesian non parametric approaches: an introduction Introduction Latent class models Latent feature models Conclusion & Perspectives Bayesian non parametric approaches: an introduction Pierre CHAINAIS Bordeaux - nov. 2012 Trajectory 1 Bayesian non parametric

More information

Dirichlet Processes: Tutorial and Practical Course

Dirichlet Processes: Tutorial and Practical Course Dirichlet Processes: Tutorial and Practical Course (updated) Yee Whye Teh Gatsby Computational Neuroscience Unit University College London August 2007 / MLSS Yee Whye Teh (Gatsby) DP August 2007 / MLSS

More information

Image segmentation combining Markov Random Fields and Dirichlet Processes

Image segmentation combining Markov Random Fields and Dirichlet Processes Image segmentation combining Markov Random Fields and Dirichlet Processes Jessica SODJO IMS, Groupe Signal Image, Talence Encadrants : A. Giremus, J.-F. Giovannelli, F. Caron, N. Dobigeon Jessica SODJO

More information

WXML Final Report: Chinese Restaurant Process

WXML Final Report: Chinese Restaurant Process WXML Final Report: Chinese Restaurant Process Dr. Noah Forman, Gerandy Brito, Alex Forney, Yiruey Chou, Chengning Li Spring 2017 1 Introduction The Chinese Restaurant Process (CRP) generates random partitions

More information

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition. Christian P. Robert The Bayesian Choice From Decision-Theoretic Foundations to Computational Implementation Second Edition With 23 Illustrations ^Springer" Contents Preface to the Second Edition Preface

More information

Hierarchical Dirichlet Processes

Hierarchical Dirichlet Processes Hierarchical Dirichlet Processes Yee Whye Teh, Michael I. Jordan, Matthew J. Beal and David M. Blei Computer Science Div., Dept. of Statistics Dept. of Computer Science University of California at Berkeley

More information

Chapter 3. Dirichlet Process

Chapter 3. Dirichlet Process Chapter 3 Dirichlet Process 3.1. The Dirichlet Process Prior 3.1.1. Definition The Dirichlet process (DP) is arguably the most popular BNP model for random probability measures (RPM), and plays a central

More information

MAD-Bayes: MAP-based Asymptotic Derivations from Bayes

MAD-Bayes: MAP-based Asymptotic Derivations from Bayes MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick Brian Kulis Michael I. Jordan Cat Clusters Mouse clusters Dog 1 Cat Clusters Dog Mouse Lizard Sheep Picture 1 Picture 2 Picture 3

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Bayesian nonparametric model for sparse dynamic networks

Bayesian nonparametric model for sparse dynamic networks Bayesian nonparametric model for sparse dynamic networks Konstantina Palla, François Caron and Yee Whye Teh 23rd of February, 2018 University of Glasgow Konstantina Palla 1 / 38 BIG PICTURE Interested

More information

Hierarchical Species Sampling Models

Hierarchical Species Sampling Models Hierarchical Species Sampling Models Federico Bassetti a Roberto Casarin b Luca Rossini c a Polytechnic University of Milan, Italy b Ca Foscari University of Venice, Italy c Free University of Bozen-Bolzano,

More information

Sharing Clusters Among Related Groups: Hierarchical Dirichlet Processes

Sharing Clusters Among Related Groups: Hierarchical Dirichlet Processes Sharing Clusters Among Related Groups: Hierarchical Dirichlet Processes Yee Whye Teh (1), Michael I. Jordan (1,2), Matthew J. Beal (3) and David M. Blei (1) (1) Computer Science Div., (2) Dept. of Statistics

More information

Applied Nonparametric Bayes

Applied Nonparametric Bayes Applied Nonparametric Bayes Michael I. Jordan Department of Electrical Engineering and Computer Science Department of Statistics University of California, Berkeley http://www.cs.berkeley.edu/ jordan Acknowledgments:

More information