Bayesian Nonparametrics: some contributions to construction and properties of prior distributions
|
|
- Dorcas Summers
- 5 years ago
- Views:
Transcription
1 Bayesian Nonparametrics: some contributions to construction and properties of prior distributions Annalisa Cerquetti Collegio Nuovo, University of Pavia, Italy Interview Day, CETL Lectureship in Statistics, PSC & Dept of Mathematics and Statistics Lancaster University, April 29, 2009
2 Outline Some background on Bayesian Nonparametrics Exchangeability and de Finetti s theorem What we need for BNP How to construct a.s. discrete priors : resorting to Pitman & Gnedin work A note on priors obtained from exponentially tilted PK models On a Gibbs characterization of N-GG priors Generalized Chinese Restaurant rules for GIBBS partitions On a class of priors based on subordination of Stable processes.
3 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background on BNP In BNP no assumption is made on the law P on (S, S ) governing (X n ) n 1 w.r.t. belonging to a parametric family, we just assume exchangeability i.e. P (X 1 B 1,..., X n B n ) = P (X π(1) B 1,..., X π(n) B n ) the order of the observed values is assumed to be irrelevant to the inference P is invariant to the family of finite permutations
4 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background: de Finetti s representation theorem exchangeable p.m.s are mixtures of product p.m.s P (B) = P P(B i )q(dp) there exists a random probability measure P on (S, S) such that X i P i.i.d P q (the prior) on (P, P) is the law of the a.s. limit of the empirical distributions: 1 n i n δ Xi ( ) P( ) q i=1
5 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background: What we need for BNP inference Propose Priors i.e. random probability measures (a.s. discrete) by specifying finite dimensional distributions by means of Infinite sum representations by Normalization of Completely Random Measures Calculate Posteriors Without conjugacy: derive Predictive distributions for generalized Pólya urn Gibbs sampler. An example: Dirichlet Process Prior (first BNP prior proposed)
6 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background: Dirichlet process prior Finite dimensional distributions (Ferguson, 1973), P(B 1 ),..., P(B k ) Dir(θH(B 1 ),..., θh(b k )) for each measurable partition (B 1,..., B k ) of S, H a non-atomic p.m. on S θh( ) a positive measure on S, with θh(s) <.
7 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background: Dirichlet process prior Infinite sum representations (Kingman, 1975) P( ) := P i δ Xi ( ) i=1 P( ) has the Ferguson-Dirichlet (θh) distribution if (P i ) = (J i / i J i) have the law of the normalized ranked lengths of the points of a Poisson prox with mean intensity ρ(x) = θx 1 e x (Gamma) (X i ) iid H( ), independent of (P i ),
8 Exchangeability and de Finetti s theorem What we need for BNP Three constructions for a.s discrete RPM: Dirichlet prior Some background: Dirichlet process prior Normalization of a Completely Random Measure (Ferguson, 1973; Kingman, 1975) P( ) := µ( ) T has the Ferguson-Dirichlet (θh) distribution if µ( ) Ga(θH( ), 1), (Gamma CRM) T = i J i = µ(s) Ga(θ, 1) a.s. positive Lévy density ρ θ (x) = θx 1 e x (infinite total mass),
9 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Motivation BNP has dramatically re-emerged during the last ten years. Many contributions, very active community, a regular series of Workshops... BNP theory actually lies at the intersection among: Bayesian statistical inference, Random partitions theory, Lévy processes theory, Combinatorics... It happens that some new contributions turn out to be reformulations of results known in collateral topics. Clarifying these intersections/connections is important for future developments
10 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Pitman & Gnedin work on Poisson-Kingman (ρ α, γ) models A large class of models for random ranked atoms (P i ) is introduced in Pitman (2003), first generalizing Kingman (1975) (P i ) = (J i / i J i) Poisson-Kingman (ρ α ) on P 1 for ρ α the α-stable Lévy density, α (0, 1) then deriving the class of mixed PK(ρ α, γ) models PK(ρ α, γ) := 0 PK(ρ α t)γ(dt) PK(ρ α, γ) are models for exchangeable random partitions
11 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Pitman & Gnedin work on Poisson-Kingman (ρ α, γ) models (X 1,..., X n,... ) from a.s. discrete P induces a random partition Π = {A 1,..., A k,... } of N by i l X i = X l for each restriction to [n], for n j = A j, Pr(Π n = {A 1,..., A k }) = p(n 1,..., n k ) p (symmetric) is the exchangeable partition probability function (EPPF) which completely describes Π.
12 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Pitman & Gnedin work on Poisson-Kingman (ρ α, γ) models Gnedin & Pitman (2005) show that EPPFs induced by PK(ρ α, γ) models, for α (0, 1), are characterized by having Gibbs form i.e. p(n 1,..., n k ) = V n,k k (1 α) nj 1. Consistency conditions imply the non-negative weights V = (V n,k ) must satisfy the recursion j=1 V n,k = (n αk)v n+1,k + V n+1,k+1
13 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions A note on RPM derived from exponentially tilted PK models [Stat & Prob Lett 07] In Lijoi et al. (JASA 2005), by mimicking the finite dimensional distributions construction of the Dirichlet process priors with the Inverse Gaussian density playing the role of the Gamma density, Normalized Inverse Gaussian priors are proposed. the Inverse Gaussian is an Infinite Divisible (ID) r.v. arising from the exponential family generated by 1/2-stable r.v. (exponential tilting)
14 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions A note on RPM derived from exponentially tilted PK models [Stat & Prob Lett 07] I have shown that: since the same operation on α-stable r.v. for α (0, 1) produces Generalized Gamma distributions (Brix, 1999) we can derive Normalized Generalized Gamma priors by Infinite sum representations for (P i ) PK(ρ GG = ρ exp α ) and all the relevant quantities for BNP can be easily derived from Pitman s results: EPPF induced by N-GG priors sequential rules to describe predictive distributions distribution of the number of blocks of the N-GG partition.
15 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions On a Gibbs characterization of N-GG [Stat & Prob Lett, 08] In Lijoi et al. (Stat Sinica 2008), an analytic procedure is proposed to prove the following characterization: N-GG priors for α (0, 1) are the unique class of priors inducing EPPFs in Gibbs form and also admitting a construction by normalization of Completely random measures. I have shown that it suffices to merge two already known characterization results: Pitman (2003), for exponentially tilted PK models Gnedin & Pitman (2005), for Gibbs partitions of type α to prove a slightly more general characterization, circumventing the need for an analytical derivation.
16 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Generalized Chinese Restaurant rules for GIBBS partitions [arxiv: ] In Lijoi et al. (AoAP 2008) a BNP analysis under Gibbs priors with applications in species sampling problems is proposed. Given a sample from P with (P i ) PK(ρ α, γ) they obtain distributional posterior results for an additional sample of size m w.r.t. number of new distinct species number of new observations belonging to new species random partition induced by the additional sample
17 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Generalized Chinese Restaurant rules for GIBBS partitions [arxiv: ] I have derived a generalized (group sequential) version of the Pitman s sequential construction of EPPFs, (known as Chinese Restaurant Process) as the theoretical framework in which such analysis should and can be effectively embedded. According to Prof. Pitman this generalization: provides some clarification of Lijoi et al. s work, connects it better to the Gnedin & Pitman development, provides a better platform for future work.
18 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Generalized Chinese Restaurant rules for GIBBS partitions [arxiv: ] Given an infinite EPPF, p(n) = V n,k k j=1 (1 α) n j 1, assume that an unlimited number of group of customers arrives sequentially in a restaurant for n 1, given (n 1,..., n k ), the placement of the first group of n customers in a (n 1,..., n k ) configuration in k tables, the new group of m customers is: all seated at the old k tables in configuration (m 1,..., m k ), with probability p(m n) = V k n+m,k (n j α) mj, (1) V n,k j=1
19 Some fundamental results of Pitman & Gnedin A note on RPM derived from exponentially tilted PK models On a Gibbs characterization of N-GG Generalized Chinese Restaurant rules for GIBBS partitions Generalized Chinese Restaurant rules for GIBBS partitions [arxiv: ] all seated at k new tables in configuration (s 1,..., s k ) with probability p(s n) = V k n+m,k+k (1 α) sj 1, (2) V n,k s < m are seated at k new tables in configuration (s 1,..., s k ) and the remaining m s customers at the k old tables in configuration (m 1,..., m k ) with probability j=1 p(m, s n) = V n+m,k+k V n,k k k (n j α) mj (1 α) sj 1. (3) j=1 j=1 These rules, by means of some combinatorial calculus and properties of generalized Stirling numbers, are enough to embed Lijoi et al. BNP analysis of Gibbs partitions.
20 On priors derived by subordination of Stable processes Ongoing work: On priors derived by convolution mixtures of Stable r.v. Each BNP prior correspond to an ID r.v. (with specific properties) Until now BNP researchers focus on Exponential Tilting of r.v.s (N-IG, N-GG) Finite Convolutions of ID r.v.s (Generalized Dirichlet Priors) A natural conjecture is to investigate the possibility to exploit the third method to obtain ID from ID, i.e. Convolution mixtures
21 On priors derived by subordination of Stable processes Ongoing work: On priors derived by convolution mixtures of Stable r.v. In Lévy process theory this corresponds to Bochner s subordination: the random time change of a subordinator by a subordinator independent of the original one. (e.g. Sato, 2000) I am investigating convolution mixtures of Stable r.v. with Generalized Gamma Convolutions (Bondesson, 1992). By exploiting recent results on BNP for normalized CRM, (James et al. 2009) it should be possible to derive a complete posterior and predictive BNP analysis.
22 Selected references Cerquetti, A. (2007) A note on Bayesian nonparametric priors derived from exponentially tilted Poisson-Kingman models. Stat. Prob. Lett.,77, Cerquetti, A. (2008) On a Gibbs characterization of normalized generalized Gamma processes. Stat. Prob. Letters, 78, Cerquetti, A. (2009) Generalized Chinese restaurant construction of exchangeable Gibbs partitions and related results. arxiv: v1 [math.pr] (Submitted) Ferguson, T. S. (1973) A Bayesian analysis of some nonparametric problems. Ann. Stat., 1, 2, Gnedin, S. and Pitman, J. (2005) Exchangeable Gibbs partitions and Stirling triangles arxiv:math.pr/ Kingman, J.F.C. (1975) Random discrete distributions. JRSS B, 37, Lijoi, A., Mena, R. and Prünster, I. (2005) Hierarchical mixture modeling with normalized Inverse-Gaussian priors. JASA, 100, Lijoi, A., Pruenster, I. and Walker, S.G. (2008) Investigating nonparametric priors with Gibbs structure. Stat. Sinica, 18, Lijoi, A. Pruenster, I. and Walker, S.G. (2008) Bayesian nonparametric estimators derived from conditional Gibbs structures. Ann. Appl. Probab. 18, Pitman, J. (2003) Poisson-Kingman partitions. IMS, LN 40, Institute of Mathematical Statistics, Hayward, California. (arxiv:math.pr/ ).
23 For those of you interested... 7th Workshop on Bayesian Nonparametrics Collegio Carlo Alberto, Moncalieri, Torino (Italy) June 21 to 25, 2009 Registration deadline: 30th April
Dependent hierarchical processes for multi armed bandits
Dependent hierarchical processes for multi armed bandits Federico Camerlenghi University of Bologna, BIDSA & Collegio Carlo Alberto First Italian meeting on Probability and Mathematical Statistics, Torino
More informationOn some distributional properties of Gibbs-type priors
On some distributional properties of Gibbs-type priors Igor Prünster University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics Workshop ICERM, 21st September 2012 Joint work with: P. De Blasi,
More informationDiscussion of On simulation and properties of the stable law by L. Devroye and L. James
Stat Methods Appl (2014) 23:371 377 DOI 10.1007/s10260-014-0269-4 Discussion of On simulation and properties of the stable law by L. Devroye and L. James Antonio Lijoi Igor Prünster Accepted: 16 May 2014
More informationA marginal sampler for σ-stable Poisson-Kingman mixture models
A marginal sampler for σ-stable Poisson-Kingman mixture models joint work with Yee Whye Teh and Stefano Favaro María Lomelí Gatsby Unit, University College London Talk at the BNP 10 Raleigh, North Carolina
More informationSlice sampling σ stable Poisson Kingman mixture models
ISSN 2279-9362 Slice sampling σ stable Poisson Kingman mixture models Stefano Favaro S.G. Walker No. 324 December 2013 www.carloalberto.org/research/working-papers 2013 by Stefano Favaro and S.G. Walker.
More informationOn the posterior structure of NRMI
On the posterior structure of NRMI Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with L.F. James and A. Lijoi Isaac Newton Institute, BNR Programme, 8th August 2007 Outline
More informationLecture 16-17: Bayesian Nonparametrics I. STAT 6474 Instructor: Hongxiao Zhu
Lecture 16-17: Bayesian Nonparametrics I STAT 6474 Instructor: Hongxiao Zhu Plan for today Why Bayesian Nonparametrics? Dirichlet Distribution and Dirichlet Processes. 2 Parameter and Patterns Reference:
More informationBayesian nonparametric models for bipartite graphs
Bayesian nonparametric models for bipartite graphs François Caron Department of Statistics, Oxford Statistics Colloquium, Harvard University November 11, 2013 F. Caron 1 / 27 Bipartite networks Readers/Customers
More informationBayesian nonparametric latent feature models
Bayesian nonparametric latent feature models Indian Buffet process, beta process, and related models François Caron Department of Statistics, Oxford Applied Bayesian Statistics Summer School Como, Italy
More informationTwo Tales About Bayesian Nonparametric Modeling
Two Tales About Bayesian Nonparametric Modeling Pierpaolo De Blasi Stefano Favaro Antonio Lijoi Ramsés H. Mena Igor Prünster Abstract Gibbs type priors represent a natural generalization of the Dirichlet
More informationDefining Predictive Probability Functions for Species Sampling Models
Defining Predictive Probability Functions for Species Sampling Models Jaeyong Lee Department of Statistics, Seoul National University leejyc@gmail.com Fernando A. Quintana Departamento de Estadísica, Pontificia
More informationTruncation error of a superposed gamma process in a decreasing order representation
Truncation error of a superposed gamma process in a decreasing order representation B julyan.arbel@inria.fr Í www.julyanarbel.com Inria, Mistis, Grenoble, France Joint work with Igor Pru nster (Bocconi
More informationCS281B / Stat 241B : Statistical Learning Theory Lecture: #22 on 19 Apr Dirichlet Process I
X i Ν CS281B / Stat 241B : Statistical Learning Theory Lecture: #22 on 19 Apr 2004 Dirichlet Process I Lecturer: Prof. Michael Jordan Scribe: Daniel Schonberg dschonbe@eecs.berkeley.edu 22.1 Dirichlet
More informationThe two-parameter generalization of Ewens random partition structure
The two-parameter generalization of Ewens random partition structure Jim Pitman Technical Report No. 345 Department of Statistics U.C. Berkeley CA 94720 March 25, 1992 Reprinted with an appendix and updated
More informationarxiv: v1 [math.st] 28 Feb 2015
Are Gibbs type priors the most natural generalization of the Dirichlet process? P. De Blasi 1, S. Favaro 1, A. Lijoi 2, R.H. Mena 3, I. Prünster 1 and M. Ruggiero 1 1 Università degli Studi di Torino and
More informationBayesian Nonparametric Models for Ranking Data
Bayesian Nonparametric Models for Ranking Data François Caron 1, Yee Whye Teh 1 and Brendan Murphy 2 1 Dept of Statistics, University of Oxford, UK 2 School of Mathematical Sciences, University College
More informationDependent Random Measures and Prediction
Dependent Random Measures and Prediction Igor Prünster University of Torino & Collegio Carlo Alberto 10th Bayesian Nonparametric Conference Raleigh, June 26, 2015 Joint wor with: Federico Camerlenghi,
More informationBayesian Nonparametrics: Dirichlet Process
Bayesian Nonparametrics: Dirichlet Process Yee Whye Teh Gatsby Computational Neuroscience Unit, UCL http://www.gatsby.ucl.ac.uk/~ywteh/teaching/npbayes2012 Dirichlet Process Cornerstone of modern Bayesian
More informationDependent mixture models: clustering and borrowing information
ISSN 2279-9362 Dependent mixture models: clustering and borrowing information Antonio Lijoi Bernardo Nipoti Igor Pruenster No. 32 June 213 www.carloalberto.org/research/working-papers 213 by Antonio Lijoi,
More informationBayesian nonparametrics
Bayesian nonparametrics 1 Some preliminaries 1.1 de Finetti s theorem We will start our discussion with this foundational theorem. We will assume throughout all variables are defined on the probability
More informationOn the Truncation Error of a Superposed Gamma Process
On the Truncation Error of a Superposed Gamma Process Julyan Arbel and Igor Prünster Abstract Completely random measures (CRMs) form a key ingredient of a wealth of stochastic models, in particular in
More informationAsymptotics for posterior hazards
Asymptotics for posterior hazards Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with P. Di Biasi and G. Peccati Workshop on Limit Theorems and Applications Paris, 16th January
More informationBayesian Nonparametrics
Bayesian Nonparametrics Peter Orbanz Columbia University PARAMETERS AND PATTERNS Parameters P(X θ) = Probability[data pattern] 3 2 1 0 1 2 3 5 0 5 Inference idea data = underlying pattern + independent
More informationFoundations of Nonparametric Bayesian Methods
1 / 27 Foundations of Nonparametric Bayesian Methods Part II: Models on the Simplex Peter Orbanz http://mlg.eng.cam.ac.uk/porbanz/npb-tutorial.html 2 / 27 Tutorial Overview Part I: Basics Part II: Models
More informationA Brief Overview of Nonparametric Bayesian Models
A Brief Overview of Nonparametric Bayesian Models Eurandom Zoubin Ghahramani Department of Engineering University of Cambridge, UK zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin Also at Machine
More informationOn Consistency of Nonparametric Normal Mixtures for Bayesian Density Estimation
On Consistency of Nonparametric Normal Mixtures for Bayesian Density Estimation Antonio LIJOI, Igor PRÜNSTER, andstepheng.walker The past decade has seen a remarkable development in the area of Bayesian
More informationAsymptotics for posterior hazards
Asymptotics for posterior hazards Pierpaolo De Blasi University of Turin 10th August 2007, BNR Workshop, Isaac Newton Intitute, Cambridge, UK Joint work with Giovanni Peccati (Université Paris VI) and
More informationPriors for Random Count Matrices with Random or Fixed Row Sums
Priors for Random Count Matrices with Random or Fixed Row Sums Mingyuan Zhou Joint work with Oscar Madrid and James Scott IROM Department, McCombs School of Business Department of Statistics and Data Sciences
More informationTruncation error of a superposed gamma process in a decreasing order representation
Truncation error of a superposed gamma process in a decreasing order representation Julyan Arbel Inria Grenoble, Université Grenoble Alpes julyan.arbel@inria.fr Igor Prünster Bocconi University, Milan
More informationBayesian nonparametric models of sparse and exchangeable random graphs
Bayesian nonparametric models of sparse and exchangeable random graphs F. Caron & E. Fox Technical Report Discussion led by Esther Salazar Duke University May 16, 2014 (Reading group) May 16, 2014 1 /
More informationExchangeable random partitions and random discrete probability measures: a brief tour guided by the Dirichlet Process
Exchangeable random partitions and random discrete probability measures: a brief tour guided by the Dirichlet Process Notes for Oxford Statistics Grad Lecture Benjamin Bloem-Reddy benjamin.bloem-reddy@stats.ox.ac.uk
More informationControlling the reinforcement in Bayesian non-parametric mixture models
J. R. Statist. Soc. B (2007) 69, Part 4, pp. 715 740 Controlling the reinforcement in Bayesian non-parametric mixture models Antonio Lijoi, Università degli Studi di Pavia, Italy Ramsés H. Mena Universidad
More informationDepartment of Statistics. University of California. Berkeley, CA May 1998
Prediction rules for exchangeable sequences related to species sampling 1 by Ben Hansen and Jim Pitman Technical Report No. 520 Department of Statistics University of California 367 Evans Hall # 3860 Berkeley,
More informationarxiv: v2 [math.st] 27 May 2014
Full Bayesian inference with hazard mixture models Julyan Arbel a, Antonio Lijoi b,a, Bernardo Nipoti c,a, arxiv:145.6628v2 [math.st] 27 May 214 a Collegio Carlo Alberto, via Real Collegio, 3, 124 Moncalieri,
More informationOutline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution
Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model
More informationNONPARAMETRIC BAYESIAN INFERENCE ON PLANAR SHAPES
NONPARAMETRIC BAYESIAN INFERENCE ON PLANAR SHAPES Author: Abhishek Bhattacharya Coauthor: David Dunson Department of Statistical Science, Duke University 7 th Workshop on Bayesian Nonparametrics Collegio
More informationBayesian Nonparametrics
Bayesian Nonparametrics Lorenzo Rosasco 9.520 Class 18 April 11, 2011 About this class Goal To give an overview of some of the basic concepts in Bayesian Nonparametrics. In particular, to discuss Dirichelet
More informationHierarchical Mixture Modeling With Normalized Inverse-Gaussian Priors
Hierarchical Mixture Modeling With Normalized Inverse-Gaussian Priors Antonio LIJOI, RamsésH.MENA, and Igor PRÜNSTER In recent years the Dirichlet process prior has experienced a great success in the context
More informationExchangeable random partitions for statistical and economic modelling
Exchangeable random partitions for statistical and economic modelling Antonio Lijoi 1, Pietro Muliere 2, Igor Prünster 3 and Filippo Taddei 4 1 Dipartimento Economia Politica e Metodi Quantitatavi, Università
More informationDefining Predictive Probability Functions for Species Sampling Models
Defining Predictive Probability Functions for Species Sampling Models Jaeyong Lee Department of Statistics, Seoul National University leejyc@gmail.com Fernando A. Quintana Departamento de Estadísica, Pontificia
More informationDirichlet Process. Yee Whye Teh, University College London
Dirichlet Process Yee Whye Teh, University College London Related keywords: Bayesian nonparametrics, stochastic processes, clustering, infinite mixture model, Blackwell-MacQueen urn scheme, Chinese restaurant
More informationAn ergodic theorem for partially exchangeable random partitions
Electron. Commun. Probab. 22 (2017), no. 64, 1 10. DOI: 10.1214/17-ECP95 ISSN: 1083-589X ELECTRONIC COMMUNICATIONS in PROBABILITY An ergodic theorem for partially exchangeable random partitions Jim Pitman
More informationBayesian estimation of the discrepancy with misspecified parametric models
Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012
More informationSpatial Normalized Gamma Process
Spatial Normalized Gamma Process Vinayak Rao Yee Whye Teh Presented at NIPS 2009 Discussion and Slides by Eric Wang June 23, 2010 Outline Introduction Motivation The Gamma Process Spatial Normalized Gamma
More informationOn Simulations form the Two-Parameter. Poisson-Dirichlet Process and the Normalized. Inverse-Gaussian Process
On Simulations form the Two-Parameter arxiv:1209.5359v1 [stat.co] 24 Sep 2012 Poisson-Dirichlet Process and the Normalized Inverse-Gaussian Process Luai Al Labadi and Mahmoud Zarepour May 8, 2018 ABSTRACT
More informationFull Bayesian inference with hazard mixture models
ISSN 2279-9362 Full Bayesian inference with hazard mixture models Julyan Arbel Antonio Lijoi Bernardo Nipoti No. 381 December 214 www.carloalberto.org/research/working-papers 214 by Julyan Arbel, Antonio
More informationDistance-Based Probability Distribution for Set Partitions with Applications to Bayesian Nonparametrics
Distance-Based Probability Distribution for Set Partitions with Applications to Bayesian Nonparametrics David B. Dahl August 5, 2008 Abstract Integration of several types of data is a burgeoning field.
More informationBayesian Nonparametrics for Speech and Signal Processing
Bayesian Nonparametrics for Speech and Signal Processing Michael I. Jordan University of California, Berkeley June 28, 2011 Acknowledgments: Emily Fox, Erik Sudderth, Yee Whye Teh, and Romain Thibaux Computer
More informationNon-Parametric Bayes
Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian
More informationNon-parametric Clustering with Dirichlet Processes
Non-parametric Clustering with Dirichlet Processes Timothy Burns SUNY at Buffalo Mar. 31 2009 T. Burns (SUNY at Buffalo) Non-parametric Clustering with Dirichlet Processes Mar. 31 2009 1 / 24 Introduction
More informationBayesian Nonparametric Models on Decomposable Graphs
Bayesian Nonparametric Models on Decomposable Graphs François Caron INRIA Bordeaux Sud Ouest Institut de Mathématiques de Bordeaux University of Bordeaux, France francois.caron@inria.fr Arnaud Doucet Departments
More informationBayesian Nonparametric Models
Bayesian Nonparametric Models David M. Blei Columbia University December 15, 2015 Introduction We have been looking at models that posit latent structure in high dimensional data. We use the posterior
More informationFeature Allocations, Probability Functions, and Paintboxes
Feature Allocations, Probability Functions, and Paintboxes Tamara Broderick, Jim Pitman, Michael I. Jordan Abstract The problem of inferring a clustering of a data set has been the subject of much research
More informationAbstract These lecture notes are largely based on Jim Pitman s textbook Combinatorial Stochastic Processes, Bertoin s textbook Random fragmentation
Abstract These lecture notes are largely based on Jim Pitman s textbook Combinatorial Stochastic Processes, Bertoin s textbook Random fragmentation and coagulation processes, and various papers. It is
More informationCSCI 5822 Probabilistic Model of Human and Machine Learning. Mike Mozer University of Colorado
CSCI 5822 Probabilistic Model of Human and Machine Learning Mike Mozer University of Colorado Topics Language modeling Hierarchical processes Pitman-Yor processes Based on work of Teh (2006), A hierarchical
More informationBayesian Nonparametric Meta-Analysis Model George Karabatsos University of Illinois-Chicago (UIC)
Bayesian Nonparametric Meta-Analysis Model George Karabatsos University of Illinois-Chicago (UIC) Collaborators: Elizabeth Talbott, UIC. Stephen Walker, UT-Austin. August 9, 5, 4:5-4:45pm JSM 5 Meeting,
More informationExchangeable Hoeffding decompositions over finite sets: a combinatorial characterization and counterexamples
Exchangeable Hoeffding decompositions over finite sets: a combinatorial characterization and counterexamples Omar El-Dakkak 1, Giovanni Peccati 2, Igor Prünster 3 1 Université Paris-Ouest, France E-mail:
More informationarxiv: v1 [stat.ml] 20 Nov 2012
A survey of non-exchangeable priors for Bayesian nonparametric models arxiv:1211.4798v1 [stat.ml] 20 Nov 2012 Nicholas J. Foti 1 and Sinead Williamson 2 1 Department of Computer Science, Dartmouth College
More informationREVERSIBLE MARKOV STRUCTURES ON DIVISIBLE SET PAR- TITIONS
Applied Probability Trust (29 September 2014) REVERSIBLE MARKOV STRUCTURES ON DIVISIBLE SET PAR- TITIONS HARRY CRANE, Rutgers University PETER MCCULLAGH, University of Chicago Abstract We study k-divisible
More informationInfinite latent feature models and the Indian Buffet Process
p.1 Infinite latent feature models and the Indian Buffet Process Tom Griffiths Cognitive and Linguistic Sciences Brown University Joint work with Zoubin Ghahramani p.2 Beyond latent classes Unsupervised
More informationBayesian nonparametric latent feature models
Bayesian nonparametric latent feature models François Caron UBC October 2, 2007 / MLRG François Caron (UBC) Bayes. nonparametric latent feature models October 2, 2007 / MLRG 1 / 29 Overview 1 Introduction
More informationProbabilistic Graphical Models
School of Computer Science Probabilistic Graphical Models Infinite Feature Models: The Indian Buffet Process Eric Xing Lecture 21, April 2, 214 Acknowledgement: slides first drafted by Sinead Williamson
More informationExchangeability. Peter Orbanz. Columbia University
Exchangeability Peter Orbanz Columbia University PARAMETERS AND PATTERNS Parameters P(X θ) = Probability[data pattern] 3 2 1 0 1 2 3 5 0 5 Inference idea data = underlying pattern + independent noise Peter
More informationNonparametric Bayesian Methods: Models, Algorithms, and Applications (Day 5)
Nonparametric Bayesian Methods: Models, Algorithms, and Applications (Day 5) Tamara Broderick ITT Career Development Assistant Professor Electrical Engineering & Computer Science MIT Bayes Foundations
More informationHybrid Dirichlet processes for functional data
Hybrid Dirichlet processes for functional data Sonia Petrone Università Bocconi, Milano Joint work with Michele Guindani - U.T. MD Anderson Cancer Center, Houston and Alan Gelfand - Duke University, USA
More informationUnit-rate Poisson representations of completely random measures
Unit-rate Poisson representations of completely random measures Peter Orbanz and Sinead Williamson Abstract: Constructive definitions of discrete random measures, which specify a sampling procedure for
More informationBayesian Nonparametrics: Models Based on the Dirichlet Process
Bayesian Nonparametrics: Models Based on the Dirichlet Process Alessandro Panella Department of Computer Science University of Illinois at Chicago Machine Learning Seminar Series February 18, 2013 Alessandro
More informationNormalized kernel-weighted random measures
Normalized kernel-weighted random measures Jim Griffin University of Kent 1 August 27 Outline 1 Introduction 2 Ornstein-Uhlenbeck DP 3 Generalisations Bayesian Density Regression We observe data (x 1,
More informationAn Infinite Product of Sparse Chinese Restaurant Processes
An Infinite Product of Sparse Chinese Restaurant Processes Yarin Gal Tomoharu Iwata Zoubin Ghahramani yg279@cam.ac.uk CRP quick recap The Chinese restaurant process (CRP) Distribution over partitions of
More informationNew Dirichlet Mean Identities
Hong Kong University of Science and Technology Isaac Newton Institute, August 10, 2007 Origins CIFARELLI, D. M. and REGAZZINI, E. (1979). Considerazioni generali sull impostazione bayesiana di problemi
More informationBayesian Statistics. Debdeep Pati Florida State University. April 3, 2017
Bayesian Statistics Debdeep Pati Florida State University April 3, 2017 Finite mixture model The finite mixture of normals can be equivalently expressed as y i N(µ Si ; τ 1 S i ), S i k π h δ h h=1 δ h
More informationBayesian Nonparametric Mixture, Admixture, and Language Models
Bayesian Nonparametric Mixture, Admixture, and Language Models Yee Whye Teh University of Oxford Nov 2015 Overview Bayesian nonparametrics and random probability measures Mixture models and clustering
More informationGentle Introduction to Infinite Gaussian Mixture Modeling
Gentle Introduction to Infinite Gaussian Mixture Modeling with an application in neuroscience By Frank Wood Rasmussen, NIPS 1999 Neuroscience Application: Spike Sorting Important in neuroscience and for
More informationarxiv: v2 [math.pr] 26 Aug 2017
Ordered and size-biased frequencies in GEM and Gibbs models for species sampling Jim Pitman Yuri Yakubovich arxiv:1704.04732v2 [math.pr] 26 Aug 2017 March 14, 2018 Abstract We describe the distribution
More informationFlexible Bayesian Nonparametric Priors and Bayesian Computational Methods
Universidad Carlos III de Madrid Repositorio institucional e-archivo Tesis http://e-archivo.uc3m.es Tesis Doctorales 2016-02-29 Flexible Bayesian Nonparametric Priors and Bayesian Computational Methods
More informationThe Indian Buffet Process: An Introduction and Review
Journal of Machine Learning Research 12 (2011) 1185-1224 Submitted 3/10; Revised 3/11; Published 4/11 The Indian Buffet Process: An Introduction and Review Thomas L. Griffiths Department of Psychology
More informationLecture 3a: Dirichlet processes
Lecture 3a: Dirichlet processes Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London c.archambeau@cs.ucl.ac.uk Advanced Topics
More informationSTAT Advanced Bayesian Inference
1 / 32 STAT 625 - Advanced Bayesian Inference Meng Li Department of Statistics Jan 23, 218 The Dirichlet distribution 2 / 32 θ Dirichlet(a 1,...,a k ) with density p(θ 1,θ 2,...,θ k ) = k j=1 Γ(a j) Γ(
More informationHierarchical Bayesian Languge Model Based on Pitman-Yor Processes. Yee Whye Teh
Hierarchical Bayesian Languge Model Based on Pitman-Yor Processes Yee Whye Teh Probabilistic model of language n-gram model Utility i-1 P(word i word i-n+1 ) Typically, trigram model (n=3) e.g., speech,
More informationSome Developments of the Normalized Random Measures with Independent Increments
Sankhyā : The Indian Journal of Statistics 2006, Volume 68, Part 3, pp. 46-487 c 2006, Indian Statistical Institute Some Developments of the Normalized Random Measures with Independent Increments Laura
More informationAn Alternative Prior Process for Nonparametric Bayesian Clustering
An Alternative Prior Process for Nonparametric Bayesian Clustering Hanna Wallach (UMass Amherst) Shane Jensen (UPenn) Lee Dicker (Harvard) Katherine Heller (Cambridge) Nonparametric Bayesian Clustering
More informationClustering using Mixture Models
Clustering using Mixture Models The full posterior of the Gaussian Mixture Model is p(x, Z, µ,, ) =p(x Z, µ, )p(z )p( )p(µ, ) data likelihood (Gaussian) correspondence prob. (Multinomial) mixture prior
More informationarxiv: v1 [stat.me] 22 Feb 2015
MIXTURE MODELS WITH A PRIOR ON THE NUMBER OF COMPONENTS JEFFREY W. MILLER AND MATTHEW T. HARRISON arxiv:1502.06241v1 [stat.me] 22 Feb 2015 Abstract. A natural Bayesian approach for mixture models with
More informationChapter 4: Modelling
Chapter 4: Modelling Exchangeability and Invariance Markus Harva 17.10. / Reading Circle on Bayesian Theory Outline 1 Introduction 2 Models via exchangeability 3 Models via invariance 4 Exercise Statistical
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 4 Problem: Density Estimation We have observed data, y 1,..., y n, drawn independently from some unknown
More informationBayesian nonparametric models for bipartite graphs
Bayesian nonparametric models for bipartite graphs François Caron INRIA IMB - University of Bordeaux Talence, France Francois.Caron@inria.fr Abstract We develop a novel Bayesian nonparametric model for
More informationInfinite Latent Feature Models and the Indian Buffet Process
Infinite Latent Feature Models and the Indian Buffet Process Thomas L. Griffiths Cognitive and Linguistic Sciences Brown University, Providence RI 292 tom griffiths@brown.edu Zoubin Ghahramani Gatsby Computational
More informationBayesian non parametric approaches: an introduction
Introduction Latent class models Latent feature models Conclusion & Perspectives Bayesian non parametric approaches: an introduction Pierre CHAINAIS Bordeaux - nov. 2012 Trajectory 1 Bayesian non parametric
More informationDirichlet Processes: Tutorial and Practical Course
Dirichlet Processes: Tutorial and Practical Course (updated) Yee Whye Teh Gatsby Computational Neuroscience Unit University College London August 2007 / MLSS Yee Whye Teh (Gatsby) DP August 2007 / MLSS
More informationImage segmentation combining Markov Random Fields and Dirichlet Processes
Image segmentation combining Markov Random Fields and Dirichlet Processes Jessica SODJO IMS, Groupe Signal Image, Talence Encadrants : A. Giremus, J.-F. Giovannelli, F. Caron, N. Dobigeon Jessica SODJO
More informationWXML Final Report: Chinese Restaurant Process
WXML Final Report: Chinese Restaurant Process Dr. Noah Forman, Gerandy Brito, Alex Forney, Yiruey Chou, Chengning Li Spring 2017 1 Introduction The Chinese Restaurant Process (CRP) generates random partitions
More informationThe Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.
Christian P. Robert The Bayesian Choice From Decision-Theoretic Foundations to Computational Implementation Second Edition With 23 Illustrations ^Springer" Contents Preface to the Second Edition Preface
More informationHierarchical Dirichlet Processes
Hierarchical Dirichlet Processes Yee Whye Teh, Michael I. Jordan, Matthew J. Beal and David M. Blei Computer Science Div., Dept. of Statistics Dept. of Computer Science University of California at Berkeley
More informationChapter 3. Dirichlet Process
Chapter 3 Dirichlet Process 3.1. The Dirichlet Process Prior 3.1.1. Definition The Dirichlet process (DP) is arguably the most popular BNP model for random probability measures (RPM), and plays a central
More informationMAD-Bayes: MAP-based Asymptotic Derivations from Bayes
MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick Brian Kulis Michael I. Jordan Cat Clusters Mouse clusters Dog 1 Cat Clusters Dog Mouse Lizard Sheep Picture 1 Picture 2 Picture 3
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationBayesian nonparametric model for sparse dynamic networks
Bayesian nonparametric model for sparse dynamic networks Konstantina Palla, François Caron and Yee Whye Teh 23rd of February, 2018 University of Glasgow Konstantina Palla 1 / 38 BIG PICTURE Interested
More informationHierarchical Species Sampling Models
Hierarchical Species Sampling Models Federico Bassetti a Roberto Casarin b Luca Rossini c a Polytechnic University of Milan, Italy b Ca Foscari University of Venice, Italy c Free University of Bozen-Bolzano,
More informationSharing Clusters Among Related Groups: Hierarchical Dirichlet Processes
Sharing Clusters Among Related Groups: Hierarchical Dirichlet Processes Yee Whye Teh (1), Michael I. Jordan (1,2), Matthew J. Beal (3) and David M. Blei (1) (1) Computer Science Div., (2) Dept. of Statistics
More informationApplied Nonparametric Bayes
Applied Nonparametric Bayes Michael I. Jordan Department of Electrical Engineering and Computer Science Department of Statistics University of California, Berkeley http://www.cs.berkeley.edu/ jordan Acknowledgments:
More information