Online Algorithms for Sum-Product
|
|
- Margaret McCarthy
- 5 years ago
- Views:
Transcription
1 Online Algorithms for Sum-Product Networks with Continuous Variables Priyank Jaini Ph.D. Seminar
2 Consistent/Robust Tensor Decomposition And Spectral Learning Offline Bayesian Learning ADF, EP, SGD, oem Non convergence and local optima Mixture Models Can be distributed; Practical problems Bayesian Moment Matching algorithm SPNs Bayesian Moment Matching for continuous SPNs Comparison with other deep networks Online
3 Streaming Data Activity Recognition Recommendation
4 Streaming Data Activity Recognition Recommendation Challenge : update model after each observation
5 Algorithm s characteristics
6 Algorithm s characteristics Online x t
7 Algorithm s characteristics Online x t Distributed x a:d x a:b x c:d
8 Algorithm s characteristics Online x t Distributed x a:d x a:b x c:d Consistent x 1:t Θ*
9 How can we learn mixture models robustly from streaming data?
10 Learning Algorithms
11 Learning Algorithms Robust : Tensor Decomposition(Anandkumar et.al, 2014),Spectral Learning(Hsu et al, 2012, Parikar and Xing, 2011); offline
12 Learning Algorithms Robust : Tensor Decomposition(Anandkumar et.al, 2014),Spectral Learning(Hsu et al, 2012, Parikar and Xing, 2011); offline Online : Assumed Density Filtering (Maybeck 1982; Lauritzen 1992; Opper & Winther 1999); not robust
13 Learning Algorithms Robust : Tensor Decomposition(Anandkumar et.al, 2014),Spectral Learning(Hsu et al, 2012, Parikar and Xing, 2011); offline Online : Assumed Density Filtering (Maybeck 1982; Lauritzen 1992; Opper & Winther 1999); not robust Expectation Propagation (Minka 2001); does not converge
14 Learning Algorithms Robust : Tensor Decomposition(Anandkumar et.al, 2014),Spectral Learning(Hsu et al, 2012, Parikar and Xing, 2011); offline Online : Assumed Density Filtering (Maybeck 1982; Lauritzen 1992; Opper & Winther 1999); not robust Expectation Propagation (Minka 2001); does not converge Stochastic Gradient Descent (Zhang 2004) online Expectation Maximization (Cappe 2012) SGD and oem : local optimum and cannot be distributed
15 Learning Algorithms Exact Bayesian Learning : Dirichlet Mixtures(Ghosal et al 1999), Gaussian Mixtures(Lijoi et al, 2005), Non-parametric Problems (Barron et al, 1999), (Freedman, 1999)
16 Learning Algorithms Exact Bayesian Learning : Dirichlet Mixtures(Ghosal et al 1999), Gaussian Mixtures(Lijoi et al, 2005), Non-parametric Problems (Barron et al, 1999), (Freedman, 1999) Exact Bayesian Learning Online
17 Learning Algorithms Exact Bayesian Learning : Dirichlet Mixtures(Ghosal et al 1999), Gaussian Mixtures(Lijoi et al, 2005), Non-parametric Problems (Barron et al, 1999), (Freedman, 1999) Exact Bayesian Learning Online Distributed
18 Learning Algorithms Exact Bayesian Learning : Dirichlet Mixtures(Ghosal et al 1999), Gaussian Mixtures(Lijoi et al, 2005), Non-parametric Problems (Barron et al, 1999), (Freedman, 1999) Exact Bayesian Learning Distributed Online Consistent
19 Learning Algorithms Exact Bayesian Learning : Dirichlet Mixtures(Ghosal et al 1999), Gaussian Mixtures(Lijoi et al, 2005), Non-parametric Problems (Barron et al, 1999), (Freedman, 1999) Exact Bayesian Learning Online Distributed Consistent In theory; practical problems!
20
21 Thomas Bayes (c )
22 Bayesian Learning Uses Bayes Theorem P Θ x = P(Θ)P(x Θ) P(x) Thomas Bayes (c ) Prior Belief New Information Bayes Theorem Updated Belief P(Θ) x P(Θ)P(x Θ) P Θ x P(x)
23 Bayesian Learning Mixture models
24 Bayesian Learning Mixture models Data : x 1:n where x i ~ M j=1 w j N(x i ; µ j, Σ j )
25 Bayesian Learning Mixture models Data : x 1:n where x i ~ M j=1 w j N(x i ; µ j, Σ j ) P n Θ = Pr Θ x 1:n
26 Bayesian Learning Mixture models Data : x 1:n where x i ~ M j=1 w j N(x i ; µ j, Σ j ) P n Θ = Pr Θ x 1:n P n 1 Θ Pr(x n Θ)
27 Bayesian Learning Mixture models Data : x 1:n where x i ~ M j=1 w j N(x i ; µ j, Σ j ) P n Θ = Pr Θ x 1:n P n 1 Θ Pr(x n Θ) P n 1 Θ M j=1 w j N(x i ; µ j, Σ j )
28 Bayesian Learning Mixture models Data : x 1:n where x i ~ M j=1 w j N(x i ; µ j, Σ j ) P n Θ = Pr Θ x 1:n P n 1 Θ Pr(x n Θ) P n 1 Θ M j=1 w j N(x i ; µ j, Σ j ) Intractable!!!
29 Bayesian Learning Mixture models Data : x 1:n where x i ~ M j=1 w j N(x i ; µ j, Σ j ) P n Θ = Pr Θ x 1:n P n 1 Θ Pr(x n Θ) P n 1 Θ M j=1 w j N(x i ; µ j, Σ j ) Intractable!!! Solution : Bayesian Moment Matching Algorithm
30
31 Karl Pearson (c )
32 Method of Moments Probability distributions defined by set of parameters Parameters can be estimated by a set of moments Karl Pearson (c ) X~ N(X; µ, σ 2 ) E X = µ E X µ 2 = σ 2
33 Make Bayesian Learning Great Again
34 Make Bayesian Learning Great Again Bayesian Learning And Method of Moments
35 Make Bayesian Learning Great Again Bayesian Learning And Method of Moments
36 Gaussian Mixture Models x i ~ M j=1 w j N(x i ; µ j, Σ j )
37 Gaussian Mixture Models x i ~ M j=1 w j N(x i ; µ j, Σ j ) Parameters : weights, means and precisions (inverse covariance matrices)
38 Bayesian Moment Matching for Gaussian Mixture Models
39 Bayesian Moment Matching for Gaussian Mixture Models P Θ x = P(Θ)P(x Θ) P(x) Parameters Parameters : weights, means and precisions (inverse covariance matrices)
40 Bayesian Moment Matching for Gaussian Mixture Models Prior P Θ x = P(Θ)P(x Θ) P(x) Parameters : weights, means and precisions (inverse covariance matrices) Prior : P w, µ, Λ ; product of Dirichlets and Normal-Wisharts Parameters
41 Bayesian Moment Matching for Gaussian Mixture Models Prior Likelihood Parameters : weights, means and precisions (inverse covariance matrices) P Θ x = P(Θ)P(x Θ) P(x) Parameters Prior : P w, µ, Λ ; product of Dirichlets and Normal-Wisharts Likelihood : P x ; w, µ, Λ = M j=1 w j N(x; µ j, Λ 1 j )
42 Bayesian Moment Matching Algorithm
43 Bayesian Moment Matching Algorithm Exact Bayesian Update x t Product of Dirichlets and Normal-Wisharts Mixture of Product of Dirichlets and Normal-Wisharts
44 Bayesian Moment Matching Algorithm Exact Bayesian Update x t Product of Dirichlets and Normal-Wisharts Projection Moment Matching Mixture of Product of Dirichlets and Normal-Wisharts
45 Sufficient Moments
46 Sufficient Moments Dirichlet : Dir w 1, w 2 w M ; α 1, α 2, α M
47 Sufficient Moments Dirichlet : Dir w 1, w 2 w M ; α 1, α 2, α M E w i = α i ; E w 2 α i = i (α i +1) j α j ( j α j )(1+ j α j )
48 Sufficient Moments Dirichlet : Dir w 1, w 2 w M ; α 1, α 2, α M E w i = α i ; E w 2 α i = i (α i +1) j α j ( j α j )(1+ j α j ) Normal-Wishart : NW µ, Λ ; µ 0, κ, W, v Λ ~ Wi(W, v) and µ Λ ~ N d µ 0, κλ 1
49 Sufficient Moments Dirichlet : Dir w 1, w 2 w M ; α 1, α 2, α M E w i = α i ; E w 2 α i = i (α i +1) j α j ( j α j )(1+ j α j ) Normal-Wishart : NW µ, Λ ; µ 0, κ, W, v Λ ~ Wi(W, v) and µ Λ ~ N d µ 0, κλ 1 E µ = µ 0 E (µ µ 0 )(µ µ 0 ) T = κ+1 κ(v d 1) W 1 E Λ = vw Var Λ ij = v(w ij 2 + W ii W jj )
50 Overall Algorithm Bayesian Step - Compute posterior P t Θ based on observation x t Sufficient Moments - Compute set of sufficient moments S for P t Θ Moment Matching - System of linear equations - Linear complexity in the number of components
51 Make Bayesian Learning Great Again Bayesian Moment Matching Algorithm - Uses Bayes Theorem + Method of Moments - Analytic solutions to Moment matching (unlike EP, ADF) - One pass over data
52 Make Bayesian Learning Great Again Bayesian Moment Matching Algorithm - Uses Bayes Theorem + Method of Moments - Analytic solutions to Moment matching (unlike EP, ADF) - One pass over data Bayesian Moment Matching Distributed Online Consistent?
53 Experiments Data Set Instances oem obmm Abalone Banknote Airfoil Arabic Transfusion CCPP Comp. Ac Kinematics Northridge Plastic
54 Experiments Data (Features) Instances oem obmm odmm Heterogeneity(16) 3,930, Magic (10) 19, Year MSD (91) 515, Miniboone (50) 130, Avg. Log-Likelihood Data (Features) Instances oem obmm odmm Heterogeneity(16) 3,930, Magic (10) 19, Year MSD (91) 515, Miniboone (50) 130, Running Time
55 Bayesian Moment Matching Discrete Data : Omar (2015, PhD Thesis) for Dirichlets; Rashwan, Zhao & Poupart (AISTATS 16) for SPNs; Hsu & Poupart (NIPS 16) for Topic Modelling Continuous Data : Jaini & Poupart, 2016 (arxiv); Jaini, Rashwan et al, (PGM 16) for SPNs; Poupart, Chen, Jaini et al (NetworksML 16) Sequence Data and Transfer Learning : Jaini, Poupart et al, (submitted to ICLR 17)
56 Consistent/Robust Tensor Decomposition And Spectral Learning Offline Bayesian Learning ADF, EP, SGD, oem Non convergence and local optima Mixture Models Can be distributed; Practical problems Bayesian Moment Matching algorithm Consistent? SPNs Bayesian Moment Matching for continuous SPNs Comparison with other deep networks Online
57 What is a Sum-Product Network? Proposed by Poon and Domingos (UAI 2011) equivalent to Arithmetic Circuits (Darwiche 2003) Deep architecture with clear semantics Tractable Probabilistic Graphical Models
58 What is a Sum-Product Network? P x 1, x 2 = P x 1, x 2 x 1 x 2 + P x 1, x 2 x 1 x 2 + P x 1, x 2 x 1 x 2 + P(x 1, x 2 )x 1 x 2 a x X x 1 x 1 x 2 x 2
59 What is a Sum-Product Network? P x 1, x 2 = P x 1, x 2 x 1 x 2 + P x 1, x 2 x 1 x 2 + P x 1, x 2 x 1 x 2 + P(x 1, x 2 )x 1 x 2 a x X SPNs - Directed Acyclic Graphs x 1 x 1 x 2 x 2
60 What is a Sum-Product Network? P x 1, x 2 = P x 1, x 2 x 1 x 2 + P x 1, x 2 x 1 x 2 + P x 1, x 2 x 1 x 2 + P(x 1, x 2 )x 1 x 2 + a x X x 1 x 1 x 2 x 2 SPNs - Directed Acyclic Graphs A valid SPN is - Complete : Each sum node has children with same scope
61 What is a Sum-Product Network? P x 1, x 2 = P x 1, x 2 x 1 x 2 + P x 1, x 2 x 1 x 2 + P x 1, x 2 x 1 x 2 + P(x 1, x 2 )x 1 x 2 + a x X x 1 x 1 x 2 x 2 SPNs - Directed Acyclic Graphs A valid SPN is - Complete : Each sum node has children with same scope - Decomposable : Each product node has children with disjoint scope
62 Probabilistic Inference - SPN SPN represents a joint distribution over a set of random variables Example : Query : Pr(X 1 = 1, X 2 = 0) Pr(X 1 = 1, X 2 = 0) = 34.8/constant
63 Probabilistic Inference - SPN SPN represents a joint distribution over a set of random variables Example : Query : Pr(X 1 = 1, X 2 = 0) Pr(X 1 = 1, X 2 = 0) = 34.8/ Linear Complexity two bottom passes for any query
64 Discrete Learning SPNS Parameter Learning : - Maximum Likelihood: SGD (Poon & Domingos, 2011) slow convergence, inaccurate; EM(Perharz, 2015) inaccurate; Signomial Programming (Zhao & Poupart, 2016) - Bayesian Learning: BMM(Rashwan et al., 2016) accurate; Collapsed Variational Inference (Zhao et al., 2016) accurate Online parameter learning for continuous SPNs (Jaini et al.2016) - extend obmm to Gaussian SPNS
65 Continuous SPNS Gaussian NN 11 (X (X 1 ) N 2 (X 1 ) N 3 (X 2 ) N 4 1 ) (X 2 )
66 Continuous SPNS Gaussian NN 11 (X (X 1 ) N 2 (X 1 ) N 3 (X 2 ) N 4 1 ) (X 2 ) mixture of Gaussians (Unnormalised)
67 Continuous SPNS Hierarchical mixture of Gaussians (Unnormalised) N 1 (X 1 ) N 2 (X 1 ) N 3 (X 2 ) N 4 (X 2 )
68 Bayesian Learning of SPNs + w 7 w Parameters : weights, means and precisions (inverse covariance matrices) Prior : P w, µ, Λ ; product of Dirichlets and Normal-Wisharts + + w 2 w 3 w 5 w 6 w 4 w N(X 2 ; µ 3, Λ 1 3 ) N(X 2 ; µ 4, Λ ) N(X 1 ; µ 1, Λ 1 1 )N(X 1 ; µ 2, Λ 2 1 ) + Likelihood : P x ; w, µ, Λ =SPN x ; w, µ, Λ Posterior : P w, µ, Λ; data P w, µ, Λ n SPN x n ; w, µ, Λ
69 Bayesian Moment Matching Algorithm Exact Bayesian Update x t Product of Dirichlets and Normal-Wisharts Projection Moment Matching Mixture of Product of Dirichlets and Normal-Wisharts
70 Overall Algorithm Recursive computation of all constants - Linear complexity in the size of SPN Moment Matching - System of linear equations - Linear complexity in the size of SPN Streaming Data - Posterior update : constant time w.r.t. amount of data
71 Empirical Results Dataset Flow Size Quake Banknote Abalone Kinematics CA Sensorless Drive Attributes obmm (random) oem (random) obmm (GMM) oem (GMM) Avg. log-likelihood and standard error based on 10-fold cross-validation obmm performs better than oem
72 Empirical Results Dataset Flow Size Quake Banknote Abalone Kinematics CA Sensorless Drive Attributes obmm (random) obmm (GMM) SRBM GenMMN Avg. log-likelihood and standard error based on 10-fold cross-validation ` obmm is competitive with SBRM and GenMMN
73 Consistent/Robust Tensor Decomposition And Spectral Learning Offline Bayesian Learning ADF, EP, SGD, oem Non convergence and local optima Online Mixture Models Can be distributed; Practical problems Bayesian Moment Matching algorithm Consistent? SPNs Bayesian Moment Matching for continuous SPNs Comparison with other deep networks obmm performs better than oem and comparable to other techniques
74 Conclusion and Future Work Contributions : - Online and distributed Bayesian Moment Matching algorithm - Performs better than oem w.r.t time and accuracy - Extended it to continuous SPNs - Comparative analysis with other deep learning methods Future Work : - Theoretical properties of BMM consistent? - Generalize obmm to exponential family - Extension to sequential data and transfer learning - Online structure learning of SPNs
75 Thank you!
Online Bayesian Transfer Learning for Sequential Data Modeling
Online Bayesian Transfer Learning for Sequential Data Modeling....? Priyank Jaini Machine Learning, Algorithms and Theory Lab Network for Aging Research 2 3 Data of personal preferences (years) Data (non-existent)
More informationSum-Product Networks. STAT946 Deep Learning Guest Lecture by Pascal Poupart University of Waterloo October 17, 2017
Sum-Product Networks STAT946 Deep Learning Guest Lecture by Pascal Poupart University of Waterloo October 17, 2017 Introduction Outline What is a Sum-Product Network? Inference Applications In more depth
More informationOnline Algorithms for Sum-Product Networks with Continuous Variables
ONLINE ALGORITHMS FOR SUM-PRODUCT NETWORKS WITH CONTINUOUS VARIABLES Online Algorithms for Sum-Product Networks with Continuous Variables Priyank Jaini 1 Abdullah Rashwan 1 Han Zhao 2 Yue Liu 1 Ershad
More informationOnline and Distributed Bayesian Moment Matching for Parameter Learning in Sum-Product Networks
Online and Distributed Bayesian Moment Matching for Parameter Learning in Sum-Product Networks Abdullah Rashwan Han Zhao Pascal Poupart Computer Science University of Waterloo arashwan@uwaterloo.ca Machine
More informationPart 1: Expectation Propagation
Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud
More informationSum-Product Networks: A New Deep Architecture
Sum-Product Networks: A New Deep Architecture Pedro Domingos Dept. Computer Science & Eng. University of Washington Joint work with Hoifung Poon 1 Graphical Models: Challenges Bayesian Network Markov Network
More informationExpectation Propagation Algorithm
Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,
More informationDiscriminative Learning of Sum-Product Networks. Robert Gens Pedro Domingos
Discriminative Learning of Sum-Product Networks Robert Gens Pedro Domingos X1 X1 X1 X1 X2 X2 X2 X2 X3 X3 X3 X3 X4 X4 X4 X4 X5 X5 X5 X5 X6 X6 X6 X6 Distributions X 1 X 1 X 1 X 1 X 2 X 2 X 2 X 2 X 3 X 3
More informationParametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012
Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood
More informationOn the Relationship between Sum-Product Networks and Bayesian Networks
On the Relationship between Sum-Product Networks and Bayesian Networks International Conference on Machine Learning, 2015 Han Zhao Mazen Melibari Pascal Poupart University of Waterloo, Waterloo, ON, Canada
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate
More informationCollapsed Variational Inference for Sum-Product Networks
for Sum-Product Networks Han Zhao 1, Tameem Adel 2, Geoff Gordon 1, Brandon Amos 1 Presented by: Han Zhao Carnegie Mellon University 1, University of Amsterdam 2 June. 20th, 2016 1 / 26 Outline Background
More informationLinear Time Computation of Moments in Sum-Product Networks
Linear Time Computation of Moments in Sum-Product Netorks Han Zhao Machine Learning Department Carnegie Mellon University Pittsburgh, PA 15213 han.zhao@cs.cmu.edu Geoff Gordon Machine Learning Department
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationGraphical Models for Collaborative Filtering
Graphical Models for Collaborative Filtering Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Sequence modeling HMM, Kalman Filter, etc.: Similarity: the same graphical model topology,
More informationExpectation propagation as a way of life
Expectation propagation as a way of life Yingzhen Li Department of Engineering Feb. 2014 Yingzhen Li (Department of Engineering) Expectation propagation as a way of life Feb. 2014 1 / 9 Reference This
More informationp L yi z n m x N n xi
y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen
More informationStreaming Variational Bayes
Streaming Variational Bayes Tamara Broderick, Nicholas Boyd, Andre Wibisono, Ashia C. Wilson, Michael I. Jordan UC Berkeley Discussion led by Miao Liu September 13, 2013 Introduction The SDA-Bayes Framework
More informationLecture 6: Graphical Models: Learning
Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationPattern Recognition and Machine Learning. Bishop Chapter 9: Mixture Models and EM
Pattern Recognition and Machine Learning Chapter 9: Mixture Models and EM Thomas Mensink Jakob Verbeek October 11, 27 Le Menu 9.1 K-means clustering Getting the idea with a simple example 9.2 Mixtures
More informationMachine Learning Summer School
Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,
More informationData Mining Techniques
Data Mining Techniques CS 622 - Section 2 - Spring 27 Pre-final Review Jan-Willem van de Meent Feedback Feedback https://goo.gl/er7eo8 (also posted on Piazza) Also, please fill out your TRACE evaluations!
More informationA graph contains a set of nodes (vertices) connected by links (edges or arcs)
BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should
More informationVariational Inference. Sargur Srihari
Variational Inference Sargur srihari@cedar.buffalo.edu 1 Plan of discussion We first describe inference with PGMs and the intractability of exact inference Then give a taxonomy of inference algorithms
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 3 Stochastic Gradients, Bayesian Inference, and Occam s Razor https://people.orie.cornell.edu/andrew/orie6741 Cornell University August
More informationPMR Learning as Inference
Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning
More informationAlgorithms for Variational Learning of Mixture of Gaussians
Algorithms for Variational Learning of Mixture of Gaussians Instructors: Tapani Raiko and Antti Honkela Bayes Group Adaptive Informatics Research Center 28.08.2008 Variational Bayesian Inference Mixture
More information13: Variational inference II
10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational
More informationBayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework
HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for
More informationDistributed Estimation, Information Loss and Exponential Families. Qiang Liu Department of Computer Science Dartmouth College
Distributed Estimation, Information Loss and Exponential Families Qiang Liu Department of Computer Science Dartmouth College Statistical Learning / Estimation Learning generative models from data Topic
More informationVariational Message Passing. By John Winn, Christopher M. Bishop Presented by Andy Miller
Variational Message Passing By John Winn, Christopher M. Bishop Presented by Andy Miller Overview Background Variational Inference Conjugate-Exponential Models Variational Message Passing Messages Univariate
More informationVCMC: Variational Consensus Monte Carlo
VCMC: Variational Consensus Monte Carlo Maxim Rabinovich, Elaine Angelino, Michael I. Jordan Berkeley Vision and Learning Center September 22, 2015 probabilistic models! sky fog bridge water grass object
More informationLinear Dynamical Systems
Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations
More informationLecture 13 : Variational Inference: Mean Field Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2017 Lecture 13 : Variational Inference: Mean Field Approximation Lecturer: Willie Neiswanger Scribes: Xupeng Tong, Minxing Liu 1 Problem Setup 1.1
More informationExpectation Propagation for Approximate Bayesian Inference
Expectation Propagation for Approximate Bayesian Inference José Miguel Hernández Lobato Universidad Autónoma de Madrid, Computer Science Department February 5, 2007 1/ 24 Bayesian Inference Inference Given
More informationProbabilistic and Bayesian Machine Learning
Probabilistic and Bayesian Machine Learning Day 4: Expectation and Belief Propagation Yee Whye Teh ywteh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/
More informationMCMC and Gibbs Sampling. Kayhan Batmanghelich
MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction
More informationVariational Mixture of Gaussians. Sargur Srihari
Variational Mixture of Gaussians Sargur srihari@cedar.buffalo.edu 1 Objective Apply variational inference machinery to Gaussian Mixture Models Demonstrates how Bayesian treatment elegantly resolves difficulties
More informationBayesian Deep Learning
Bayesian Deep Learning Mohammad Emtiyaz Khan AIP (RIKEN), Tokyo http://emtiyaz.github.io emtiyaz.khan@riken.jp June 06, 2018 Mohammad Emtiyaz Khan 2018 1 What will you learn? Why is Bayesian inference
More informationCSci 8980: Advanced Topics in Graphical Models Gaussian Processes
CSci 8980: Advanced Topics in Graphical Models Gaussian Processes Instructor: Arindam Banerjee November 15, 2007 Gaussian Processes Outline Gaussian Processes Outline Parametric Bayesian Regression Gaussian
More informationNon-Parametric Bayes
Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian
More informationUnderstanding Covariance Estimates in Expectation Propagation
Understanding Covariance Estimates in Expectation Propagation William Stephenson Department of EECS Massachusetts Institute of Technology Cambridge, MA 019 wtstephe@csail.mit.edu Tamara Broderick Department
More informationProbabilistic Graphical Networks: Definitions and Basic Results
This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationBayesian Deep Learning
Bayesian Deep Learning Mohammad Emtiyaz Khan RIKEN Center for AI Project, Tokyo http://emtiyaz.github.io Keywords Bayesian Statistics Gaussian distribution, Bayes rule Continues Optimization Gradient descent,
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More informationLearning Bayesian network : Given structure and completely observed data
Learning Bayesian network : Given structure and completely observed data Probabilistic Graphical Models Sharif University of Technology Spring 2017 Soleymani Learning problem Target: true distribution
More informationCSC 412 (Lecture 4): Undirected Graphical Models
CSC 412 (Lecture 4): Undirected Graphical Models Raquel Urtasun University of Toronto Feb 2, 2016 R Urtasun (UofT) CSC 412 Feb 2, 2016 1 / 37 Today Undirected Graphical Models: Semantics of the graph:
More informationProbabilistic Graphical Models & Applications
Probabilistic Graphical Models & Applications Learning of Graphical Models Bjoern Andres and Bernt Schiele Max Planck Institute for Informatics The slides of today s lecture are authored by and shown with
More informationVariational Scoring of Graphical Model Structures
Variational Scoring of Graphical Model Structures Matthew J. Beal Work with Zoubin Ghahramani & Carl Rasmussen, Toronto. 15th September 2003 Overview Bayesian model selection Approximations using Variational
More informationBayesian Approaches Data Mining Selected Technique
Bayesian Approaches Data Mining Selected Technique Henry Xiao xiao@cs.queensu.ca School of Computing Queen s University Henry Xiao CISC 873 Data Mining p. 1/17 Probabilistic Bases Review the fundamentals
More informationECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning
ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning Topics Summary of Class Advanced Topics Dhruv Batra Virginia Tech HW1 Grades Mean: 28.5/38 ~= 74.9%
More informationDeep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści
Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, 2017 Spis treści Website Acknowledgments Notation xiii xv xix 1 Introduction 1 1.1 Who Should Read This Book?
More informationLearning Gaussian Graphical Models with Unknown Group Sparsity
Learning Gaussian Graphical Models with Unknown Group Sparsity Kevin Murphy Ben Marlin Depts. of Statistics & Computer Science Univ. British Columbia Canada Connections Graphical models Density estimation
More informationTractable Learning in Structured Probability Spaces
Tractable Learning in Structured Probability Spaces Guy Van den Broeck UCLA Stats Seminar Jan 17, 2017 Outline 1. Structured probability spaces? 2. Specification language Logic 3. Deep architecture Logic
More informationCh 4. Linear Models for Classification
Ch 4. Linear Models for Classification Pattern Recognition and Machine Learning, C. M. Bishop, 2006. Department of Computer Science and Engineering Pohang University of Science and echnology 77 Cheongam-ro,
More informationPILCO: A Model-Based and Data-Efficient Approach to Policy Search
PILCO: A Model-Based and Data-Efficient Approach to Policy Search (M.P. Deisenroth and C.E. Rasmussen) CSC2541 November 4, 2016 PILCO Graphical Model PILCO Probabilistic Inference for Learning COntrol
More informationBasic math for biology
Basic math for biology Lei Li Florida State University, Feb 6, 2002 The EM algorithm: setup Parametric models: {P θ }. Data: full data (Y, X); partial data Y. Missing data: X. Likelihood and maximum likelihood
More information13 : Variational Inference: Loopy Belief Propagation and Mean Field
10-708: Probabilistic Graphical Models 10-708, Spring 2012 13 : Variational Inference: Loopy Belief Propagation and Mean Field Lecturer: Eric P. Xing Scribes: Peter Schulam and William Wang 1 Introduction
More informationNote Set 5: Hidden Markov Models
Note Set 5: Hidden Markov Models Probabilistic Learning: Theory and Algorithms, CS 274A, Winter 2016 1 Hidden Markov Models (HMMs) 1.1 Introduction Consider observed data vectors x t that are d-dimensional
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft
More informationECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4
ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian
More informationLecture 21: Spectral Learning for Graphical Models
10-708: Probabilistic Graphical Models 10-708, Spring 2016 Lecture 21: Spectral Learning for Graphical Models Lecturer: Eric P. Xing Scribes: Maruan Al-Shedivat, Wei-Cheng Chang, Frederick Liu 1 Motivation
More informationExpectation Propagation in Factor Graphs: A Tutorial
DRAFT: Version 0.1, 28 October 2005. Do not distribute. Expectation Propagation in Factor Graphs: A Tutorial Charles Sutton October 28, 2005 Abstract Expectation propagation is an important variational
More informationScaling Neighbourhood Methods
Quick Recap Scaling Neighbourhood Methods Collaborative Filtering m = #items n = #users Complexity : m * m * n Comparative Scale of Signals ~50 M users ~25 M items Explicit Ratings ~ O(1M) (1 per billion)
More informationIntroduction to Gaussian Processes
Introduction to Gaussian Processes Iain Murray murray@cs.toronto.edu CSC255, Introduction to Machine Learning, Fall 28 Dept. Computer Science, University of Toronto The problem Learn scalar function of
More informationFast yet Simple Natural-Gradient Variational Inference in Complex Models
Fast yet Simple Natural-Gradient Variational Inference in Complex Models Mohammad Emtiyaz Khan RIKEN Center for AI Project, Tokyo, Japan http://emtiyaz.github.io Joint work with Wu Lin (UBC) DidrikNielsen,
More informationProbabilistic and Logistic Circuits: A New Synthesis of Logic and Machine Learning
Probabilistic and Logistic Circuits: A New Synthesis of Logic and Machine Learning Guy Van den Broeck KULeuven Symposium Dec 12, 2018 Outline Learning Adding knowledge to deep learning Logistic circuits
More informationTractable Learning in Structured Probability Spaces
Tractable Learning in Structured Probability Spaces Guy Van den Broeck DTAI Seminar - KU Leuven Dec 20, 2016 Structured probability spaces? Running Example Courses: Logic (L) Knowledge Representation (K)
More informationLightweight Probabilistic Deep Networks Supplemental Material
Lightweight Probabilistic Deep Networks Supplemental Material Jochen Gast Stefan Roth Department of Computer Science, TU Darmstadt In this supplemental material we derive the recipe to create uncertainty
More informationModel Selection for Gaussian Processes
Institute for Adaptive and Neural Computation School of Informatics,, UK December 26 Outline GP basics Model selection: covariance functions and parameterizations Criteria for model selection Marginal
More informationProbabilistic Time Series Classification
Probabilistic Time Series Classification Y. Cem Sübakan Boğaziçi University 25.06.2013 Y. Cem Sübakan (Boğaziçi University) M.Sc. Thesis Defense 25.06.2013 1 / 54 Problem Statement The goal is to assign
More informationAPPROXIMATION COMPLEXITY OF MAP INFERENCE IN SUM-PRODUCT NETWORKS
APPROXIMATION COMPLEXITY OF MAP INFERENCE IN SUM-PRODUCT NETWORKS Diarmaid Conaty Queen s University Belfast, UK Denis D. Mauá Universidade de São Paulo, Brazil Cassio P. de Campos Queen s University Belfast,
More informationChris Bishop s PRML Ch. 8: Graphical Models
Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular
More informationLearning MN Parameters with Alternative Objective Functions. Sargur Srihari
Learning MN Parameters with Alternative Objective Functions Sargur srihari@cedar.buffalo.edu 1 Topics Max Likelihood & Contrastive Objectives Contrastive Objective Learning Methods Pseudo-likelihood Gradient
More informationMixtures of Gaussians. Sargur Srihari
Mixtures of Gaussians Sargur srihari@cedar.buffalo.edu 1 9. Mixture Models and EM 0. Mixture Models Overview 1. K-Means Clustering 2. Mixtures of Gaussians 3. An Alternative View of EM 4. The EM Algorithm
More information1 Bayesian Linear Regression (BLR)
Statistical Techniques in Robotics (STR, S15) Lecture#10 (Wednesday, February 11) Lecturer: Byron Boots Gaussian Properties, Bayesian Linear Regression 1 Bayesian Linear Regression (BLR) In linear regression,
More informationIntroduction to Probabilistic Graphical Models: Exercises
Introduction to Probabilistic Graphical Models: Exercises Cédric Archambeau Xerox Research Centre Europe cedric.archambeau@xrce.xerox.com Pascal Bootcamp Marseille, France, July 2010 Exercise 1: basics
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 2: Bayesian Basics https://people.orie.cornell.edu/andrew/orie6741 Cornell University August 25, 2016 1 / 17 Canonical Machine Learning
More informationOn the Relationship between Sum-Product Networks and Bayesian Networks
On the Relationship between Sum-Product Networks and Bayesian Networks by Han Zhao A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of
More informationBased on slides by Richard Zemel
CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we
More informationClassification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012
Classification CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Topics Discriminant functions Logistic regression Perceptron Generative models Generative vs. discriminative
More informationAn Introduction to Bayesian Machine Learning
1 An Introduction to Bayesian Machine Learning José Miguel Hernández-Lobato Department of Engineering, Cambridge University April 8, 2013 2 What is Machine Learning? The design of computational systems
More informationVariational Learning : From exponential families to multilinear systems
Variational Learning : From exponential families to multilinear systems Ananth Ranganathan th February 005 Abstract This note aims to give a general overview of variational inference on graphical models.
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 12 Dynamical Models CS/CNS/EE 155 Andreas Krause Homework 3 out tonight Start early!! Announcements Project milestones due today Please email to TAs 2 Parameter learning
More informationEstimating Latent Variable Graphical Models with Moments and Likelihoods
Estimating Latent Variable Graphical Models with Moments and Likelihoods Arun Tejasvi Chaganty Percy Liang Stanford University June 18, 2014 Chaganty, Liang (Stanford University) Moments and Likelihoods
More informationNeutron inverse kinetics via Gaussian Processes
Neutron inverse kinetics via Gaussian Processes P. Picca Politecnico di Torino, Torino, Italy R. Furfaro University of Arizona, Tucson, Arizona Outline Introduction Review of inverse kinetics techniques
More informationEM & Variational Bayes
EM & Variational Bayes Hanxiao Liu September 9, 2014 1 / 19 Outline 1. EM Algorithm 1.1 Introduction 1.2 Example: Mixture of vmfs 2. Variational Bayes 2.1 Introduction 2.2 Example: Bayesian Mixture of
More informationIntroduction to Bayesian inference
Introduction to Bayesian inference Thomas Alexander Brouwer University of Cambridge tab43@cam.ac.uk 17 November 2015 Probabilistic models Describe how data was generated using probability distributions
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationIntroduction to Gaussian Process
Introduction to Gaussian Process CS 778 Chris Tensmeyer CS 478 INTRODUCTION 1 What Topic? Machine Learning Regression Bayesian ML Bayesian Regression Bayesian Non-parametric Gaussian Process (GP) GP Regression
More informationCS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas
CS839: Probabilistic Graphical Models Lecture 7: Learning Fully Observed BNs Theo Rekatsinas 1 Exponential family: a basic building block For a numeric random variable X p(x ) =h(x)exp T T (x) A( ) = 1
More informationProbabilistic Latent Semantic Analysis
Probabilistic Latent Semantic Analysis Yuriy Sverchkov Intelligent Systems Program University of Pittsburgh October 6, 2011 Outline Latent Semantic Analysis (LSA) A quick review Probabilistic LSA (plsa)
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS Parametric Distributions Basic building blocks: Need to determine given Representation: or? Recall Curve Fitting Binary Variables
More information