Chapter 04: Exact Inference in Bayesian Networks

Size: px
Start display at page:

Download "Chapter 04: Exact Inference in Bayesian Networks"

Transcription

1 LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 04: Exact Inference in Bayesian Networks Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute of Technology Institute of Measurement and Control Systems Learning and Inference in Graphical Models. Chapter 04 p. /23

2 References for this chapter Christopher M. Bishop, Pattern Recognition and Machine Learning, ch. 8, Springer, 2006 Stuart Russell and Peter Norvig, Artificial Intelligenece: A Modern Approach, ch. 4, Prentice Hall, 2003 Judea Pearl, Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference, Morgan Kaufmann, 989 Steffen L. Lauritzen and David J. Spiegelhalter, Local computations with probabilities on graphical structures and their applications to expert systems, In: The Journal of the Royal Statistical Society, vol. 50, no. 2, pp , 988 Brendan J. Frey and David J. C. MacKay, A Revolution: Belief Propagation in Graphs with Cycles, In: Advances in Neural Information Processing Systems (NIPS), vol. 0, pp , 997, http: //books.nips.cc/papers/files/nips0/0479.pdf Learning and Inference in Graphical Models. Chapter 04 p. 2/23

3 Inference Given a graphical model with unobserved nodes U and observed nodeso, we want to conclude about the distribution of unobserved nodes calculate the single node marginal distribution p(x O) withx U calculate the maximum-a-posterior estimator arg max u p(u = u O = o) Exact inference is not always possible. We focus on the easier cases, polytrees example of a polytree a polytree is a directed acyclic graph whose underlying undirected graph is acyclic. counterexample of a polytree Learning and Inference in Graphical Models. Chapter 04 p. 3/23

4 Marginalization Calculatep(X = x O = o,o 2 = o 2 ) for the network on the right. p (X = x O = o,o 2 = o 2 ) with = p(x = x,o = o,o 2 = o 2 ) p(o = o,o 2 = o 2 ) p(x = x,o = o,o 2 = o 2 ) = p(x = x,o = o,o 2 = o 2 )dx p (X = x,o = o,o 2 = o 2 ) = p(x=x,o =o,o 2 =o 2,U =u,..., U U 2 U 3 X U 4 U 5 O continue on blackboard U 7 =u 7 ) du du 2 du 3 du 4 du 5 du 6 du 7 U 6 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 4/23

5 Factor graph A factor graph is a bipartite graph with two kind of nodes: variable nodes that model random variables like in a Bayesian network factor nodes that model a probabilistic relationship between variable nodes. Each factor node is assigned with a factor, i.e. a function that models the stochastic relationship Variable nodes and factor nodes are connected by undirected links. For each Bayesian polytree we can create a factor graph as follows: the set of variable nodes is taken from the nodes of the Bayesian polytree for each factorp(x Pred(X) in the Bayesian network we create a new factor nodef we connectx andpred(x) withf we assignf(x,y,...,y n ) p(x = x Pred(X) = (y,...,y n )) Hence, the joint probability of the Bayesian polytree is equal to the product of all factors of the factor tree. Learning and Inference in Graphical Models. Chapter 04 p. 5/23

6 Factor graphs Example: factor graph for the Bayesian network on the right f (u )=p(u = u ) f 2 (u,u 2 )=p(u 2 = u 2 U = u ) f 3 (u 3 )=p(u 3 = u 3 ) f 4 (u 2,u 3,x)=p(X = x U 2 = u 2,U 3 = u 3 ) f 5 (u 4 )=p(u 4 = u 4 ) f 6 (x,u 4,u 5 )=p(u 5 = u 5 X = x,u 4 = u 4 ) f 7 (u 4,o )=p(o = o U 4 = u 4 ) f 8 (u 5,u 6 )=p(u 6 = u 6 U 5 = u 5 ) f 9 (u 5,o 2 )=p(o 2 = o 2 U 5 = u 5 ) f 0 (o,u 7 )=p(u 7 = u 7 O = o ) f U f 2 f 3 U 2 U 3 f 4 f 5 X U 4 f 6 f 7 U 5 O f 8 f 9 f 0 U 6 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 6/23

7 Marginalization on factor graphs Task: calculate I(x) = f (u )f 2 (u,u 2 )f 3 (u 3 ) f 4 (u 2,u 3,x)f 5 (u 4 )f 6 (x,u 4,u 5 ) f 7 (u 4,o )f 8 (u 5,u 6 )f 9 (u 5,o 2 )f 0 (o,u 7 ) du du 2 du 3 du 4 du 5 du 6 du 7 Observations: factor graph is tree with rootx I(x) can be split into two large factors m f4 X(x)= f (u )f 2 (u,u 2 )f 3 (u 3 ) f 4 (u 2,u 3,x)du du 2 du 3 m f6 X(x)= f 5 (u 4 )f 6 (x,u 4,u 5 ) f U f 2 f 3 U 2 U 3 f 4 f 5 X U 4 f 6 f 7 U 5 O f 8 f 9 f 0 U 6 O 2 U 7 f 7 (u 4,o )f 8 (u 5,u 6 )f 9 (u 5,o 2 )f 0 (o,u 7 )du 4 du 5 du 6 du 7 Learning and Inference in Graphical Models. Chapter 04 p. 7/23

8 Marginalization on factor graphs Inm f4 x we can factor outf 3 andf 4 ( m f4 X(x)= ) f (u )f 2 (u,u 2 )du } {{ } =:m f2 U 2 (u 2 ) and rewritem f2 U 2 (u 2 ) as m f2 U 2 (u 2 )= f (u ) f }{{} 2 (u,u 2 )du =:m f U (u ) Observations: calculation can be split along the branches of the tree the lead nodes can serve as starting points for the calculation f 3 (u 3 ) f }{{} 4 (u 2,u 3,x)du 2 du 3 =:m f3 U 3 (u 3 ) f U f 2 f 3 U 2 U 3 f 4 only multiplication and integration/summation occur intermediate results can be interpreted as messages sent from one node to its neighbors X Learning and Inference in Graphical Models. Chapter 04 p. 8/23

9 Marginalization on factor graphs m f6 X(x) can be split in a similar manner m f6 X(x)= f 6 (x,u 4,u 5 )m U4 f 6 (u 4 )m U5 f 6 (u 5 )du 4 du 5 m U4 f 6 (u 4 )=m f5 U 4 (u 4 ) m f7 U 4 (u 4 ) m U5 f 6 (u 4 )=m f8 U 5 (u 5 ) m f9 U 5 (u 5 ) m f8 U 5 (u 5 )= f 8 (u 5,u 6 )m U6 f 8 (u 6 )du 6 m U6 f 8 (u 6 )= m f9 U 5 (u 5 )=f 9 (u 5,o 2 ) with observedo 2 If we want to extend the procedure to observed nodes, we could also argue m f9 U 5 (u 5 )= f 9 (u 5,o 2)m O2 f 9 (o 2)do 2 f 5 X U 4 f 6 f 7 U 5 O f 8 f 9 f 0 U 6 O 2 U 7 m O2 f 9 (o 2) = δ(o 2 o 2 ) withδ is the Dirac distribution Learning and Inference in Graphical Models. Chapter 04 p. 9/23

10 Side topic: Dirac distribution The Diracδ is a distribution that can be used to model discrete distributions in continuous space: δ(x) = so that Examples { 0 ifx 0 ifx = 0 δ(x)dx = ifx is distributed w.r.t. the Dirac distribution,x can only take the value0 ify is distributed w.r.t. 0.3 δ(y 2)+0.7 δ(y +5.),Y will take the value2with probability of0.3 and 5. with probability of0.7 Learning and Inference in Graphical Models. Chapter 04 p. 0/23

11 Belief propagation Example motivates a generic algorithm known as sum-product algorithm or belief propagation (Pearl, 989; Lauritzen and Spiegelhalter, 988) factor nodesf generate messages and send them to variable nodesv: m f V (v) variable nodesv generate messages and send them to factor nodesf : m V f (v) messages are like distributions (but not necessarily normalized) a message from a nodento a neighboring noden can be generated as soon asnreceived messages from all its neighbors exceptn hence, the method can start at the leaf nodes and follow the branches of the tree until the node of interest is met (dynamic programing principle) Learning and Inference in Graphical Models. Chapter 04 p. /23

12 Belief propagation How are the messages created? messages from unobserved variable nodes to factor nodes: m X f (x) = m fi X(x) i iff,f,...,f n are the neighbors ofx messages from observed variable nodes to factor nodes: m X f (x) = δ(x x ) m fi X(x) i iff,f,...,f n are the neighbors ofx andx is the observed value atx messages from factor nodes to variable nodes: m f X (x) = ifx,y,...,y n are the neighbors off f(x,y,...,y n ) m Yi f(y i )dy...dy n i Learning and Inference in Graphical Models. Chapter 04 p. 2/23

13 Belief propagation Example ( exam problem ) c P(c=c) m + P(M=m) 3 4 P (E=e C=c,M=m) = ife=c+m ife=c ife=c m 0 otherwise level of comprehension mental mood result of exam apply belief propagation to calculate P(C E) for E = 2 blackboard Learning and Inference in Graphical Models. Chapter 04 p. 3/23

14 Belief propagation Belief propagation works if either or all distributions are categorical (e.g. example) all distributions are conjugate all distributions are Gaussian and variables depend linearly, i.e. X Y,...,Y n N(c Y + +c n Y n,σ 2 ) with fixed values c,...,c n,σ 2 Otherwise, integrals might become untreatable analytically Gauss-linear example: exam problem with C N(3,4) M N(0,) E C,M N(C +M,) blackboard/homework Learning and Inference in Graphical Models. Chapter 04 p. 4/23

15 MAP estimator Second task, calculate the maximum-a-posterior estimator arg maxp(u = u O = o) u Again, we focus on polytrees only. Calculate the MAP estimator for the model on the right. arg max u = argmax u = argmax u = argmax u p(u = u O = o) p(u = u,o = o) p(o = o) p(u = u,o = o) logp(u = u,o = o) U U 2 U 3 U 8 U 4 U 5 O U 6 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 5/23

16 MAP estimator logp(u = u,o = o) = log ( f(u i,pred(u i )) i f(o j,pred(o j )) ) = i j logf(u i,pred(u i ))+ logf(o j,pred(o j )) j choose one node as root node (e.g. U 8 ): max logp(u = u,o = o) u,...,u 8 ( = max mf4 U 8 (u 8 )+m f6 U 8 (u 8 ) ) u 8 wherem f4 U 8 contains all terms reated tof,...,f 4 andm f6 U 8 contains all terms related tof 5,...,f 0 f U f 2 f 3 U 2 U 3 f 4 f 5 U 8 U 4 f 6 f 7 U 6 U 5 O f 8 f 9 f 0 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 6/23

17 MAP estimator m f6 U 8 (u 8 )=max u 4,u 5 ( logf6 (u 4,u 5,u 8 )+m U4 f 6 (u 4 )+m U5 f 6 (u 5 ) ) m U5 f 6 (u 5 )=m f8 U 5 (u 5 )+m f9 U 5 (u 5 ) ( m f8 U 5 (u 5 )=max logf8 (u 5,u 6 )+m U6 f 8 (u 6 ) ) u 6 m U6 f 8 (u 6 )=0 m f9 U 5 (u 5 )=logf 9 (u 5,o 2 ) ( =max logf9 (u 5,o 2)+m O2 o f 9 (o 2) ) 2 { m O2 f 9 (o 0 ifo 2 = o 2 2)= otherwise f 5 U 8 U 4 f 6 f 7 U 5 O f 8 f 9 f 0 U 6 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 7/23

18 MAP estimator m U4 f 6 (u 4 )=m f5 U 4 (u 4 )+m f7 U 4 (u 4 ) m f5 U 4 (u 4 )=logf 4 (u 4 ) m f7 U 4 (u 4 )=logf 7 (u 4,o )+something ( =max logf7 (u 4,o )+m O o f 7 (o ) ) m O f 7 (o )=m f0 O (o )+ { 0 ifo = o otherwise m f0 O (o )=max u 7 ( logf0 (o,u 7 )+m U7 f 0 (u 7 ) ) m U7 f 0 (u 7 )=0 f 5 U 8 U 4 f 6 f 7 U 5 O f 8 f 9 f 0 U 6 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 8/23

19 Max-sum algorithm Example motivates a generic algorithm to calculatemax u logp(u = u,o = o) known as max-sum algorithm factor nodesf generate messages and send them to variable nodesv: m f V (v) variable nodesv generate messages and send them to factor nodesf : m V f (v) messages are like functions depending on one variable a message from a nodento a neighboring noden can be generated as soon asnreceived messages from all its neighbors exceptn hence, the method can start at the leaf nodes and follow the branches of the tree until the node of interest is met (dynamic programing principle) Learning and Inference in Graphical Models. Chapter 04 p. 9/23

20 Max-sum algorithm How are the messages created? messages from unobserved variable nodes to factor nodes: m X f (x) = m fi X(x)+0 i iff,f,...,f n are the neighbors ofx messages from observed variable nodes to factor nodes: m X f (x) = { 0 ifx = x m fi X(x)+ otherwise i iff,f,...,f n are the neighbors ofx andx is the observed value atx messages from factor nodes to variable nodes: ( m f X (x) = max logf(x,y,...,y n )+ y,...,y n i ifx,y,...,y n are the neighbors off m Yi f(y i ) ) Learning and Inference in Graphical Models. Chapter 04 p. 20/23

21 Max-sum algorithm How do we calculateargmax u logp(u = u,o = o) with the max-sum algorithm? basic idea: backtracking in each maximization step we remind the maximizing value of each variable after having calculated the maximum over the whole tree, we backtrack through all branches following the reminded values Example: f f 2 U U 2 U 3 with binary variablesu,u 2,U 3 blackboard f (u,u 2 ) u 2 = 0 u 2 = u = 0 4 u = f 2 (u 2,u 3 ) u 3 = 0 u 3 = u 2 = 0 u 2 = Learning and Inference in Graphical Models. Chapter 04 p. 2/23

22 Non-polytrees Can we apply max-sum or sum-product also in non-polytree structures? no loopy belief propagation (Frey and MacKay, 998) EM/ECM algorithm variational methods Monte Carlo methods Learning and Inference in Graphical Models. Chapter 04 p. 22/23

23 Summary sum-product algorithm (belief propagation) max-sum algorithm Learning and Inference in Graphical Models. Chapter 04 p. 23/23

Chapter 03: Bayesian Networks

Chapter 03: Bayesian Networks LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 03: Bayesian Networks Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute of Technology Institute of Measurement and Control

More information

Chapter 05: Hidden Markov Models

Chapter 05: Hidden Markov Models LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 05: Hidden Markov Models Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute of Technology Institute of Measurement and Control

More information

Chapter 08: Direct Maximum Likelihood/MAP Estimation and Incomplete Data Problems

Chapter 08: Direct Maximum Likelihood/MAP Estimation and Incomplete Data Problems LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 08: Direct Maximum Likelihood/MAP Estimation and Incomplete Data Problems Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute

More information

Machine Learning Lecture 14

Machine Learning Lecture 14 Many slides adapted from B. Schiele, S. Roth, Z. Gharahmani Machine Learning Lecture 14 Undirected Graphical Models & Inference 23.06.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de

More information

Bayesian Networks: Belief Propagation in Singly Connected Networks

Bayesian Networks: Belief Propagation in Singly Connected Networks Bayesian Networks: Belief Propagation in Singly Connected Networks Huizhen u janey.yu@cs.helsinki.fi Dept. Computer Science, Univ. of Helsinki Probabilistic Models, Spring, 2010 Huizhen u (U.H.) Bayesian

More information

Inference in Bayesian Networks

Inference in Bayesian Networks Andrea Passerini passerini@disi.unitn.it Machine Learning Inference in graphical models Description Assume we have evidence e on the state of a subset of variables E in the model (i.e. Bayesian Network)

More information

Intelligent Systems I

Intelligent Systems I Intelligent Systems I 00 INTRODUCTION Stefan Harmeling & Philipp Hennig 24. October 2013 Max Planck Institute for Intelligent Systems Dptmt. of Empirical Inference Which Card? Opening Experiment Which

More information

Chapter 10: Random Fields

Chapter 10: Random Fields LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 10: Random Fields Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute of Technology Institute of Measurement and Control

More information

p L yi z n m x N n xi

p L yi z n m x N n xi y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen

More information

9 Forward-backward algorithm, sum-product on factor graphs

9 Forward-backward algorithm, sum-product on factor graphs Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous

More information

Graphical Models and Kernel Methods

Graphical Models and Kernel Methods Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.

More information

Outline. Bayesian Networks: Belief Propagation in Singly Connected Networks. Form of Evidence and Notation. Motivation

Outline. Bayesian Networks: Belief Propagation in Singly Connected Networks. Form of Evidence and Notation. Motivation Outline Bayesian : in Singly Connected Huizhen u Dept. Computer Science, Uni. of Helsinki Huizhen u (U.H.) Bayesian : in Singly Connected Feb. 16 1 / 25 Huizhen u (U.H.) Bayesian : in Singly Connected

More information

Belief Update in CLG Bayesian Networks With Lazy Propagation

Belief Update in CLG Bayesian Networks With Lazy Propagation Belief Update in CLG Bayesian Networks With Lazy Propagation Anders L Madsen HUGIN Expert A/S Gasværksvej 5 9000 Aalborg, Denmark Anders.L.Madsen@hugin.com Abstract In recent years Bayesian networks (BNs)

More information

Bayesian Machine Learning

Bayesian Machine Learning Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September

More information

6.867 Machine learning, lecture 23 (Jaakkola)

6.867 Machine learning, lecture 23 (Jaakkola) Lecture topics: Markov Random Fields Probabilistic inference Markov Random Fields We will briefly go over undirected graphical models or Markov Random Fields (MRFs) as they will be needed in the context

More information

Introduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah

Introduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah Introduction to Graphical Models Srikumar Ramalingam School of Computing University of Utah Reference Christopher M. Bishop, Pattern Recognition and Machine Learning, Jonathan S. Yedidia, William T. Freeman,

More information

Fractional Belief Propagation

Fractional Belief Propagation Fractional Belief Propagation im iegerinck and Tom Heskes S, niversity of ijmegen Geert Grooteplein 21, 6525 EZ, ijmegen, the etherlands wimw,tom @snn.kun.nl Abstract e consider loopy belief propagation

More information

Introduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah

Introduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah Introduction to Graphical Models Srikumar Ramalingam School of Computing University of Utah Reference Christopher M. Bishop, Pattern Recognition and Machine Learning, Jonathan S. Yedidia, William T. Freeman,

More information

Learning in Markov Random Fields An Empirical Study

Learning in Markov Random Fields An Empirical Study Learning in Markov Random Fields An Empirical Study Sridevi Parise, Max Welling University of California, Irvine sparise,welling@ics.uci.edu Abstract Learning the parameters of an undirected graphical

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft

More information

Introduction to Probabilistic Graphical Models

Introduction to Probabilistic Graphical Models Introduction to Probabilistic Graphical Models Kyu-Baek Hwang and Byoung-Tak Zhang Biointelligence Lab School of Computer Science and Engineering Seoul National University Seoul 151-742 Korea E-mail: kbhwang@bi.snu.ac.kr

More information

A Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004

A Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004 A Brief Introduction to Graphical Models Presenter: Yijuan Lu November 12,2004 References Introduction to Graphical Models, Kevin Murphy, Technical Report, May 2001 Learning in Graphical Models, Michael

More information

Fixed Point Solutions of Belief Propagation

Fixed Point Solutions of Belief Propagation Fixed Point Solutions of Belief Propagation Christian Knoll Signal Processing and Speech Comm. Lab. Graz University of Technology christian.knoll@tugraz.at Dhagash Mehta Dept of ACMS University of Notre

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Approximate Inference Part 1 of 2

Approximate Inference Part 1 of 2 Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ Bayesian paradigm Consistent use of probability theory

More information

Probabilistic and Bayesian Machine Learning

Probabilistic and Bayesian Machine Learning Probabilistic and Bayesian Machine Learning Day 4: Expectation and Belief Propagation Yee Whye Teh ywteh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Probabilistic Graphical Networks: Definitions and Basic Results

Probabilistic Graphical Networks: Definitions and Basic Results This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical

More information

Artificial Intelligence: Cognitive Agents

Artificial Intelligence: Cognitive Agents Artificial Intelligence: Cognitive Agents AI, Uncertainty & Bayesian Networks 2015-03-10 / 03-12 Kim, Byoung-Hee Biointelligence Laboratory Seoul National University http://bi.snu.ac.kr A Bayesian network

More information

Junction Tree, BP and Variational Methods

Junction Tree, BP and Variational Methods Junction Tree, BP and Variational Methods Adrian Weller MLSALT4 Lecture Feb 21, 2018 With thanks to David Sontag (MIT) and Tony Jebara (Columbia) for use of many slides and illustrations For more information,

More information

4 : Exact Inference: Variable Elimination

4 : Exact Inference: Variable Elimination 10-708: Probabilistic Graphical Models 10-708, Spring 2014 4 : Exact Inference: Variable Elimination Lecturer: Eric P. ing Scribes: Soumya Batra, Pradeep Dasigi, Manzil Zaheer 1 Probabilistic Inference

More information

Lecture 6: Graphical Models: Learning

Lecture 6: Graphical Models: Learning Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)

More information

Approximate Message Passing

Approximate Message Passing Approximate Message Passing Mohammad Emtiyaz Khan CS, UBC February 8, 2012 Abstract In this note, I summarize Sections 5.1 and 5.2 of Arian Maleki s PhD thesis. 1 Notation We denote scalars by small letters

More information

Probabilistic Graphical Models. Guest Lecture by Narges Razavian Machine Learning Class April

Probabilistic Graphical Models. Guest Lecture by Narges Razavian Machine Learning Class April Probabilistic Graphical Models Guest Lecture by Narges Razavian Machine Learning Class April 14 2017 Today What is probabilistic graphical model and why it is useful? Bayesian Networks Basic Inference

More information

Probabilistic Graphical Models (I)

Probabilistic Graphical Models (I) Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random

More information

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005 Instructor: Professor Jeff A. Bilmes Uncertainty & Bayesian Networks

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 16 Undirected Graphs Undirected Separation Inferring Marginals & Conditionals Moralization Junction Trees Triangulation Undirected Graphs Separation

More information

Graphical models: parameter learning

Graphical models: parameter learning Graphical models: parameter learning Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London London WC1N 3AR, England http://www.gatsby.ucl.ac.uk/ zoubin/ zoubin@gatsby.ucl.ac.uk

More information

Message Passing and Junction Tree Algorithms. Kayhan Batmanghelich

Message Passing and Junction Tree Algorithms. Kayhan Batmanghelich Message Passing and Junction Tree Algorithms Kayhan Batmanghelich 1 Review 2 Review 3 Great Ideas in ML: Message Passing Each soldier receives reports from all branches of tree 3 here 7 here 1 of me 11

More information

Machine Learning Summer School

Machine Learning Summer School Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,

More information

Nash Propagation for Loopy Graphical Games

Nash Propagation for Loopy Graphical Games Nash Propagation for Loopy Graphical Games Luis E. Ortiz Michael Kearns Department of Computer and Information Science University of Pennsylvania fleortiz,mkearnsg@cis.upenn.edu Abstract We introduce NashProp,

More information

A New Approach for Hybrid Bayesian Networks Using Full Densities

A New Approach for Hybrid Bayesian Networks Using Full Densities A New Approach for Hybrid Bayesian Networks Using Full Densities Oliver C. Schrempf Institute of Computer Design and Fault Tolerance Universität Karlsruhe e-mail: schrempf@ira.uka.de Uwe D. Hanebeck Institute

More information

Correctness of belief propagation in Gaussian graphical models of arbitrary topology

Correctness of belief propagation in Gaussian graphical models of arbitrary topology Correctness of belief propagation in Gaussian graphical models of arbitrary topology Yair Weiss Computer Science Division UC Berkeley, 485 Soda Hall Berkeley, CA 94720-1776 Phone: 510-642-5029 yweiss@cs.berkeley.edu

More information

Lecture 16 Deep Neural Generative Models

Lecture 16 Deep Neural Generative Models Lecture 16 Deep Neural Generative Models CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 22, 2017 Approach so far: We have considered simple models and then constructed

More information

Statistical Approaches to Learning and Discovery

Statistical Approaches to Learning and Discovery Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University

More information

Approximate Inference Part 1 of 2

Approximate Inference Part 1 of 2 Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ 1 Bayesian paradigm Consistent use of probability theory

More information

Improving Search-Based Inference in Bayesian Networks

Improving Search-Based Inference in Bayesian Networks From: Proceedings of the Eleventh International FLAIRS Conference. Copyright 1998, AAAI (www.aaai.org). All rights reserved. Improving Search-Based Inference in Bayesian Networks E. Castillo, J. M. Guti~rrez

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

13: Variational inference II

13: Variational inference II 10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational

More information

Density Propagation for Continuous Temporal Chains Generative and Discriminative Models

Density Propagation for Continuous Temporal Chains Generative and Discriminative Models $ Technical Report, University of Toronto, CSRG-501, October 2004 Density Propagation for Continuous Temporal Chains Generative and Discriminative Models Cristian Sminchisescu and Allan Jepson Department

More information

Part I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS

Part I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Part I C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Probabilistic Graphical Models Graphical representation of a probabilistic model Each variable corresponds to a

More information

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception COURSE INTRODUCTION COMPUTATIONAL MODELING OF VISUAL PERCEPTION 2 The goal of this course is to provide a framework and computational tools for modeling visual inference, motivated by interesting examples

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture Notes Fall 2009 November, 2009 Byoung-Ta Zhang School of Computer Science and Engineering & Cognitive Science, Brain Science, and Bioinformatics Seoul National University

More information

Applying Bayesian networks in the game of Minesweeper

Applying Bayesian networks in the game of Minesweeper Applying Bayesian networks in the game of Minesweeper Marta Vomlelová Faculty of Mathematics and Physics Charles University in Prague http://kti.mff.cuni.cz/~marta/ Jiří Vomlel Institute of Information

More information

Bayesian networks: approximate inference

Bayesian networks: approximate inference Bayesian networks: approximate inference Machine Intelligence Thomas D. Nielsen September 2008 Approximative inference September 2008 1 / 25 Motivation Because of the (worst-case) intractability of exact

More information

Modeling High-Dimensional Discrete Data with Multi-Layer Neural Networks

Modeling High-Dimensional Discrete Data with Multi-Layer Neural Networks Modeling High-Dimensional Discrete Data with Multi-Layer Neural Networks Yoshua Bengio Dept. IRO Université de Montréal Montreal, Qc, Canada, H3C 3J7 bengioy@iro.umontreal.ca Samy Bengio IDIAP CP 592,

More information

Bayesian network modeling. 1

Bayesian network modeling.  1 Bayesian network modeling http://springuniversity.bc3research.org/ 1 Probabilistic vs. deterministic modeling approaches Probabilistic Explanatory power (e.g., r 2 ) Explanation why Based on inductive

More information

Bayesian belief networks. Inference.

Bayesian belief networks. Inference. Lecture 13 Bayesian belief networks. Inference. Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Midterm exam Monday, March 17, 2003 In class Closed book Material covered by Wednesday, March 12 Last

More information

The computational power of dynamic bayesian networks

The computational power of dynamic bayesian networks The computational power of dynamic bayesian networks Joshua Brulé Department of Computer Science University of Maryland, College Park jbrule@cs.umd.edu Abstract. This paper considers the computational

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 9 Undirected Models CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due next Wednesday (Nov 4) in class Start early!!! Project milestones due Monday (Nov 9)

More information

Machine Learning for Data Science (CS4786) Lecture 24

Machine Learning for Data Science (CS4786) Lecture 24 Machine Learning for Data Science (CS4786) Lecture 24 Graphical Models: Approximate Inference Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ BELIEF PROPAGATION OR MESSAGE PASSING Each

More information

Bayesian networks. Chapter 14, Sections 1 4

Bayesian networks. Chapter 14, Sections 1 4 Bayesian networks Chapter 14, Sections 1 4 Artificial Intelligence, spring 2013, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 14, Sections 1 4 1 Bayesian networks

More information

Graphical models. Sunita Sarawagi IIT Bombay

Graphical models. Sunita Sarawagi IIT Bombay 1 Graphical models Sunita Sarawagi IIT Bombay http://www.cse.iitb.ac.in/~sunita 2 Probabilistic modeling Given: several variables: x 1,... x n, n is large. Task: build a joint distribution function Pr(x

More information

B C D E E E" D C B D" C" A A" A A A A

B C D E E E D C B D C A A A A A A 1 MERL { A MITSUBISHI ELECTRIC RESEARCH LABORATORY http://www.merl.com Correctness of belief propagation in Gaussian graphical models of arbitrary topology Yair Weiss William T. Freeman y TR-99-38 October

More information

Bayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling

Bayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling 12735: Urban Systems Modeling Lec. 09 Bayesian Networks instructor: Matteo Pozzi x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 1 outline example of applications how to shape a problem as a BN complexity of the inference

More information

Artificial Intelligence

Artificial Intelligence ICS461 Fall 2010 Nancy E. Reed nreed@hawaii.edu 1 Lecture #14B Outline Inference in Bayesian Networks Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic

More information

Expectation Propagation Algorithm

Expectation Propagation Algorithm Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,

More information

CS281A/Stat241A Lecture 19

CS281A/Stat241A Lecture 19 CS281A/Stat241A Lecture 19 p. 1/4 CS281A/Stat241A Lecture 19 Junction Tree Algorithm Peter Bartlett CS281A/Stat241A Lecture 19 p. 2/4 Announcements My office hours: Tuesday Nov 3 (today), 1-2pm, in 723

More information

Induction of Decision Trees

Induction of Decision Trees Induction of Decision Trees Peter Waiganjo Wagacha This notes are for ICS320 Foundations of Learning and Adaptive Systems Institute of Computer Science University of Nairobi PO Box 30197, 00200 Nairobi.

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Expectation Propagation for Approximate Bayesian Inference

Expectation Propagation for Approximate Bayesian Inference Expectation Propagation for Approximate Bayesian Inference José Miguel Hernández Lobato Universidad Autónoma de Madrid, Computer Science Department February 5, 2007 1/ 24 Bayesian Inference Inference Given

More information

Maximum Weight Matching via Max-Product Belief Propagation

Maximum Weight Matching via Max-Product Belief Propagation Maximum Weight Matching via Max-Product Belief Propagation Mohsen Bayati Department of EE Stanford University Stanford, CA 94305 Email: bayati@stanford.edu Devavrat Shah Departments of EECS & ESD MIT Boston,

More information

Bayesian Networks Inference with Probabilistic Graphical Models

Bayesian Networks Inference with Probabilistic Graphical Models 4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning

More information

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

A graph contains a set of nodes (vertices) connected by links (edges or arcs) BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,

More information

Bayesian Methods in Artificial Intelligence

Bayesian Methods in Artificial Intelligence WDS'10 Proceedings of Contributed Papers, Part I, 25 30, 2010. ISBN 978-80-7378-139-2 MATFYZPRESS Bayesian Methods in Artificial Intelligence M. Kukačka Charles University, Faculty of Mathematics and Physics,

More information

CS Lecture 4. Markov Random Fields

CS Lecture 4. Markov Random Fields CS 6347 Lecture 4 Markov Random Fields Recap Announcements First homework is available on elearning Reminder: Office hours Tuesday from 10am-11am Last Time Bayesian networks Today Markov random fields

More information

Probabilistic Graphical Models for Image Analysis - Lecture 1

Probabilistic Graphical Models for Image Analysis - Lecture 1 Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.

More information

Probabilistic Graphical Models: Representation and Inference

Probabilistic Graphical Models: Representation and Inference Probabilistic Graphical Models: Representation and Inference Aaron C. Courville Université de Montréal Note: Material for the slides is taken directly from a presentation prepared by Andrew Moore 1 Overview

More information

Inference in Graphical Models Variable Elimination and Message Passing Algorithm

Inference in Graphical Models Variable Elimination and Message Passing Algorithm Inference in Graphical Models Variable Elimination and Message Passing lgorithm Le Song Machine Learning II: dvanced Topics SE 8803ML, Spring 2012 onditional Independence ssumptions Local Markov ssumption

More information

Variational Inference (11/04/13)

Variational Inference (11/04/13) STA561: Probabilistic machine learning Variational Inference (11/04/13) Lecturer: Barbara Engelhardt Scribes: Matt Dickenson, Alireza Samany, Tracy Schifeling 1 Introduction In this lecture we will further

More information

14 : Theory of Variational Inference: Inner and Outer Approximation

14 : Theory of Variational Inference: Inner and Outer Approximation 10-708: Probabilistic Graphical Models 10-708, Spring 2014 14 : Theory of Variational Inference: Inner and Outer Approximation Lecturer: Eric P. Xing Scribes: Yu-Hsin Kuo, Amos Ng 1 Introduction Last lecture

More information

Factor Graphs and Message Passing Algorithms Part 1: Introduction

Factor Graphs and Message Passing Algorithms Part 1: Introduction Factor Graphs and Message Passing Algorithms Part 1: Introduction Hans-Andrea Loeliger December 2007 1 The Two Basic Problems 1. Marginalization: Compute f k (x k ) f(x 1,..., x n ) x 1,..., x n except

More information

Decisiveness in Loopy Propagation

Decisiveness in Loopy Propagation Decisiveness in Loopy Propagation Janneke H. Bolt and Linda C. van der Gaag Department of Information and Computing Sciences, Utrecht University P.O. Box 80.089, 3508 TB Utrecht, The Netherlands Abstract.

More information

Human-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg

Human-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg Temporal Reasoning Kai Arras, University of Freiburg 1 Temporal Reasoning Contents Introduction Temporal Reasoning Hidden Markov Models Linear Dynamical Systems (LDS) Kalman Filter 2 Temporal Reasoning

More information

13 : Variational Inference: Loopy Belief Propagation and Mean Field

13 : Variational Inference: Loopy Belief Propagation and Mean Field 10-708: Probabilistic Graphical Models 10-708, Spring 2012 13 : Variational Inference: Loopy Belief Propagation and Mean Field Lecturer: Eric P. Xing Scribes: Peter Schulam and William Wang 1 Introduction

More information

12 : Variational Inference I

12 : Variational Inference I 10-708: Probabilistic Graphical Models, Spring 2015 12 : Variational Inference I Lecturer: Eric P. Xing Scribes: Fattaneh Jabbari, Eric Lei, Evan Shapiro 1 Introduction Probabilistic inference is one of

More information

Intelligent Systems:

Intelligent Systems: Intelligent Systems: Undirected Graphical models (Factor Graphs) (2 lectures) Carsten Rother 15/01/2015 Intelligent Systems: Probabilistic Inference in DGM and UGM Roadmap for next two lectures Definition

More information

Inference and Representation

Inference and Representation Inference and Representation David Sontag New York University Lecture 5, Sept. 30, 2014 David Sontag (NYU) Inference and Representation Lecture 5, Sept. 30, 2014 1 / 16 Today s lecture 1 Running-time of

More information

Part 1: Expectation Propagation

Part 1: Expectation Propagation Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud

More information

Chapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang

Chapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang Chapter 4 Dynamic Bayesian Networks 2016 Fall Jin Gu, Michael Zhang Reviews: BN Representation Basic steps for BN representations Define variables Define the preliminary relations between variables Check

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 10 Alternatives to Monte Carlo Computation Since about 1990, Markov chain Monte Carlo has been the dominant

More information

Symbolic Variable Elimination in Discrete and Continuous Graphical Models. Scott Sanner Ehsan Abbasnejad

Symbolic Variable Elimination in Discrete and Continuous Graphical Models. Scott Sanner Ehsan Abbasnejad Symbolic Variable Elimination in Discrete and Continuous Graphical Models Scott Sanner Ehsan Abbasnejad Inference for Dynamic Tracking No one previously did this inference exactly in closed-form! Exact

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information

Multivariate Bayesian Linear Regression MLAI Lecture 11

Multivariate Bayesian Linear Regression MLAI Lecture 11 Multivariate Bayesian Linear Regression MLAI Lecture 11 Neil D. Lawrence Department of Computer Science Sheffield University 21st October 2012 Outline Univariate Bayesian Linear Regression Multivariate

More information

Introduction To Graphical Models

Introduction To Graphical Models Peter Gehler Introduction to Graphical Models Introduction To Graphical Models Peter V. Gehler Max Planck Institute for Intelligent Systems, Tübingen, Germany ENS/INRIA Summer School, Paris, July 2013

More information

Readings: K&F: 16.3, 16.4, Graphical Models Carlos Guestrin Carnegie Mellon University October 6 th, 2008

Readings: K&F: 16.3, 16.4, Graphical Models Carlos Guestrin Carnegie Mellon University October 6 th, 2008 Readings: K&F: 16.3, 16.4, 17.3 Bayesian Param. Learning Bayesian Structure Learning Graphical Models 10708 Carlos Guestrin Carnegie Mellon University October 6 th, 2008 10-708 Carlos Guestrin 2006-2008

More information

Outline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination

Outline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination Probabilistic Graphical Models COMP 790-90 Seminar Spring 2011 The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Outline It Introduction ti Representation Bayesian network Conditional Independence Inference:

More information

T Machine Learning: Basic Principles

T Machine Learning: Basic Principles Machine Learning: Basic Principles Bayesian Networks Laboratory of Computer and Information Science (CIS) Department of Computer Science and Engineering Helsinki University of Technology (TKK) Autumn 2007

More information

Lecture 9: PGM Learning

Lecture 9: PGM Learning 13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and

More information