Chapter 04: Exact Inference in Bayesian Networks
|
|
- Domenic Harvey
- 6 years ago
- Views:
Transcription
1 LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 04: Exact Inference in Bayesian Networks Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute of Technology Institute of Measurement and Control Systems Learning and Inference in Graphical Models. Chapter 04 p. /23
2 References for this chapter Christopher M. Bishop, Pattern Recognition and Machine Learning, ch. 8, Springer, 2006 Stuart Russell and Peter Norvig, Artificial Intelligenece: A Modern Approach, ch. 4, Prentice Hall, 2003 Judea Pearl, Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference, Morgan Kaufmann, 989 Steffen L. Lauritzen and David J. Spiegelhalter, Local computations with probabilities on graphical structures and their applications to expert systems, In: The Journal of the Royal Statistical Society, vol. 50, no. 2, pp , 988 Brendan J. Frey and David J. C. MacKay, A Revolution: Belief Propagation in Graphs with Cycles, In: Advances in Neural Information Processing Systems (NIPS), vol. 0, pp , 997, http: //books.nips.cc/papers/files/nips0/0479.pdf Learning and Inference in Graphical Models. Chapter 04 p. 2/23
3 Inference Given a graphical model with unobserved nodes U and observed nodeso, we want to conclude about the distribution of unobserved nodes calculate the single node marginal distribution p(x O) withx U calculate the maximum-a-posterior estimator arg max u p(u = u O = o) Exact inference is not always possible. We focus on the easier cases, polytrees example of a polytree a polytree is a directed acyclic graph whose underlying undirected graph is acyclic. counterexample of a polytree Learning and Inference in Graphical Models. Chapter 04 p. 3/23
4 Marginalization Calculatep(X = x O = o,o 2 = o 2 ) for the network on the right. p (X = x O = o,o 2 = o 2 ) with = p(x = x,o = o,o 2 = o 2 ) p(o = o,o 2 = o 2 ) p(x = x,o = o,o 2 = o 2 ) = p(x = x,o = o,o 2 = o 2 )dx p (X = x,o = o,o 2 = o 2 ) = p(x=x,o =o,o 2 =o 2,U =u,..., U U 2 U 3 X U 4 U 5 O continue on blackboard U 7 =u 7 ) du du 2 du 3 du 4 du 5 du 6 du 7 U 6 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 4/23
5 Factor graph A factor graph is a bipartite graph with two kind of nodes: variable nodes that model random variables like in a Bayesian network factor nodes that model a probabilistic relationship between variable nodes. Each factor node is assigned with a factor, i.e. a function that models the stochastic relationship Variable nodes and factor nodes are connected by undirected links. For each Bayesian polytree we can create a factor graph as follows: the set of variable nodes is taken from the nodes of the Bayesian polytree for each factorp(x Pred(X) in the Bayesian network we create a new factor nodef we connectx andpred(x) withf we assignf(x,y,...,y n ) p(x = x Pred(X) = (y,...,y n )) Hence, the joint probability of the Bayesian polytree is equal to the product of all factors of the factor tree. Learning and Inference in Graphical Models. Chapter 04 p. 5/23
6 Factor graphs Example: factor graph for the Bayesian network on the right f (u )=p(u = u ) f 2 (u,u 2 )=p(u 2 = u 2 U = u ) f 3 (u 3 )=p(u 3 = u 3 ) f 4 (u 2,u 3,x)=p(X = x U 2 = u 2,U 3 = u 3 ) f 5 (u 4 )=p(u 4 = u 4 ) f 6 (x,u 4,u 5 )=p(u 5 = u 5 X = x,u 4 = u 4 ) f 7 (u 4,o )=p(o = o U 4 = u 4 ) f 8 (u 5,u 6 )=p(u 6 = u 6 U 5 = u 5 ) f 9 (u 5,o 2 )=p(o 2 = o 2 U 5 = u 5 ) f 0 (o,u 7 )=p(u 7 = u 7 O = o ) f U f 2 f 3 U 2 U 3 f 4 f 5 X U 4 f 6 f 7 U 5 O f 8 f 9 f 0 U 6 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 6/23
7 Marginalization on factor graphs Task: calculate I(x) = f (u )f 2 (u,u 2 )f 3 (u 3 ) f 4 (u 2,u 3,x)f 5 (u 4 )f 6 (x,u 4,u 5 ) f 7 (u 4,o )f 8 (u 5,u 6 )f 9 (u 5,o 2 )f 0 (o,u 7 ) du du 2 du 3 du 4 du 5 du 6 du 7 Observations: factor graph is tree with rootx I(x) can be split into two large factors m f4 X(x)= f (u )f 2 (u,u 2 )f 3 (u 3 ) f 4 (u 2,u 3,x)du du 2 du 3 m f6 X(x)= f 5 (u 4 )f 6 (x,u 4,u 5 ) f U f 2 f 3 U 2 U 3 f 4 f 5 X U 4 f 6 f 7 U 5 O f 8 f 9 f 0 U 6 O 2 U 7 f 7 (u 4,o )f 8 (u 5,u 6 )f 9 (u 5,o 2 )f 0 (o,u 7 )du 4 du 5 du 6 du 7 Learning and Inference in Graphical Models. Chapter 04 p. 7/23
8 Marginalization on factor graphs Inm f4 x we can factor outf 3 andf 4 ( m f4 X(x)= ) f (u )f 2 (u,u 2 )du } {{ } =:m f2 U 2 (u 2 ) and rewritem f2 U 2 (u 2 ) as m f2 U 2 (u 2 )= f (u ) f }{{} 2 (u,u 2 )du =:m f U (u ) Observations: calculation can be split along the branches of the tree the lead nodes can serve as starting points for the calculation f 3 (u 3 ) f }{{} 4 (u 2,u 3,x)du 2 du 3 =:m f3 U 3 (u 3 ) f U f 2 f 3 U 2 U 3 f 4 only multiplication and integration/summation occur intermediate results can be interpreted as messages sent from one node to its neighbors X Learning and Inference in Graphical Models. Chapter 04 p. 8/23
9 Marginalization on factor graphs m f6 X(x) can be split in a similar manner m f6 X(x)= f 6 (x,u 4,u 5 )m U4 f 6 (u 4 )m U5 f 6 (u 5 )du 4 du 5 m U4 f 6 (u 4 )=m f5 U 4 (u 4 ) m f7 U 4 (u 4 ) m U5 f 6 (u 4 )=m f8 U 5 (u 5 ) m f9 U 5 (u 5 ) m f8 U 5 (u 5 )= f 8 (u 5,u 6 )m U6 f 8 (u 6 )du 6 m U6 f 8 (u 6 )= m f9 U 5 (u 5 )=f 9 (u 5,o 2 ) with observedo 2 If we want to extend the procedure to observed nodes, we could also argue m f9 U 5 (u 5 )= f 9 (u 5,o 2)m O2 f 9 (o 2)do 2 f 5 X U 4 f 6 f 7 U 5 O f 8 f 9 f 0 U 6 O 2 U 7 m O2 f 9 (o 2) = δ(o 2 o 2 ) withδ is the Dirac distribution Learning and Inference in Graphical Models. Chapter 04 p. 9/23
10 Side topic: Dirac distribution The Diracδ is a distribution that can be used to model discrete distributions in continuous space: δ(x) = so that Examples { 0 ifx 0 ifx = 0 δ(x)dx = ifx is distributed w.r.t. the Dirac distribution,x can only take the value0 ify is distributed w.r.t. 0.3 δ(y 2)+0.7 δ(y +5.),Y will take the value2with probability of0.3 and 5. with probability of0.7 Learning and Inference in Graphical Models. Chapter 04 p. 0/23
11 Belief propagation Example motivates a generic algorithm known as sum-product algorithm or belief propagation (Pearl, 989; Lauritzen and Spiegelhalter, 988) factor nodesf generate messages and send them to variable nodesv: m f V (v) variable nodesv generate messages and send them to factor nodesf : m V f (v) messages are like distributions (but not necessarily normalized) a message from a nodento a neighboring noden can be generated as soon asnreceived messages from all its neighbors exceptn hence, the method can start at the leaf nodes and follow the branches of the tree until the node of interest is met (dynamic programing principle) Learning and Inference in Graphical Models. Chapter 04 p. /23
12 Belief propagation How are the messages created? messages from unobserved variable nodes to factor nodes: m X f (x) = m fi X(x) i iff,f,...,f n are the neighbors ofx messages from observed variable nodes to factor nodes: m X f (x) = δ(x x ) m fi X(x) i iff,f,...,f n are the neighbors ofx andx is the observed value atx messages from factor nodes to variable nodes: m f X (x) = ifx,y,...,y n are the neighbors off f(x,y,...,y n ) m Yi f(y i )dy...dy n i Learning and Inference in Graphical Models. Chapter 04 p. 2/23
13 Belief propagation Example ( exam problem ) c P(c=c) m + P(M=m) 3 4 P (E=e C=c,M=m) = ife=c+m ife=c ife=c m 0 otherwise level of comprehension mental mood result of exam apply belief propagation to calculate P(C E) for E = 2 blackboard Learning and Inference in Graphical Models. Chapter 04 p. 3/23
14 Belief propagation Belief propagation works if either or all distributions are categorical (e.g. example) all distributions are conjugate all distributions are Gaussian and variables depend linearly, i.e. X Y,...,Y n N(c Y + +c n Y n,σ 2 ) with fixed values c,...,c n,σ 2 Otherwise, integrals might become untreatable analytically Gauss-linear example: exam problem with C N(3,4) M N(0,) E C,M N(C +M,) blackboard/homework Learning and Inference in Graphical Models. Chapter 04 p. 4/23
15 MAP estimator Second task, calculate the maximum-a-posterior estimator arg maxp(u = u O = o) u Again, we focus on polytrees only. Calculate the MAP estimator for the model on the right. arg max u = argmax u = argmax u = argmax u p(u = u O = o) p(u = u,o = o) p(o = o) p(u = u,o = o) logp(u = u,o = o) U U 2 U 3 U 8 U 4 U 5 O U 6 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 5/23
16 MAP estimator logp(u = u,o = o) = log ( f(u i,pred(u i )) i f(o j,pred(o j )) ) = i j logf(u i,pred(u i ))+ logf(o j,pred(o j )) j choose one node as root node (e.g. U 8 ): max logp(u = u,o = o) u,...,u 8 ( = max mf4 U 8 (u 8 )+m f6 U 8 (u 8 ) ) u 8 wherem f4 U 8 contains all terms reated tof,...,f 4 andm f6 U 8 contains all terms related tof 5,...,f 0 f U f 2 f 3 U 2 U 3 f 4 f 5 U 8 U 4 f 6 f 7 U 6 U 5 O f 8 f 9 f 0 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 6/23
17 MAP estimator m f6 U 8 (u 8 )=max u 4,u 5 ( logf6 (u 4,u 5,u 8 )+m U4 f 6 (u 4 )+m U5 f 6 (u 5 ) ) m U5 f 6 (u 5 )=m f8 U 5 (u 5 )+m f9 U 5 (u 5 ) ( m f8 U 5 (u 5 )=max logf8 (u 5,u 6 )+m U6 f 8 (u 6 ) ) u 6 m U6 f 8 (u 6 )=0 m f9 U 5 (u 5 )=logf 9 (u 5,o 2 ) ( =max logf9 (u 5,o 2)+m O2 o f 9 (o 2) ) 2 { m O2 f 9 (o 0 ifo 2 = o 2 2)= otherwise f 5 U 8 U 4 f 6 f 7 U 5 O f 8 f 9 f 0 U 6 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 7/23
18 MAP estimator m U4 f 6 (u 4 )=m f5 U 4 (u 4 )+m f7 U 4 (u 4 ) m f5 U 4 (u 4 )=logf 4 (u 4 ) m f7 U 4 (u 4 )=logf 7 (u 4,o )+something ( =max logf7 (u 4,o )+m O o f 7 (o ) ) m O f 7 (o )=m f0 O (o )+ { 0 ifo = o otherwise m f0 O (o )=max u 7 ( logf0 (o,u 7 )+m U7 f 0 (u 7 ) ) m U7 f 0 (u 7 )=0 f 5 U 8 U 4 f 6 f 7 U 5 O f 8 f 9 f 0 U 6 O 2 U 7 Learning and Inference in Graphical Models. Chapter 04 p. 8/23
19 Max-sum algorithm Example motivates a generic algorithm to calculatemax u logp(u = u,o = o) known as max-sum algorithm factor nodesf generate messages and send them to variable nodesv: m f V (v) variable nodesv generate messages and send them to factor nodesf : m V f (v) messages are like functions depending on one variable a message from a nodento a neighboring noden can be generated as soon asnreceived messages from all its neighbors exceptn hence, the method can start at the leaf nodes and follow the branches of the tree until the node of interest is met (dynamic programing principle) Learning and Inference in Graphical Models. Chapter 04 p. 9/23
20 Max-sum algorithm How are the messages created? messages from unobserved variable nodes to factor nodes: m X f (x) = m fi X(x)+0 i iff,f,...,f n are the neighbors ofx messages from observed variable nodes to factor nodes: m X f (x) = { 0 ifx = x m fi X(x)+ otherwise i iff,f,...,f n are the neighbors ofx andx is the observed value atx messages from factor nodes to variable nodes: ( m f X (x) = max logf(x,y,...,y n )+ y,...,y n i ifx,y,...,y n are the neighbors off m Yi f(y i ) ) Learning and Inference in Graphical Models. Chapter 04 p. 20/23
21 Max-sum algorithm How do we calculateargmax u logp(u = u,o = o) with the max-sum algorithm? basic idea: backtracking in each maximization step we remind the maximizing value of each variable after having calculated the maximum over the whole tree, we backtrack through all branches following the reminded values Example: f f 2 U U 2 U 3 with binary variablesu,u 2,U 3 blackboard f (u,u 2 ) u 2 = 0 u 2 = u = 0 4 u = f 2 (u 2,u 3 ) u 3 = 0 u 3 = u 2 = 0 u 2 = Learning and Inference in Graphical Models. Chapter 04 p. 2/23
22 Non-polytrees Can we apply max-sum or sum-product also in non-polytree structures? no loopy belief propagation (Frey and MacKay, 998) EM/ECM algorithm variational methods Monte Carlo methods Learning and Inference in Graphical Models. Chapter 04 p. 22/23
23 Summary sum-product algorithm (belief propagation) max-sum algorithm Learning and Inference in Graphical Models. Chapter 04 p. 23/23
Chapter 03: Bayesian Networks
LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 03: Bayesian Networks Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute of Technology Institute of Measurement and Control
More informationChapter 05: Hidden Markov Models
LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 05: Hidden Markov Models Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute of Technology Institute of Measurement and Control
More informationChapter 08: Direct Maximum Likelihood/MAP Estimation and Incomplete Data Problems
LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 08: Direct Maximum Likelihood/MAP Estimation and Incomplete Data Problems Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute
More informationMachine Learning Lecture 14
Many slides adapted from B. Schiele, S. Roth, Z. Gharahmani Machine Learning Lecture 14 Undirected Graphical Models & Inference 23.06.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de
More informationBayesian Networks: Belief Propagation in Singly Connected Networks
Bayesian Networks: Belief Propagation in Singly Connected Networks Huizhen u janey.yu@cs.helsinki.fi Dept. Computer Science, Univ. of Helsinki Probabilistic Models, Spring, 2010 Huizhen u (U.H.) Bayesian
More informationInference in Bayesian Networks
Andrea Passerini passerini@disi.unitn.it Machine Learning Inference in graphical models Description Assume we have evidence e on the state of a subset of variables E in the model (i.e. Bayesian Network)
More informationIntelligent Systems I
Intelligent Systems I 00 INTRODUCTION Stefan Harmeling & Philipp Hennig 24. October 2013 Max Planck Institute for Intelligent Systems Dptmt. of Empirical Inference Which Card? Opening Experiment Which
More informationChapter 10: Random Fields
LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 10: Random Fields Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute of Technology Institute of Measurement and Control
More informationp L yi z n m x N n xi
y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen
More information9 Forward-backward algorithm, sum-product on factor graphs
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous
More informationGraphical Models and Kernel Methods
Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.
More informationOutline. Bayesian Networks: Belief Propagation in Singly Connected Networks. Form of Evidence and Notation. Motivation
Outline Bayesian : in Singly Connected Huizhen u Dept. Computer Science, Uni. of Helsinki Huizhen u (U.H.) Bayesian : in Singly Connected Feb. 16 1 / 25 Huizhen u (U.H.) Bayesian : in Singly Connected
More informationBelief Update in CLG Bayesian Networks With Lazy Propagation
Belief Update in CLG Bayesian Networks With Lazy Propagation Anders L Madsen HUGIN Expert A/S Gasværksvej 5 9000 Aalborg, Denmark Anders.L.Madsen@hugin.com Abstract In recent years Bayesian networks (BNs)
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September
More information6.867 Machine learning, lecture 23 (Jaakkola)
Lecture topics: Markov Random Fields Probabilistic inference Markov Random Fields We will briefly go over undirected graphical models or Markov Random Fields (MRFs) as they will be needed in the context
More informationIntroduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah
Introduction to Graphical Models Srikumar Ramalingam School of Computing University of Utah Reference Christopher M. Bishop, Pattern Recognition and Machine Learning, Jonathan S. Yedidia, William T. Freeman,
More informationFractional Belief Propagation
Fractional Belief Propagation im iegerinck and Tom Heskes S, niversity of ijmegen Geert Grooteplein 21, 6525 EZ, ijmegen, the etherlands wimw,tom @snn.kun.nl Abstract e consider loopy belief propagation
More informationIntroduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah
Introduction to Graphical Models Srikumar Ramalingam School of Computing University of Utah Reference Christopher M. Bishop, Pattern Recognition and Machine Learning, Jonathan S. Yedidia, William T. Freeman,
More informationLearning in Markov Random Fields An Empirical Study
Learning in Markov Random Fields An Empirical Study Sridevi Parise, Max Welling University of California, Irvine sparise,welling@ics.uci.edu Abstract Learning the parameters of an undirected graphical
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft
More informationIntroduction to Probabilistic Graphical Models
Introduction to Probabilistic Graphical Models Kyu-Baek Hwang and Byoung-Tak Zhang Biointelligence Lab School of Computer Science and Engineering Seoul National University Seoul 151-742 Korea E-mail: kbhwang@bi.snu.ac.kr
More informationA Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004
A Brief Introduction to Graphical Models Presenter: Yijuan Lu November 12,2004 References Introduction to Graphical Models, Kevin Murphy, Technical Report, May 2001 Learning in Graphical Models, Michael
More informationFixed Point Solutions of Belief Propagation
Fixed Point Solutions of Belief Propagation Christian Knoll Signal Processing and Speech Comm. Lab. Graz University of Technology christian.knoll@tugraz.at Dhagash Mehta Dept of ACMS University of Notre
More informationBayesian Networks BY: MOHAMAD ALSABBAGH
Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationApproximate Inference Part 1 of 2
Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ Bayesian paradigm Consistent use of probability theory
More informationProbabilistic and Bayesian Machine Learning
Probabilistic and Bayesian Machine Learning Day 4: Expectation and Belief Propagation Yee Whye Teh ywteh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationProbabilistic Graphical Networks: Definitions and Basic Results
This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical
More informationArtificial Intelligence: Cognitive Agents
Artificial Intelligence: Cognitive Agents AI, Uncertainty & Bayesian Networks 2015-03-10 / 03-12 Kim, Byoung-Hee Biointelligence Laboratory Seoul National University http://bi.snu.ac.kr A Bayesian network
More informationJunction Tree, BP and Variational Methods
Junction Tree, BP and Variational Methods Adrian Weller MLSALT4 Lecture Feb 21, 2018 With thanks to David Sontag (MIT) and Tony Jebara (Columbia) for use of many slides and illustrations For more information,
More information4 : Exact Inference: Variable Elimination
10-708: Probabilistic Graphical Models 10-708, Spring 2014 4 : Exact Inference: Variable Elimination Lecturer: Eric P. ing Scribes: Soumya Batra, Pradeep Dasigi, Manzil Zaheer 1 Probabilistic Inference
More informationLecture 6: Graphical Models: Learning
Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)
More informationApproximate Message Passing
Approximate Message Passing Mohammad Emtiyaz Khan CS, UBC February 8, 2012 Abstract In this note, I summarize Sections 5.1 and 5.2 of Arian Maleki s PhD thesis. 1 Notation We denote scalars by small letters
More informationProbabilistic Graphical Models. Guest Lecture by Narges Razavian Machine Learning Class April
Probabilistic Graphical Models Guest Lecture by Narges Razavian Machine Learning Class April 14 2017 Today What is probabilistic graphical model and why it is useful? Bayesian Networks Basic Inference
More informationProbabilistic Graphical Models (I)
Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random
More informationEE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS
EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005 Instructor: Professor Jeff A. Bilmes Uncertainty & Bayesian Networks
More informationMachine Learning 4771
Machine Learning 4771 Instructor: Tony Jebara Topic 16 Undirected Graphs Undirected Separation Inferring Marginals & Conditionals Moralization Junction Trees Triangulation Undirected Graphs Separation
More informationGraphical models: parameter learning
Graphical models: parameter learning Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London London WC1N 3AR, England http://www.gatsby.ucl.ac.uk/ zoubin/ zoubin@gatsby.ucl.ac.uk
More informationMessage Passing and Junction Tree Algorithms. Kayhan Batmanghelich
Message Passing and Junction Tree Algorithms Kayhan Batmanghelich 1 Review 2 Review 3 Great Ideas in ML: Message Passing Each soldier receives reports from all branches of tree 3 here 7 here 1 of me 11
More informationMachine Learning Summer School
Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,
More informationNash Propagation for Loopy Graphical Games
Nash Propagation for Loopy Graphical Games Luis E. Ortiz Michael Kearns Department of Computer and Information Science University of Pennsylvania fleortiz,mkearnsg@cis.upenn.edu Abstract We introduce NashProp,
More informationA New Approach for Hybrid Bayesian Networks Using Full Densities
A New Approach for Hybrid Bayesian Networks Using Full Densities Oliver C. Schrempf Institute of Computer Design and Fault Tolerance Universität Karlsruhe e-mail: schrempf@ira.uka.de Uwe D. Hanebeck Institute
More informationCorrectness of belief propagation in Gaussian graphical models of arbitrary topology
Correctness of belief propagation in Gaussian graphical models of arbitrary topology Yair Weiss Computer Science Division UC Berkeley, 485 Soda Hall Berkeley, CA 94720-1776 Phone: 510-642-5029 yweiss@cs.berkeley.edu
More informationLecture 16 Deep Neural Generative Models
Lecture 16 Deep Neural Generative Models CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 22, 2017 Approach so far: We have considered simple models and then constructed
More informationStatistical Approaches to Learning and Discovery
Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University
More informationApproximate Inference Part 1 of 2
Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ 1 Bayesian paradigm Consistent use of probability theory
More informationImproving Search-Based Inference in Bayesian Networks
From: Proceedings of the Eleventh International FLAIRS Conference. Copyright 1998, AAAI (www.aaai.org). All rights reserved. Improving Search-Based Inference in Bayesian Networks E. Castillo, J. M. Guti~rrez
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More information13: Variational inference II
10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational
More informationDensity Propagation for Continuous Temporal Chains Generative and Discriminative Models
$ Technical Report, University of Toronto, CSRG-501, October 2004 Density Propagation for Continuous Temporal Chains Generative and Discriminative Models Cristian Sminchisescu and Allan Jepson Department
More informationPart I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS
Part I C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Probabilistic Graphical Models Graphical representation of a probabilistic model Each variable corresponds to a
More informationCOURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception
COURSE INTRODUCTION COMPUTATIONAL MODELING OF VISUAL PERCEPTION 2 The goal of this course is to provide a framework and computational tools for modeling visual inference, motivated by interesting examples
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture Notes Fall 2009 November, 2009 Byoung-Ta Zhang School of Computer Science and Engineering & Cognitive Science, Brain Science, and Bioinformatics Seoul National University
More informationApplying Bayesian networks in the game of Minesweeper
Applying Bayesian networks in the game of Minesweeper Marta Vomlelová Faculty of Mathematics and Physics Charles University in Prague http://kti.mff.cuni.cz/~marta/ Jiří Vomlel Institute of Information
More informationBayesian networks: approximate inference
Bayesian networks: approximate inference Machine Intelligence Thomas D. Nielsen September 2008 Approximative inference September 2008 1 / 25 Motivation Because of the (worst-case) intractability of exact
More informationModeling High-Dimensional Discrete Data with Multi-Layer Neural Networks
Modeling High-Dimensional Discrete Data with Multi-Layer Neural Networks Yoshua Bengio Dept. IRO Université de Montréal Montreal, Qc, Canada, H3C 3J7 bengioy@iro.umontreal.ca Samy Bengio IDIAP CP 592,
More informationBayesian network modeling. 1
Bayesian network modeling http://springuniversity.bc3research.org/ 1 Probabilistic vs. deterministic modeling approaches Probabilistic Explanatory power (e.g., r 2 ) Explanation why Based on inductive
More informationBayesian belief networks. Inference.
Lecture 13 Bayesian belief networks. Inference. Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Midterm exam Monday, March 17, 2003 In class Closed book Material covered by Wednesday, March 12 Last
More informationThe computational power of dynamic bayesian networks
The computational power of dynamic bayesian networks Joshua Brulé Department of Computer Science University of Maryland, College Park jbrule@cs.umd.edu Abstract. This paper considers the computational
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 9 Undirected Models CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due next Wednesday (Nov 4) in class Start early!!! Project milestones due Monday (Nov 9)
More informationMachine Learning for Data Science (CS4786) Lecture 24
Machine Learning for Data Science (CS4786) Lecture 24 Graphical Models: Approximate Inference Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ BELIEF PROPAGATION OR MESSAGE PASSING Each
More informationBayesian networks. Chapter 14, Sections 1 4
Bayesian networks Chapter 14, Sections 1 4 Artificial Intelligence, spring 2013, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 14, Sections 1 4 1 Bayesian networks
More informationGraphical models. Sunita Sarawagi IIT Bombay
1 Graphical models Sunita Sarawagi IIT Bombay http://www.cse.iitb.ac.in/~sunita 2 Probabilistic modeling Given: several variables: x 1,... x n, n is large. Task: build a joint distribution function Pr(x
More informationB C D E E E" D C B D" C" A A" A A A A
1 MERL { A MITSUBISHI ELECTRIC RESEARCH LABORATORY http://www.merl.com Correctness of belief propagation in Gaussian graphical models of arbitrary topology Yair Weiss William T. Freeman y TR-99-38 October
More informationBayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling
12735: Urban Systems Modeling Lec. 09 Bayesian Networks instructor: Matteo Pozzi x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 1 outline example of applications how to shape a problem as a BN complexity of the inference
More informationArtificial Intelligence
ICS461 Fall 2010 Nancy E. Reed nreed@hawaii.edu 1 Lecture #14B Outline Inference in Bayesian Networks Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic
More informationExpectation Propagation Algorithm
Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,
More informationCS281A/Stat241A Lecture 19
CS281A/Stat241A Lecture 19 p. 1/4 CS281A/Stat241A Lecture 19 Junction Tree Algorithm Peter Bartlett CS281A/Stat241A Lecture 19 p. 2/4 Announcements My office hours: Tuesday Nov 3 (today), 1-2pm, in 723
More informationInduction of Decision Trees
Induction of Decision Trees Peter Waiganjo Wagacha This notes are for ICS320 Foundations of Learning and Adaptive Systems Institute of Computer Science University of Nairobi PO Box 30197, 00200 Nairobi.
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationExpectation Propagation for Approximate Bayesian Inference
Expectation Propagation for Approximate Bayesian Inference José Miguel Hernández Lobato Universidad Autónoma de Madrid, Computer Science Department February 5, 2007 1/ 24 Bayesian Inference Inference Given
More informationMaximum Weight Matching via Max-Product Belief Propagation
Maximum Weight Matching via Max-Product Belief Propagation Mohsen Bayati Department of EE Stanford University Stanford, CA 94305 Email: bayati@stanford.edu Devavrat Shah Departments of EECS & ESD MIT Boston,
More informationBayesian Networks Inference with Probabilistic Graphical Models
4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning
More informationA graph contains a set of nodes (vertices) connected by links (edges or arcs)
BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,
More informationBayesian Methods in Artificial Intelligence
WDS'10 Proceedings of Contributed Papers, Part I, 25 30, 2010. ISBN 978-80-7378-139-2 MATFYZPRESS Bayesian Methods in Artificial Intelligence M. Kukačka Charles University, Faculty of Mathematics and Physics,
More informationCS Lecture 4. Markov Random Fields
CS 6347 Lecture 4 Markov Random Fields Recap Announcements First homework is available on elearning Reminder: Office hours Tuesday from 10am-11am Last Time Bayesian networks Today Markov random fields
More informationProbabilistic Graphical Models for Image Analysis - Lecture 1
Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.
More informationProbabilistic Graphical Models: Representation and Inference
Probabilistic Graphical Models: Representation and Inference Aaron C. Courville Université de Montréal Note: Material for the slides is taken directly from a presentation prepared by Andrew Moore 1 Overview
More informationInference in Graphical Models Variable Elimination and Message Passing Algorithm
Inference in Graphical Models Variable Elimination and Message Passing lgorithm Le Song Machine Learning II: dvanced Topics SE 8803ML, Spring 2012 onditional Independence ssumptions Local Markov ssumption
More informationVariational Inference (11/04/13)
STA561: Probabilistic machine learning Variational Inference (11/04/13) Lecturer: Barbara Engelhardt Scribes: Matt Dickenson, Alireza Samany, Tracy Schifeling 1 Introduction In this lecture we will further
More information14 : Theory of Variational Inference: Inner and Outer Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2014 14 : Theory of Variational Inference: Inner and Outer Approximation Lecturer: Eric P. Xing Scribes: Yu-Hsin Kuo, Amos Ng 1 Introduction Last lecture
More informationFactor Graphs and Message Passing Algorithms Part 1: Introduction
Factor Graphs and Message Passing Algorithms Part 1: Introduction Hans-Andrea Loeliger December 2007 1 The Two Basic Problems 1. Marginalization: Compute f k (x k ) f(x 1,..., x n ) x 1,..., x n except
More informationDecisiveness in Loopy Propagation
Decisiveness in Loopy Propagation Janneke H. Bolt and Linda C. van der Gaag Department of Information and Computing Sciences, Utrecht University P.O. Box 80.089, 3508 TB Utrecht, The Netherlands Abstract.
More informationHuman-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg
Temporal Reasoning Kai Arras, University of Freiburg 1 Temporal Reasoning Contents Introduction Temporal Reasoning Hidden Markov Models Linear Dynamical Systems (LDS) Kalman Filter 2 Temporal Reasoning
More information13 : Variational Inference: Loopy Belief Propagation and Mean Field
10-708: Probabilistic Graphical Models 10-708, Spring 2012 13 : Variational Inference: Loopy Belief Propagation and Mean Field Lecturer: Eric P. Xing Scribes: Peter Schulam and William Wang 1 Introduction
More information12 : Variational Inference I
10-708: Probabilistic Graphical Models, Spring 2015 12 : Variational Inference I Lecturer: Eric P. Xing Scribes: Fattaneh Jabbari, Eric Lei, Evan Shapiro 1 Introduction Probabilistic inference is one of
More informationIntelligent Systems:
Intelligent Systems: Undirected Graphical models (Factor Graphs) (2 lectures) Carsten Rother 15/01/2015 Intelligent Systems: Probabilistic Inference in DGM and UGM Roadmap for next two lectures Definition
More informationInference and Representation
Inference and Representation David Sontag New York University Lecture 5, Sept. 30, 2014 David Sontag (NYU) Inference and Representation Lecture 5, Sept. 30, 2014 1 / 16 Today s lecture 1 Running-time of
More informationPart 1: Expectation Propagation
Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud
More informationChapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang
Chapter 4 Dynamic Bayesian Networks 2016 Fall Jin Gu, Michael Zhang Reviews: BN Representation Basic steps for BN representations Define variables Define the preliminary relations between variables Check
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 10 Alternatives to Monte Carlo Computation Since about 1990, Markov chain Monte Carlo has been the dominant
More informationSymbolic Variable Elimination in Discrete and Continuous Graphical Models. Scott Sanner Ehsan Abbasnejad
Symbolic Variable Elimination in Discrete and Continuous Graphical Models Scott Sanner Ehsan Abbasnejad Inference for Dynamic Tracking No one previously did this inference exactly in closed-form! Exact
More informationNaïve Bayes classification
Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss
More informationMultivariate Bayesian Linear Regression MLAI Lecture 11
Multivariate Bayesian Linear Regression MLAI Lecture 11 Neil D. Lawrence Department of Computer Science Sheffield University 21st October 2012 Outline Univariate Bayesian Linear Regression Multivariate
More informationIntroduction To Graphical Models
Peter Gehler Introduction to Graphical Models Introduction To Graphical Models Peter V. Gehler Max Planck Institute for Intelligent Systems, Tübingen, Germany ENS/INRIA Summer School, Paris, July 2013
More informationReadings: K&F: 16.3, 16.4, Graphical Models Carlos Guestrin Carnegie Mellon University October 6 th, 2008
Readings: K&F: 16.3, 16.4, 17.3 Bayesian Param. Learning Bayesian Structure Learning Graphical Models 10708 Carlos Guestrin Carnegie Mellon University October 6 th, 2008 10-708 Carlos Guestrin 2006-2008
More informationOutline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination
Probabilistic Graphical Models COMP 790-90 Seminar Spring 2011 The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Outline It Introduction ti Representation Bayesian network Conditional Independence Inference:
More informationT Machine Learning: Basic Principles
Machine Learning: Basic Principles Bayesian Networks Laboratory of Computer and Information Science (CIS) Department of Computer Science and Engineering Helsinki University of Technology (TKK) Autumn 2007
More informationLecture 9: PGM Learning
13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and
More information