COMPSCI 276 Fall 2007

Size: px
Start display at page:

Download "COMPSCI 276 Fall 2007"

Transcription

1 Exact Inference lgorithms for Probabilistic Reasoning; OMPSI 276 Fall

2 elief Updating Smoking lung ancer ronchitis X-ray Dyspnoea P lung cancer=yes smoking=no, dyspnoea=yes =? 2

3 Probabilistic Inference Tasks elief updating: ELXi = PXi = x i evidence Finding most probable explanation MPE x* = argmaxpx, e Finding maximum a-posteriory hypothesis * * a1,...,ak = argmax Px, e a X/ Finding maximum-expected-utility MEU decision * * d1,..., dk = arg max Px, eux d X/D x X : hypothesis variables D X : decision variables U x : utility function 3

4 elief updating is NP-hard Each sat formula can be mapped to a bayesian network query. Example: u,~v,w and ~u,~w,y sat? 4

5 Motivation Given a chain show how it works How can we compute PD? PD =0? P D=0? D rute force Ok^4 5

6 elief updating: PX evidence=? Pa e=0 Pa,e=0= D E PaPb apc apd b,ape b,c= e = 0, d, c, b Moral graph Pa e=0 d Pc a c b Pb apd b,ape b,c Variable Elimination h a, d, c, e 6

7 7

8 8

9 9

10 10

11 ucket elimination lgorithm elim-bel Dechter 1996 b bucket : bucket : bucket D: bucket E: bucket : Pb a Pd b,a Pe b,c Pc a e=0 Pa h Pa e=0 a, h a, h D h E d, a, d, a e e c, e Elimination operator W*=4 induced width max clique size D E 11

12 12

13 E D D E 13

14 omplexity of elimination w * d O n exp w * d the induced width of moral graph along ordering d The effect of the ordering: E D D E D E Moral graph * * w d 1 = 4 w d 2 = 2 14

15 Finding small induced-width NP-complete tree has induced-width of? Greedy algorithms: Min width Min induced-width Max-cardinality Fill-in thought as the best See anytime min-width Gogate and Dechter 15

16 Different Induced graphs F E E D D D E F E D F F a b c d 16

17 17

18 The impact of observations G G G G D D D D F F F F a b c d 18

19 Use the ancestral graph only 19

20 From ucket elimination to bucket-tree elimination ucket G: PG F ucket F: PF, F λ G F ucket D: PD, ucket : P ucket : P ucket : P λ G λ F, λ,, D λ λ F PF, P P λ F, λ F λ G F, G PG F D PD, λ D λ, P T 20

21 21

22 22

23 23

24 ucket-tree propagation x 1 x 2 u hu,v v x n cluster 2 u =ψ u { h x1, u, h x, u,..., h xn, u, h v, u} ompute the message : h u, v = elim u, v f cluster u { h v, u} f 24

25 25

26 TE: full Execution ucket G: PG F ucket F: PF, ucket D: PD, ucket : P ucket : P F λ G λ F ucket : P λ F, λ,, D λ G F F D F,,, λ F F PF,, P λ P F Π,, Π F λ G F Π G Π F F λ D, G PG F D PD,, P λ D Π, 26

27 From buckets to superbucket to clusters G,F G,F G,F F F F F,, D,, F,, D,,,,,D,F,,,,,,,,, D F G 27

28 luster Tree Decomposition and elimination TE cluster-tree decomposition is a set of subsets of variables clusters connected by a tree structure: 1. Every function PT has at least one cluster that contains its scope. The function is assigned to one such cluster. 2. The cluster-tree obeys the running intersection property. Proposition: If T is a cluster-tree decomposition, then any tree obtained by merging adjacent clusters the variable set and the functions is also a tree-decomposition. Join-Tree: a tree-decomposition where all clusters are maximal. 28

29 Tree Decomposition for belief updating pa pb a pe b, f pc a, b D Pd b E pf c, d F pg e, f G 29

30 Tree Decomposition pa pa, pb a, pc a,b pc a, b pb a D Pd b E pe b, f D F pd b, pf c,d F E F pe b,f pf c, d F EF pg e, f G E F G pg e,f 30

31 Tree decompositions more formal tree decomposition for a belief network N =< X,D,G,P > is a triple < T, χ, ψ >, where T = V,E is a tree andχ andψ are labeling functions, associating with each vertex v V two sets, χv X and ψv P satisfying : 1. For each function pi P there is exactly one vertex such that pi ψv and scopepi χv 2. For each variable X X the set {v V X χv} forms a i connected subtree running intersection property i pa, pb a, pc a,b D F pd b, pf c,d F E F pe b,f EF elief network F D E G E F G pg e,f Tree decomposition 31

32 Same Message Passing x 1 x 2 u hu,v v x n cluster 2 u =ψ u { h x1, u, h x, u,..., h xn, u, h v, u} ompute the message : h u, v = elim u, v f cluster u { h v, u} f 32

33 luster Tree Elimination = h 1,2 b, c p a p b a p c a, b a 1 pa, pb a, pc a,b 2 D D F D F 2pd b, pd b, h h 1,2 1,2 b,c, pf c,d pf c,d 1 h 2,3 b, f = p d b h1,2 b, c p f c, d c, d F sep2,3 = {,F} elim2,3 = {,D} D E 3 E F pe b,f EF F G 4 E F G pg e,f 33

34 ,, 1,2 b a c p a b p a p c b h a =,,, 3,2, 2,1 f b h d c f p b d p c b h f d = 2 1 DF TE: luster Tree Elimination 34,,, 1,2, 2,3 c b h d c f p b d p f b h d c =,,, 4,3 3,2 f e h f b e p f b h e =,,, 2,3 3,4 f b h f b e p f e h b =,, 4,3 f e g G p f e h e = = G E F D 4 3 EF EFG EF F Time: O expw+1 Space: O expsep For each cluster PX e is computed

35 Tree-Width & Separator 35

36 Finding tree-decompositions Good Join-trees using triangulation Tree-width can be generated using induced-width ordering heuristics 36

37 TE - properties orrectness and completeness: lgorithm TE is correct, i.e. it computes the exact joint probability of a single variable and the evidence. Time complexity: O deg n+n d w*+1 Space complexity: O N d sep where deg = the maximum degree of a node n = number of variables = number of PTs N = number of nodes in the tree decomposition d = the maximum domain size of a variable w* = the induced width sep = the separator size 37

38 Inference on trees is easy and distributed m XY X m YX X PY X m YT Y m TY Y PX m XZ X m ZX X Z = Z = m ZX X = P Z X m MZ Z m PZ X Z m YR Y Z m ZM Z m X = P X m X m RY Y m ZL Z Z m LZ m MZ PT Y PR Y PL Z PM Z m m m m MZ LZ XZ ZL ZM Z = Z = M L X X P M Z P L Z YX P Z X m XZ P Z X m XZ LZ X m X m Z MZ LZ Z Z elief updating = sum-prod Inference is time and space linear on trees 38

39 Pearl s elief Propagation polytree: a tree with Larger families polytree decomposition λ Z u = 1 P z 1 1 u 1 Z 1 Z 2 Z 3 λ Z 2 u 2 λ Z3 u 3 U 1 U 2 U 3 u1 Pu2 Pu3 X 1 Pz1 u1 Pz2 u2 Pz3 u3 Y 1 PX1,u1,1,u2,u3 Running TE = running Pearl s P over the dual graph Dual-graph: nodes are cpts, arcs connect non-empty intersections. Time and space linear propagatin Py1 x1 39

40 Iterative elief Propagation IP Loopy belief propagation elief propagation is exact for poly-trees IP - applying P iteratively to cyclic networks One step: update ELU 1 λ X 1 u 1 U 1 π U2 x 1 U 2 U 3 π U3 x 1 X 1 λ X u 2 1 X 2 No guarantees for convergence Works well for many coding networks 40

41 41

42 Finding MPE lgorithm elim-mpe Dechter 1996 MPE = max Px max b Elimination operator bucket : Pb a Pd b,a Pe b,c bucket : bucket D: bucket E: bucket : = is replaced by max : max P a P c a P b a P d a, e, d, c, b Pc a e=0 Pa h h MPE a, h D h E a, d, a, a d, e e c, x e a, b P e W*=4 induced width max clique size b, c D E 42

43 Generating the MPE-tuple 5. b' = arg max b Pd' b, a' Pb a' Pe' b, c' : Pb a Pd b,a Pe b,c 4. c' = arg max Pc h c a',d',c, e' a' : Pc a h a, d, c, e 3. d' = arg max h d a',d,e' D: h a,d,e 2. e' = 0 E: e=0 h D a, e 1. a' = arg max a Pa h E a : Pa h E a Return a', b',c',d',e' 43

44 44

45 45

46 46

47 47

48 onditioning generates the probability tree = = = 0,, 0, e b c b c b e P b a d P a c P a b P a P e a P 48 omplexity of conditioning: exponential time, linear space

49 onditioning+elimination = = = 0,, 0, e d c b c b e P b a d P a c P a b P a P e a P 49 Idea: conditioning until of a subproblem gets small w *

50 50

51 Loop-cutset decomposition You condition until you get a polytree =0 =1 =0 =0 =0 =1 =1 =1 F F F P F=0 = P, =0 F=0+P,=1 F=0 Loop-cutset method is time exp in loop-cutset size nd linear space 51

52 W-cutset algorithms Elim-cond-bel: Identify a w-cutset, _w, of the network For each assignment to the cutset solve by TE the conditioned subproblems ggregate the solutions over all cutset assignments. Time complexity: exp _w +w Space complexity: expw 52

53 ll algorithms genaralize to any graphical models Through general operations of combination and marginalization General E, TE, TE, P pplicable to Markov networks, to constraint optimization, to counting number of solutions in a ST formula, etc. 53

54 Tree-solving elief updating sum-prod SP consistency projection-join m XY X m YX X PX m XZ X m ZX X PY X m YT Y m YR Y m TY Y m RY Y PZ X m ZL Z m ZM Z m LZ Z m MZ Z PT Y PR Y PL Z PM Z MPE max-prod #SP sum-prod Inference is time and space linear on trees 54

55 Graphical Models graphical model X,D,F: X = {X 1, X n } variables D = {D 1, D n } domains F = {f 1,,f r } Operators: functions constraints, PTS, NFs combination elimination projection onditional Probability Table PT F PF, E D Relation F red green blue blue red red blue blue green green red blue F f i : = F = + Primal graph interaction graph Tasks: elief updating: Σ X-y j P i MPE: max X j P j SP: X j j Max-SP: min X Σ j F j ll these tasks are NP-hard exploit problem structure identify special cases approximate 55

56 56

57 57

58 luster-tree Decomposition Let P=<X,D,F,,,{Z i }> be an automated reasoning problem. tree decomposition is <T,χ,ψ>, such that T=V,E is a tree χχ associates a set of variables χv X with each node ψ associates a set of functions ψv F with each node such that f i F, there is exactly one v such that f i ψv and scopef i χv. x X, the set {v V x χv} induces a connected subtree. i Z i χv for some v V. 58

59 Example: luster-tree 6 6 χ6 = {1,5,6} ψ 6 = { f 1 5,6, f 2 1,6} ij = X i X j 5 4 χ5 = {1,2,5} ψ 5 = { 2,5} f χ3 = {2,3} ψ 3 = { 2,3} f χ2 = {1,2} ψ 2 = { 1,2} f χ4 = {1,4} ψ 4 = { 1,4} f χ 1 = {1} ψ 1 = Ø 1 a b c Tree-width = 3 sep5,6 = {1, 5} 59

60 luster-tree Elimination TE m u,w = sepu,w [ j {m vj,u} ψu] χu ψu m v1,u... m v2,u m vi,u 60

61 Example: luster-tree Elimination 6 χ 6 = {1,5,6} ψ 6 = { f 1 5,6, f 2 1,6} χ 5 = {1,2,5} χ 3 = {2,3} ψ 5 = { 2,5} 5 3 ψ 3 = { 2,3} f 3 f 5 m m m 6,5 5,6 5,2 {1,5,6} { f 1, f 2 } 1,5 : = min 6{ f15, f 1,6} 1,5 : = min { f 2 + m 3 2 2,5 1,2 : = min { f 5 + m 3 6,5 2,5 + 1,2} {1,2,5} { f 3 } 2,5 + 1,5} 5 {2,3} { f 5 } f 5 3 m 3,2 2 : = min 3{ f52,3} χ2 = {1,2} ψ 2 = { 1,2} f 6 χ 1 = {1} ψ 1 = Ø χ4 = {1,4} ψ 4 = { 1,4} f 4 m 2,1 m2,5 1,2 : = f61,2 + + m 2 + m 1 + m 3,2 1 : = min { f 5,2 2 1,2 + m 6 1,2 1,2 + 3,2 2} m 1,2 1 : = m4,1 1 {1,2} { f 6 {1} } 2 1 m 2,3 2 : = min { f + m 5,2 1,2 + m {1,4} { f 4 1 } 6 1,2 + 1,2 4 1} m 4,1 1 : = min 4{ f41,4} m 1,4 1 : = m2,1 1 61

Exact Inference Algorithms Bucket-elimination

Exact Inference Algorithms Bucket-elimination Exact Inference Algorithms Bucket-elimination COMPSCI 276, Spring 2011 Class 5: Rina Dechter (Reading: class notes chapter 4, Darwiche chapter 6) 1 Belief Updating Smoking lung Cancer Bronchitis X-ray

More information

Exact Inference Algorithms Bucket-elimination

Exact Inference Algorithms Bucket-elimination Exact Inference Algorithms Bucket-elimination COMPSCI 276, Spring 2013 Class 5: Rina Dechter (Reading: class notes chapter 4, Darwiche chapter 6) 1 Belief Updating Smoking lung Cancer Bronchitis X-ray

More information

Inference in Graphical Models Variable Elimination and Message Passing Algorithm

Inference in Graphical Models Variable Elimination and Message Passing Algorithm Inference in Graphical Models Variable Elimination and Message Passing lgorithm Le Song Machine Learning II: dvanced Topics SE 8803ML, Spring 2012 onditional Independence ssumptions Local Markov ssumption

More information

Reasoning with Deterministic and Probabilistic graphical models Class 2: Inference in Constraint Networks Rina Dechter

Reasoning with Deterministic and Probabilistic graphical models Class 2: Inference in Constraint Networks Rina Dechter Reasoning with Deterministic and Probabilistic graphical models Class 2: Inference in Constraint Networks Rina Dechter Road Map n Graphical models n Constraint networks Model n Inference n Search n Probabilistic

More information

Bayesian Networks Representation and Reasoning

Bayesian Networks Representation and Reasoning ayesian Networks Representation and Reasoning Marco F. Ramoni hildren s Hospital Informatics Program Harvard Medical School (2003) Harvard-MIT ivision of Health Sciences and Technology HST.951J: Medical

More information

Bayesian Networks: Representation, Variable Elimination

Bayesian Networks: Representation, Variable Elimination Bayesian Networks: Representation, Variable Elimination CS 6375: Machine Learning Class Notes Instructor: Vibhav Gogate The University of Texas at Dallas We can view a Bayesian network as a compact representation

More information

Exact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = =

Exact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = = Exact Inference I Mark Peot In this lecture we will look at issues associated with exact inference 10 Queries The objective of probabilistic inference is to compute a joint distribution of a set of query

More information

Bounded Inference Non-iteratively; Mini-Bucket Elimination

Bounded Inference Non-iteratively; Mini-Bucket Elimination Bounded Inference Non-iteratively; Mini-Bucket Elimination COMPSCI 276,Spring 2018 Class 6: Rina Dechter Reading: Class Notes (8, Darwiche chapters 14 Outline Mini-bucket elimination Weighted Mini-bucket

More information

CS281A/Stat241A Lecture 19

CS281A/Stat241A Lecture 19 CS281A/Stat241A Lecture 19 p. 1/4 CS281A/Stat241A Lecture 19 Junction Tree Algorithm Peter Bartlett CS281A/Stat241A Lecture 19 p. 2/4 Announcements My office hours: Tuesday Nov 3 (today), 1-2pm, in 723

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 16 Undirected Graphs Undirected Separation Inferring Marginals & Conditionals Moralization Junction Trees Triangulation Undirected Graphs Separation

More information

Variable Elimination: Algorithm

Variable Elimination: Algorithm Variable Elimination: Algorithm Sargur srihari@cedar.buffalo.edu 1 Topics 1. Types of Inference Algorithms 2. Variable Elimination: the Basic ideas 3. Variable Elimination Sum-Product VE Algorithm Sum-Product

More information

Implementing Machine Reasoning using Bayesian Network in Big Data Analytics

Implementing Machine Reasoning using Bayesian Network in Big Data Analytics Implementing Machine Reasoning using Bayesian Network in Big Data Analytics Steve Cheng, Ph.D. Guest Speaker for EECS 6893 Big Data Analytics Columbia University October 26, 2017 Outline Introduction Probability

More information

COMP538: Introduction to Bayesian Networks

COMP538: Introduction to Bayesian Networks COMP538: Introduction to ayesian Networks Lecture 4: Inference in ayesian Networks: The VE lgorithm Nevin L. Zhang lzhang@cse.ust.hk Department of Computer Science and Engineering Hong Kong University

More information

Inference as Optimization

Inference as Optimization Inference as Optimization Sargur Srihari srihari@cedar.buffalo.edu 1 Topics in Inference as Optimization Overview Exact Inference revisited The Energy Functional Optimizing the Energy Functional 2 Exact

More information

Weighted Heuristic Anytime Search

Weighted Heuristic Anytime Search Weighted Heuristic Anytime Search Flerova, Marinescu, and Dechter 1 1 University of South Carolina April 24, 2017 Basic Heuristic Best-first Search Blindly follows the heuristic Weighted A Search For w

More information

Lecture 17: May 29, 2002

Lecture 17: May 29, 2002 EE596 Pat. Recog. II: Introduction to Graphical Models University of Washington Spring 2000 Dept. of Electrical Engineering Lecture 17: May 29, 2002 Lecturer: Jeff ilmes Scribe: Kurt Partridge, Salvador

More information

Variable Elimination (VE) Barak Sternberg

Variable Elimination (VE) Barak Sternberg Variable Elimination (VE) Barak Sternberg Basic Ideas in VE Example 1: Let G be a Chain Bayesian Graph: X 1 X 2 X n 1 X n How would one compute P X n = k? Using the CPDs: P X 2 = x = x Val X1 P X 1 = x

More information

Probabilistic Graphical Models (I)

Probabilistic Graphical Models (I) Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random

More information

Statistical Approaches to Learning and Discovery

Statistical Approaches to Learning and Discovery Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University

More information

Chapter 8 Cluster Graph & Belief Propagation. Probabilistic Graphical Models 2016 Fall

Chapter 8 Cluster Graph & Belief Propagation. Probabilistic Graphical Models 2016 Fall Chapter 8 Cluster Graph & elief ropagation robabilistic Graphical Models 2016 Fall Outlines Variable Elimination 消元法 imple case: linear chain ayesian networks VE in complex graphs Inferences in HMMs and

More information

Lecture 9: PGM Learning

Lecture 9: PGM Learning 13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and

More information

Machine Learning Lecture 14

Machine Learning Lecture 14 Many slides adapted from B. Schiele, S. Roth, Z. Gharahmani Machine Learning Lecture 14 Undirected Graphical Models & Inference 23.06.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de

More information

CS 188: Artificial Intelligence. Bayes Nets

CS 188: Artificial Intelligence. Bayes Nets CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew

More information

Inference in Bayesian Networks

Inference in Bayesian Networks Andrea Passerini passerini@disi.unitn.it Machine Learning Inference in graphical models Description Assume we have evidence e on the state of a subset of variables E in the model (i.e. Bayesian Network)

More information

Variable Elimination: Algorithm

Variable Elimination: Algorithm Variable Elimination: Algorithm Sargur srihari@cedar.buffalo.edu 1 Topics 1. Types of Inference Algorithms 2. Variable Elimination: the Basic ideas 3. Variable Elimination Sum-Product VE Algorithm Sum-Product

More information

4 : Exact Inference: Variable Elimination

4 : Exact Inference: Variable Elimination 10-708: Probabilistic Graphical Models 10-708, Spring 2014 4 : Exact Inference: Variable Elimination Lecturer: Eric P. ing Scribes: Soumya Batra, Pradeep Dasigi, Manzil Zaheer 1 Probabilistic Inference

More information

Sampling Algorithms for Probabilistic Graphical models

Sampling Algorithms for Probabilistic Graphical models Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir

More information

Bayesian Networks. Exact Inference by Variable Elimination. Emma Rollon and Javier Larrosa Q

Bayesian Networks. Exact Inference by Variable Elimination. Emma Rollon and Javier Larrosa Q Bayesian Networks Exact Inference by Variable Elimination Emma Rollon and Javier Larrosa Q1-2015-2016 Emma Rollon and Javier Larrosa Bayesian Networks Q1-2015-2016 1 / 25 Recall the most usual queries

More information

Message Passing and Junction Tree Algorithms. Kayhan Batmanghelich

Message Passing and Junction Tree Algorithms. Kayhan Batmanghelich Message Passing and Junction Tree Algorithms Kayhan Batmanghelich 1 Review 2 Review 3 Great Ideas in ML: Message Passing Each soldier receives reports from all branches of tree 3 here 7 here 1 of me 11

More information

Inference and Representation

Inference and Representation Inference and Representation David Sontag New York University Lecture 5, Sept. 30, 2014 David Sontag (NYU) Inference and Representation Lecture 5, Sept. 30, 2014 1 / 16 Today s lecture 1 Running-time of

More information

Junction Tree, BP and Variational Methods

Junction Tree, BP and Variational Methods Junction Tree, BP and Variational Methods Adrian Weller MLSALT4 Lecture Feb 21, 2018 With thanks to David Sontag (MIT) and Tony Jebara (Columbia) for use of many slides and illustrations For more information,

More information

Bayes Networks 6.872/HST.950

Bayes Networks 6.872/HST.950 Bayes Networks 6.872/HST.950 What Probabilistic Models Should We Use? Full joint distribution Completely expressive Hugely data-hungry Exponential computational complexity Naive Bayes (full conditional

More information

Capturing Independence Graphically; Undirected Graphs

Capturing Independence Graphically; Undirected Graphs Capturing Independence Graphically; Undirected Graphs COMPSCI 276, Spring 2017 Set 2: Rina Dechter (Reading: Pearl chapters 3, Darwiche chapter 4) 1 Outline Graphical models: The constraint network, Probabilistic

More information

On finding minimal w-cutset problem

On finding minimal w-cutset problem On finding minimal w-cutset problem Bozhena Bidyuk and Rina Dechter Information and Computer Science University Of California Irvine Irvine, CA 92697-3425 Abstract The complexity of a reasoning task over

More information

Undirected Graphical Models

Undirected Graphical Models Undirected Graphical Models 1 Conditional Independence Graphs Let G = (V, E) be an undirected graph with vertex set V and edge set E, and let A, B, and C be subsets of vertices. We say that C separates

More information

Conditional Independence

Conditional Independence Conditional Independence Sargur Srihari srihari@cedar.buffalo.edu 1 Conditional Independence Topics 1. What is Conditional Independence? Factorization of probability distribution into marginals 2. Why

More information

Alternative Parameterizations of Markov Networks. Sargur Srihari

Alternative Parameterizations of Markov Networks. Sargur Srihari Alternative Parameterizations of Markov Networks Sargur srihari@cedar.buffalo.edu 1 Topics Three types of parameterization 1. Gibbs Parameterization 2. Factor Graphs 3. Log-linear Models with Energy functions

More information

Methods of Exact Inference. Cutset Conditioning. Cutset Conditioning Example 1. Cutset Conditioning Example 2. Cutset Conditioning Example 3

Methods of Exact Inference. Cutset Conditioning. Cutset Conditioning Example 1. Cutset Conditioning Example 2. Cutset Conditioning Example 3 Methods of xact Inference Lecture 11: xact Inference means modelling all the dependencies in the data. xact Inference his is only computationally feasible for small or sparse networks. Methods: utset onditioning

More information

Computational Complexity of Inference

Computational Complexity of Inference Computational Complexity of Inference Sargur srihari@cedar.buffalo.edu 1 Topics 1. What is Inference? 2. Complexity Classes 3. Exact Inference 1. Variable Elimination Sum-Product Algorithm 2. Factor Graphs

More information

Cutset sampling for Bayesian networks

Cutset sampling for Bayesian networks Cutset sampling for Bayesian networks Cutset sampling for Bayesian networks Bozhena Bidyuk Rina Dechter School of Information and Computer Science University Of California Irvine Irvine, CA 92697-3425

More information

Exact Inference: Clique Trees. Sargur Srihari

Exact Inference: Clique Trees. Sargur Srihari Exact Inference: Clique Trees Sargur srihari@cedar.buffalo.edu 1 Topics 1. Overview 2. Variable Elimination and Clique Trees 3. Message Passing: Sum-Product VE in a Clique Tree Clique-Tree Calibration

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Uncertainty and knowledge. Uncertainty and knowledge. Reasoning with uncertainty. Notes

Uncertainty and knowledge. Uncertainty and knowledge. Reasoning with uncertainty. Notes Approximate reasoning Uncertainty and knowledge Introduction All knowledge representation formalism and problem solving mechanisms that we have seen until now are based on the following assumptions: All

More information

CSC 412 (Lecture 4): Undirected Graphical Models

CSC 412 (Lecture 4): Undirected Graphical Models CSC 412 (Lecture 4): Undirected Graphical Models Raquel Urtasun University of Toronto Feb 2, 2016 R Urtasun (UofT) CSC 412 Feb 2, 2016 1 / 37 Today Undirected Graphical Models: Semantics of the graph:

More information

Chris Bishop s PRML Ch. 8: Graphical Models

Chris Bishop s PRML Ch. 8: Graphical Models Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular

More information

Directed Graphical Models or Bayesian Networks

Directed Graphical Models or Bayesian Networks Directed Graphical Models or Bayesian Networks Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Bayesian Networks One of the most exciting recent advancements in statistical AI Compact

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Matrix Data: Classification: Part 2 Instructor: Yizhou Sun yzsun@ccs.neu.edu September 21, 2014 Methods to Learn Matrix Data Set Data Sequence Data Time Series Graph & Network

More information

Graphical Models and Kernel Methods

Graphical Models and Kernel Methods Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft

More information

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2016/2017 Lesson 13 24 march 2017 Reasoning with Bayesian Networks Naïve Bayesian Systems...2 Example

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Variational algorithms for marginal MAP

Variational algorithms for marginal MAP Variational algorithms for marginal MAP Alexander Ihler UC Irvine CIOG Workshop November 2011 Variational algorithms for marginal MAP Alexander Ihler UC Irvine CIOG Workshop November 2011 Work with Qiang

More information

Probabilistic Graphical Models Homework 2: Due February 24, 2014 at 4 pm

Probabilistic Graphical Models Homework 2: Due February 24, 2014 at 4 pm Probabilistic Graphical Models 10-708 Homework 2: Due February 24, 2014 at 4 pm Directions. This homework assignment covers the material presented in Lectures 4-8. You must complete all four problems to

More information

Graph structure in polynomial systems: chordal networks

Graph structure in polynomial systems: chordal networks Graph structure in polynomial systems: chordal networks Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute of Technology

More information

Part I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS

Part I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Part I C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Probabilistic Graphical Models Graphical representation of a probabilistic model Each variable corresponds to a

More information

Graphical Model Inference with Perfect Graphs

Graphical Model Inference with Perfect Graphs Graphical Model Inference with Perfect Graphs Tony Jebara Columbia University July 25, 2013 joint work with Adrian Weller Graphical models and Markov random fields We depict a graphical model G as a bipartite

More information

12 : Variational Inference I

12 : Variational Inference I 10-708: Probabilistic Graphical Models, Spring 2015 12 : Variational Inference I Lecturer: Eric P. Xing Scribes: Fattaneh Jabbari, Eric Lei, Evan Shapiro 1 Introduction Probabilistic inference is one of

More information

Chordal networks of polynomial ideals

Chordal networks of polynomial ideals Chordal networks of polynomial ideals Diego Cifuentes Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute of Technology Joint work with Pablo

More information

Alternative Parameterizations of Markov Networks. Sargur Srihari

Alternative Parameterizations of Markov Networks. Sargur Srihari Alternative Parameterizations of Markov Networks Sargur srihari@cedar.buffalo.edu 1 Topics Three types of parameterization 1. Gibbs Parameterization 2. Factor Graphs 3. Log-linear Models Features (Ising,

More information

Mini-Buckets: A General Scheme for Bounded Inference

Mini-Buckets: A General Scheme for Bounded Inference Mini-Buckets: A General Scheme for Bounded Inference RINA DECHTER University of California, Irvine, Irvine, California AND IRINA RISH IBM T. J. Watson Research Center, Hawthorne, New York Abstract. This

More information

Message Passing Algorithms and Junction Tree Algorithms

Message Passing Algorithms and Junction Tree Algorithms Message Passing lgorithms and Junction Tree lgorithms Le Song Machine Learning II: dvanced Topics S 8803ML, Spring 2012 Inference in raphical Models eneral form of the inference problem P X 1,, X n Ψ(

More information

15-780: Graduate Artificial Intelligence. Bayesian networks: Construction and inference

15-780: Graduate Artificial Intelligence. Bayesian networks: Construction and inference 15-780: Graduate Artificial Intelligence ayesian networks: Construction and inference ayesian networks: Notations ayesian networks are directed acyclic graphs. Conditional probability tables (CPTs) P(Lo)

More information

Example: multivariate Gaussian Distribution

Example: multivariate Gaussian Distribution School of omputer Science Probabilistic Graphical Models Representation of undirected GM (continued) Eric Xing Lecture 3, September 16, 2009 Reading: KF-chap4 Eric Xing @ MU, 2005-2009 1 Example: multivariate

More information

Clique trees & Belief Propagation. Siamak Ravanbakhsh Winter 2018

Clique trees & Belief Propagation. Siamak Ravanbakhsh Winter 2018 Graphical Models Clique trees & Belief Propagation Siamak Ravanbakhsh Winter 2018 Learning objectives message passing on clique trees its relation to variable elimination two different forms of belief

More information

5. Sum-product algorithm

5. Sum-product algorithm Sum-product algorithm 5-1 5. Sum-product algorithm Elimination algorithm Sum-product algorithm on a line Sum-product algorithm on a tree Sum-product algorithm 5-2 Inference tasks on graphical models consider

More information

Learning MN Parameters with Approximation. Sargur Srihari

Learning MN Parameters with Approximation. Sargur Srihari Learning MN Parameters with Approximation Sargur srihari@cedar.buffalo.edu 1 Topics Iterative exact learning of MN parameters Difficulty with exact methods Approximate methods Approximate Inference Belief

More information

Graph structure in polynomial systems: chordal networks

Graph structure in polynomial systems: chordal networks Graph structure in polynomial systems: chordal networks Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute of Technology

More information

Bayesian Networks: Belief Propagation in Singly Connected Networks

Bayesian Networks: Belief Propagation in Singly Connected Networks Bayesian Networks: Belief Propagation in Singly Connected Networks Huizhen u janey.yu@cs.helsinki.fi Dept. Computer Science, Univ. of Helsinki Probabilistic Models, Spring, 2010 Huizhen u (U.H.) Bayesian

More information

13 : Variational Inference: Loopy Belief Propagation and Mean Field

13 : Variational Inference: Loopy Belief Propagation and Mean Field 10-708: Probabilistic Graphical Models 10-708, Spring 2012 13 : Variational Inference: Loopy Belief Propagation and Mean Field Lecturer: Eric P. Xing Scribes: Peter Schulam and William Wang 1 Introduction

More information

Undirected Graphical Models: Markov Random Fields

Undirected Graphical Models: Markov Random Fields Undirected Graphical Models: Markov Random Fields 40-956 Advanced Topics in AI: Probabilistic Graphical Models Sharif University of Technology Soleymani Spring 2015 Markov Random Field Structure: undirected

More information

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network. ecall from last time Lecture 3: onditional independence and graph structure onditional independencies implied by a belief network Independence maps (I-maps) Factorization theorem The Bayes ball algorithm

More information

Capturing Independence Graphically; Undirected Graphs

Capturing Independence Graphically; Undirected Graphs Capturing Independence Graphically; Undirected Graphs COMPSCI 276, Spring 2014 Set 2: Rina Dechter (Reading: Pearl chapters 3, Darwiche chapter 4) 1 Constraint Networks Example: map coloring Variables

More information

Bayesian Machine Learning

Bayesian Machine Learning Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Chapter 8&9: Classification: Part 3 Instructor: Yizhou Sun yzsun@ccs.neu.edu March 12, 2013 Midterm Report Grade Distribution 90-100 10 80-89 16 70-79 8 60-69 4

More information

Lecture 6: Graphical Models

Lecture 6: Graphical Models Lecture 6: Graphical Models Kai-Wei Chang CS @ Uniersity of Virginia kw@kwchang.net Some slides are adapted from Viek Skirmar s course on Structured Prediction 1 So far We discussed sequence labeling tasks:

More information

Generative and Discriminative Approaches to Graphical Models CMSC Topics in AI

Generative and Discriminative Approaches to Graphical Models CMSC Topics in AI Generative and Discriminative Approaches to Graphical Models CMSC 35900 Topics in AI Lecture 2 Yasemin Altun January 26, 2007 Review of Inference on Graphical Models Elimination algorithm finds single

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 3, 2016 CPSC 422, Lecture 11 Slide 1 422 big picture: Where are we? Query Planning Deterministic Logics First Order Logics Ontologies

More information

Markov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials)

Markov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials) Markov Networks l Like Bayes Nets l Graphical model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov

More information

Introduc)on to Ar)ficial Intelligence

Introduc)on to Ar)ficial Intelligence Introduc)on to Ar)ficial Intelligence Lecture 13 Approximate Inference CS/CNS/EE 154 Andreas Krause Bayesian networks! Compact representa)on of distribu)ons over large number of variables! (OQen) allows

More information

Inference in Bayes Nets

Inference in Bayes Nets Inference in Bayes Nets Bruce D Ambrosio Oregon State University and Robert Fung Institute for Decision Systems Research July 23, 1994 Summer Institute on Probablility in AI 1994 Inference 1 1 Introduction

More information

Context-specific independence Parameter learning: MLE

Context-specific independence Parameter learning: MLE Use hapter 3 of K&F as a reference for SI Reading for parameter learning: hapter 12 of K&F ontext-specific independence Parameter learning: MLE Graphical Models 10708 arlos Guestrin arnegie Mellon University

More information

6.867 Machine learning, lecture 23 (Jaakkola)

6.867 Machine learning, lecture 23 (Jaakkola) Lecture topics: Markov Random Fields Probabilistic inference Markov Random Fields We will briefly go over undirected graphical models or Markov Random Fields (MRFs) as they will be needed in the context

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 9 Undirected Models CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due next Wednesday (Nov 4) in class Start early!!! Project milestones due Monday (Nov 9)

More information

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, Networks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches

CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches A presentation by Evan Ettinger November 11, 2005 Outline Introduction Motivation and Background

More information

Bayes Nets: Independence

Bayes Nets: Independence Bayes Nets: Independence [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Bayes Nets A Bayes

More information

Lecture 8: PGM Inference

Lecture 8: PGM Inference 15 September 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I 1 Variable elimination Max-product Sum-product 2 LP Relaxations QP Relaxations 3 Marginal and MAP X1 X2 X3 X4

More information

13 : Variational Inference: Loopy Belief Propagation

13 : Variational Inference: Loopy Belief Propagation 10-708: Probabilistic Graphical Models 10-708, Spring 2014 13 : Variational Inference: Loopy Belief Propagation Lecturer: Eric P. Xing Scribes: Rajarshi Das, Zhengzhong Liu, Dishan Gupta 1 Introduction

More information

Rapid Introduction to Machine Learning/ Deep Learning

Rapid Introduction to Machine Learning/ Deep Learning Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian

More information

9 Forward-backward algorithm, sum-product on factor graphs

9 Forward-backward algorithm, sum-product on factor graphs Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous

More information

Chapter 16. Structured Probabilistic Models for Deep Learning

Chapter 16. Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 18 The Junction Tree Algorithm Collect & Distribute Algorithmic Complexity ArgMax Junction Tree Algorithm Review: Junction Tree Algorithm end message

More information

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 Plan for today and next week Today and next time: Bayesian networks (Bishop Sec. 8.1) Conditional

More information

Introduction to Probabilistic Reasoning. Image credit: NASA. Assignment

Introduction to Probabilistic Reasoning. Image credit: NASA. Assignment Introduction to Probabilistic Reasoning Brian C. Williams 16.410/16.413 November 17 th, 2010 11/17/10 copyright Brian Williams, 2005-10 1 Brian C. Williams, copyright 2000-09 Image credit: NASA. Assignment

More information

Directed and Undirected Graphical Models

Directed and Undirected Graphical Models Directed and Undirected Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Last Lecture Refresher Lecture Plan Directed

More information

Graphical Models. Andrea Passerini Statistical relational learning. Graphical Models

Graphical Models. Andrea Passerini Statistical relational learning. Graphical Models Andrea Passerini passerini@disi.unitn.it Statistical relational learning Probability distributions Bernoulli distribution Two possible values (outcomes): 1 (success), 0 (failure). Parameters: p probability

More information

Bayesian Networks Inference with Probabilistic Graphical Models

Bayesian Networks Inference with Probabilistic Graphical Models 4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning

More information

Bayesian networks. Instructor: Vincent Conitzer

Bayesian networks. Instructor: Vincent Conitzer Bayesian networks Instructor: Vincent Conitzer Rain and sprinklers example raining (X) sprinklers (Y) P(X=1) =.3 P(Y=1) =.4 grass wet (Z) Each node has a conditional probability table (CPT) P(Z=1 X=0,

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture Notes Fall 2009 November, 2009 Byoung-Ta Zhang School of Computer Science and Engineering & Cognitive Science, Brain Science, and Bioinformatics Seoul National University

More information

{ p if x = 1 1 p if x = 0

{ p if x = 1 1 p if x = 0 Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =

More information