Bayesian Networks. B. Inference / B.3 Approximate Inference / Sampling
|
|
- Solomon Harrell
- 6 years ago
- Views:
Transcription
1 Bayesian etworks B. Inference / B.3 Approximate Inference / Sampling Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL) Institute of Computer Science University of Hildesheim Course on Bayesian etworks, winter term 2016/2017 1/20
2 Bayesian etworks Mon Syllabus (1) 0. Introduction A. Fundamentals & A.1. Probability Calculus revisited Mon (2) A.2. Separation in Graphs Mon (3) A.3. Bayesian and Markov etworks Mon (4) (ctd.) B. Inference Mon (5) B.1. Exact Inference Mon (6) B.2. Exact Inference / Join Tree Mon (7) (ctd.) Mon (8) B.3. Approximate Inference / Sampling Mon (9) B.4. Approximate Inference / Propagation C. Learning Mon (10) C.1. Parameter Learning Mon (11) C.2. Parameter Learning with Missing Values Mon (12) C.3. Structure Learning / PC Algorithm Mon (13) C.4. Structure Learning / Search Methods Course on Bayesian etworks, winter term 2016/2017 1/20
3 Bayesian etworks 1. Why exact inference may not be good enough 2. Acceptance-Rejection Sampling 3. Importance Sampling 4. Self and Adaptive Importance Sampling 5. Stochastic / Loopy Propagation Course on Bayesian etworks, winter term 2016/2017 1/20
4 Bayesian etworks / 1. Why exact inference may not be good enough bayesian network # variables time for exact inference studfarm s Hailfinder s Pathfinder s Link s 1) on a 1.6MHz Pentium-M notebook ( 1) on a 2.5 MHz Pentium-IV) though w/o optimized implementation with very simple triangulation heuristics (minimal degree). Course on Bayesian etworks, winter term 2016/2017 1/20
5 Bayesian etworks 1. Why exact inference may not be good enough 2. Acceptance-Rejection Sampling 3. Importance Sampling 4. Self and Adaptive Importance Sampling 5. Stochastic / Loopy Propagation Course on Bayesian etworks, winter term 2016/2017 2/20
6 Bayesian etworks / 2. Acceptance-Rejection Sampling Estimating marginals from data case family-out light-on bowel-problem dog-out hear-bark Figure 1: Example data for the dog-problem. family-out bowel-problem light-on dog-out hear-bark Figure 2: Bayesian network for dog-problem. family-out a) counts Figure 3: Estimating absolute probabilities (root node tables). family-out b) probabilities Course on Bayesian etworks, winter term 2016/2017 2/20
7 Bayesian etworks / 2. Acceptance-Rejection Sampling Estimating marginals from data case family-out light-on bowel-problem dog-out hear-bark Figure 1: Example data for the dog-problem. family-out bowel-problem light-on dog-out hear-bark Figure 2: Bayesian network for dog-problem. family-out 0 1 bowel dog-out a) counts family-out 0 1 bowel dog-out b) absolute probabilities Figure 4: Estimating conditional probabilities (inner node tables). family-out 0 1 bowel dog-out c) cond. probabilities Course on Bayesian etworks, winter term 2016/2017 3/20
8 Bayesian etworks / 2. Acceptance-Rejection Sampling Estimating marginals from data given evidence If we want to estimate the probabilities for family-out given the evidence that dog-out is 1, we have (i) identify all cases that are compatible with the given evidence, (ii) estimate the target potential p(familiy-out) from these cases. family-out light-on bowel-problem dog-out hear-bark case rejected rejected accepted accepted rejected rejected accepted rejected accepted accepted Figure 5: Accepted and rejected cases for evidence dog-out = 1. family-out a) counts family-out b) probabilities Figure 6: Estimating target potentials given evidence, here p(family-out dog-out = 1). Course on Bayesian etworks, winter term 2016/2017 4/20
9 Bayesian etworks / 2. Acceptance-Rejection Sampling Learning and inferencing data learn: a) parameters: estimate vertex potentials b) graph structure model infer target potentials Figure 7: Learing models from data for inferencing. Course on Bayesian etworks, winter term 2016/2017 4/20
10 Bayesian etworks / 2. Acceptance-Rejection Sampling Sampling and estimating data learn: a) parameters: estimate vertex potentials b) graph structure generate synthetic data / sample model estimate infer target potentials Figure 7: Learing models from data for inferencing vs. sampling from models and estimating. Course on Bayesian etworks, winter term 2016/2017 5/20
11 Bayesian etworks / 2. Acceptance-Rejection Sampling Given a discrete distribution, e.g., Sampling a discrete distribution Pain Weightloss Vomiting Adeno Figure 8: Example for a discrete distribution. How do we draw samples from this distribution? = generate synthetic data that is distributed according to it? Course on Bayesian etworks, winter term 2016/2017 6/20
12 Bayesian etworks / 2. Acceptance-Rejection Sampling Sampling a discrete distribution (i) Fix an enumeration of all states of the distribution p, i.e., σ : {1,..., Ω } Ω bijective with Ω the set of all states, (ii) compute the cumulative distribution function in the state index, i.e., cump,σ : {1,..., Ω } [0, X1] i 7 p(σ(j)) j i (iii) draw a random real value r uniformly from [0,1], (iv) search the state ω with cump,σ (ω) r and maximal cump,σ (ω). Pain Weightloss Vomiting Adeno Figure 8: Example for a discrete distribution. Adeno Pain Weightloss Vomiting cump,σ (i) index i Figure 9: Cumulative distribution function. Course on Bayesian etworks, winter term 2016/2017 7/20
13 Bayesian etworks / 2. Acceptance-Rejection Sampling Sampling a Bayesian etwork / naive approach As a bayesian network encodes a discrete distribution, we can use the method from the former slide to draw samples from a bayesian network: (i) Compute the full JPD table from the bayesian network, (ii) draw a sample from the table as on the slide before. This approach is not sensible though, as we actually used bayesian networks s.t. we not have to compute the full JPD (as it normally is way to large to handle). How can we make use of the independencies encoded in the bayesian network structure? Course on Bayesian etworks, winter term 2016/2017 8/20
14 Bayesian etworks / 2. Acceptance-Rejection Sampling Sampling a Bayesian etwork Idea: sample variables separately, one at a time. 1 2 sample-forward(b := (G := (V, E), (pv )v V )) : σ := topological-ordering (G) x := 0V for i = 1,..., σ do v := σ(i) q := pv x pa(v) draw xv q od return x 3 If we have sampled 4 X1,..., X k 5 6 already and Xk+1 is a vertex s.t. 7 desc(xk+1) {X1,..., Xk } = 8 then 9 p(xk+1 X1,..., Xk ) = p(xk+1 pa(xk+1)) Figure 10: Algorithm for sampling a bayesian i.e., we can sample Xk+1 from its vertex network. potential given the evidence of its parents (as sampled before). Course on Bayesian etworks, winter term 2016/2017 9/20
15 Bayesian etworks / 2. Acceptance-Rejection Sampling Sampling a Bayesian etwork / example Let σ := (F, B, L, D, H). 1. x F p F = (0.85, 0.15) say with outcome x B p B = (0.8, 0.2) say with outcome 1. family-out light-on bowel-problem dog-out 3. x L p L (F = 0) = (0.95, 0.05) say with outcome x D p D (F = 0, B = 1) = (0.03, 0.97) say with outcome x H p H (D = 1) = (0.3, 0.7) say with outcome 1. The result is x = (0, 1, 0, 1, 1) F = F 0 1 L = hear-bark B = D 0 1 H = B 0 1 F D = Figure 11: Bayesian network for dog-problem. Course on Bayesian etworks, winter term 2016/ /20
16 Bayesian etworks / 2. Acceptance-Rejection Sampling Acceptance-rejection sampling Inferencing by acceptance-rejection sampling means: (i) draw a sample from the bayesian network (w/o evidence entered), (ii) drop all data from the sample that are not conformant with the evidence, (iii) estimate target potentials from the remaining data For bayesian networks sampling is done by forward-sampling. Forward sampling is stopped as soon as an evidence variable has been instantiated that contradicts the evidence infer-acceptance-rejection(b : bayesian network, W : target domain, E : evidence, n : sample size ) : D := (sample-forward (B) i = 1,..., n) return estimate (D, W, E) sample-forward(b := (G := (V, E), (pv )v V )) : σ := topological-ordering (G) x := 0V for i = 1,..., σ do v := σ(i) q := pv x pa(v) draw xv q od return x estimate(d : data, W : target domain, E : evidence ) : D 0 := (d D d dom(e) = val(e)) return estimate (D 0, W ) estimate(d : data, W : target domain ) : q := zero-potential on W for d D do q(d)++ od q := q/ data return q Acceptance-rejection sampling for 7 bayesian networks is also called logic Figure 12: Algorithm for acceptance-rejection sampling [Hen88]. sampling. Course on Bayesian etworks, winter term 2016/ /20
17 Bayesian etworks 1. Why exact inference may not be good enough 2. Acceptance-Rejection Sampling 3. Importance Sampling 4. Self and Adaptive Importance Sampling 5. Stochastic / Loopy Propagation Course on Bayesian etworks, winter term 2016/ /20
18 Bayesian etworks / 3. Importance Sampling Acceptance rate of acceptance-rejection sampling How efficient acceptance-rejection sampling is depends on the acceptance rate. Let E be evidence. Then the acceptance rate, i.e., the fraction of samples conformant with E, is p(e) the marginal probability of the evidence. Thus, acceptance-rejection sampling performs poorly if the probability of evidence is small. In the studfarm example p(j = aa) = i.e., from 2326 sampled cases 2325 are rejected. Course on Bayesian etworks, winter term 2016/ /20
19 Bayesian etworks / 3. Importance Sampling Idea of importance sampling Idea: do not sample the evidence variables, but instantiate them to the values of the evidence. A B C Instantiating the evidence variables first, Figure 13: If C is evidential and already instanmeans, we have to sample the other tiated, say C = c, then A is dependent on C, so variables from we would have to sample A from p(a C = c). Even worse, B is dependant on C and A (dp(xk+1 X1 = x1,..., Xk = xk, separation), so we would have to sample B from E1 = e1,..., Em = em) even for a topological ordering of nonevidential variables. Problem: if there is an evidence variable that is a descendant of a nonevidential variable Xk+1 that has to be sampled, then p(b A = a, C = c). But neither of these cpdfs is known in advance. it does neither occur among its parents nor is independent from Xk+1, and it may open dependency chains to other variables! Course on Bayesian etworks, winter term 2016/ /20
20 Bayesian etworks / 3. Importance Sampling Inference from a stochastic point of view LetQ V be a set of variables and p a pdf on dom(v ). Infering the marginal on a given set of variables W V and given evidence E means to compute i.e., for all x (pe ) W (x) = = Q (pe ) W dom(w ) X Q y dom(v \W \dom(e)) X Q y dom(v ) So we can reformulate the inference problem as the problem of averaging a given random variable f (here: f := I(x,e)) over a given pdf p, i.e., to compute / estimate the mean X Ep(f ) := f (x) p(x) x dom(p) p(x, y, e) I(x,e)(y) p(y) with the indicator function Q Ix : dom(v ) {0, ( 1} 1, if y dom(x) = x y 7 0, else Theorem 1 (strong law of large numbers). Let p : Ω [0, 1] be a pdf, f : Ω R be a random variable with Ep( f ) <, and Xi f, i independently. Then n 1X Xi a.s. Ep(f ) n i=1 Proof. See, e.g., [Sha03, p. 62] Course on Bayesian etworks, winter term 2016/ /20
21 Bayesian etworks / 3. Importance Sampling Sampling from the wrong distribution Inference by sampling applies the SLL: X f (x) p(x) =:Ep(f ) x dom(p) 1 X f (x) n x p (n draws) ow let q be any other pdf with p(x) > 0 = q(x) > 0, Due to X x dom(p) f (x) p(x) = x dom(p) = dom(q) X x dom(p) f (x) p =:Eq (f ) q 1 X p(x) f (x) n x q q(x) (n draws) note: p(x) q(x) q(x) X i we can sample from q instead from p if we adjust the function values of f accordingly. The pdf q is called importance function, the function w := p/q is called score or case weight. Often we know the case weight only up to a multiplicative constant, i.e., w0 := c w p/q with unknown constant c. For a sample x1,..., xn q, we then can approximate Ep(f ) by n 1X f (xi) w(xi) Ep(f ) n i=1 n X 1 f (xi) w0(xi) P 0 i w (xi ) i=1 X p(xi) 1 X p(xi) w (xi) = c = cn 1 c n Ep(1) = c n q(x ) n q(x ) i i i i 0 Course on Bayesian etworks, winter term 2016/ /20
22 Bayesian etworks / 3. Importance Sampling Case weight Back to sampling from the true distribution pe vs. sampling from the bayesian network with pre-instantiated evidence variables (the wrong distribution) The probability for a sample x from a Bayesian network among samples con- The probability for a sample x from a Bayesian network with pre-instantiated sistent with a given evidence E is evidence variables is pv (xv x pa(v)) p(x) qe (x) = pv (xv x pa(v)) v V p(x E) = = v V \dom(e) p(e) p(e) Thus, the case weight is w(x) := p(x E) = qe (x) v dom(e) pv (xv x pa(v)) p(e) Course on Bayesian etworks, winter term 2016/ /20
23 Bayesian etworks / 3. Importance Sampling Likelihood weighting sampling Inferencing by importance sampling means: infer-likelihood-weighting(b : bayesian network, W : target domain, E : evidence, n : sample size ) : (D, w) := (sample-likelihood-weighting (B, E) i = 1,..., n) return estimate (D, w, W ) 1 2 (i) choose a sampling distribution q, 3 4 (ii) draw a weighted sample from q, 1 (iii) estimate target potentials from these sample data For bayesian networks using sampling from bayesian networks with pre-instantiated evidence variables and the case weight w(x) := pv (xv x pa(v)) v dom(e) is called likelihood [FC90, SP90] weighting sampling sample-likelihood-weighting(b := (G, (pv )v VG ), E : evidence ) : σ := topological-ordering (G \ dom(e)) x := 0VG x dom(e) := val(e) for i = 1,..., σ do v := σ(i) q := pv x pa(v) draw xv q od w(x) := pv (xv x pa(v) ) v dom(e) return (x, w(x)) estimate(d : data, w : case weight, W : target domain ) : q := zero-potential on W wtot := 0 for d D do q(d) := q(d) + w(d) wtot := wtot + w(d) od q := q/wtot return q Figure 14: Algorithm for inference by likelihood weighting sampling. Course on Bayesian etworks, winter term 2016/ /20
24 Bayesian etworks / 3. Importance Sampling Likelihood weighting sampling / example Let the evidence be D = 1. Fix σ := (F, B, L, H). 1. x F p F = (0.85, 0.15) say with outcome x B p B = (0.8, 0.2) say with outcome x L p L (F = 0) = (0.95, 0.05) say with outcome x H p H (D = 1) = (0.3, 0.7) say with outcome 1. The result is family-out light-on F = F 0 1 L = bowel-problem dog-out hear-bark B = D 0 1 H = x = (0, 1, 0, 1, 1) and the case weight w(x) = p D (D = 1 F = 0, B = 1) = 0.97 B 0 1 F D = Figure 15: Bayesian network for dog-problem. Course on Bayesian etworks, winter term 2016/ /20
25 Bayesian etworks / 3. Importance Sampling Acceptance-rejection sampling Acceptance-rejection sampling can be viewed as another instance of importance sampling. Here, the sampling distribution is q := p (i.e., the distribution without evidence entered; the target distribtion is pe!) and the case weight ( 1, if x dom(e) = val(e) w(x) := Ie(x) := 0, otherwise Course on Bayesian etworks, winter term 2016/ /20
26 Bayesian etworks / 3. Importance Sampling References [FC90] R. Fung and K. Chang. Weighting and integrating evidence for stochastic simulation in bayesian networks. In M. Henrion, R.D. Shachter, L.. Kanal, and J. F. Lemmer, editors, Uncertainty in Artificial Intelligence 5, pages orth Holland, Amsterdam, [Hen88] M. Henrion. Propagation of unvertainty by logic sampling in bayes networks. In J. F. Lemmer and L.. Kanal, editors, Uncertainty in Artificial Intelligence 2, pages orth Holland, Amsterdam, [Sha03] Jun Shao. Mathematical Statistics. Springer, [SP90] R. D. Shachter and M. Peot. Simulation approaches to general probabilistic inference on belief networks. In M. Henrion, R. D. Shachter, L.. Kanal, and J. F. Lemmer, editors, Uncertainty in Artificial Intelligence 5, pages orth Holland, Amsterdam, Course on Bayesian etworks, winter term 2016/ /20
Bayesian Networks. 7. Approximate Inference / Sampling
Bayesian Networks Bayesian Networks 7. Approximate Inference / Sampling Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL) Institute for Business Economics and Information Systems
More informationProf. Dr. Lars Schmidt-Thieme, L. B. Marinho, K. Buza Information Systems and Machine Learning Lab (ISMLL), University of Hildesheim, Germany, Course
Course on Bayesian Networks, winter term 2007 0/31 Bayesian Networks Bayesian Networks I. Bayesian Networks / 1. Probabilistic Independence and Separation in Graphs Prof. Dr. Lars Schmidt-Thieme, L. B.
More informationBayesian Networks. 10. Parameter Learning / Missing Values
Bayesian Networks Bayesian Networks 10. Parameter Learning / Missing Values Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL) Institute for Business Economics and Information Systems
More informationMachine Learning for Data Science (CS4786) Lecture 24
Machine Learning for Data Science (CS4786) Lecture 24 Graphical Models: Approximate Inference Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ BELIEF PROPAGATION OR MESSAGE PASSING Each
More informationLars Schmidt-Thieme, Information Systems and Machine Learning Lab (ISMLL), Institute BW/WI & Institute for Computer Science, University of Hildesheim
Course on Bayesian Networks, summer term 2010 0/42 Bayesian Networks Bayesian Networks 11. Structure Learning / Constrained-based Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL)
More informationBayesian networks: approximate inference
Bayesian networks: approximate inference Machine Intelligence Thomas D. Nielsen September 2008 Approximative inference September 2008 1 / 25 Motivation Because of the (worst-case) intractability of exact
More informationAnnouncements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic
CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return
More informationMachine Learning Lecture 14
Many slides adapted from B. Schiele, S. Roth, Z. Gharahmani Machine Learning Lecture 14 Undirected Graphical Models & Inference 23.06.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de
More informationArtificial Intelligence Bayesian Networks
Artificial Intelligence Bayesian Networks Stephan Dreiseitl FH Hagenberg Software Engineering & Interactive Media Stephan Dreiseitl (Hagenberg/SE/IM) Lecture 11: Bayesian Networks Artificial Intelligence
More informationBayes Networks 6.872/HST.950
Bayes Networks 6.872/HST.950 What Probabilistic Models Should We Use? Full joint distribution Completely expressive Hugely data-hungry Exponential computational complexity Naive Bayes (full conditional
More informationBayesian Approaches Data Mining Selected Technique
Bayesian Approaches Data Mining Selected Technique Henry Xiao xiao@cs.queensu.ca School of Computing Queen s University Henry Xiao CISC 873 Data Mining p. 1/17 Probabilistic Bases Review the fundamentals
More informationBayesian Networks BY: MOHAMAD ALSABBAGH
Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional
More informationProbabilistic Graphical Models (I)
Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random
More informationAnnouncements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.
Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides
More informationBayes Nets III: Inference
1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy
More informationExact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = =
Exact Inference I Mark Peot In this lecture we will look at issues associated with exact inference 10 Queries The objective of probabilistic inference is to compute a joint distribution of a set of query
More informationCS 484 Data Mining. Classification 7. Some slides are from Professor Padhraic Smyth at UC Irvine
CS 484 Data Mining Classification 7 Some slides are from Professor Padhraic Smyth at UC Irvine Bayesian Belief networks Conditional independence assumption of Naïve Bayes classifier is too strong. Allows
More informationStatistical Approaches to Learning and Discovery
Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University
More informationGraphical Models and Kernel Methods
Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.
More informationDynamic importance sampling in Bayesian networks using factorisation of probability trees
Dynamic importance sampling in Bayesian networks using factorisation of probability trees Irene Martínez Department of Languages and Computation University of Almería Almería, 4, Spain Carmelo Rodríguez
More informationCSEP 573: Artificial Intelligence
CSEP 573: Artificial Intelligence Bayesian Networks: Inference Ali Farhadi Many slides over the course adapted from either Luke Zettlemoyer, Pieter Abbeel, Dan Klein, Stuart Russell or Andrew Moore 1 Outline
More informationPROBABILISTIC REASONING SYSTEMS
PROBABILISTIC REASONING SYSTEMS In which we explain how to build reasoning systems that use network models to reason with uncertainty according to the laws of probability theory. Outline Knowledge in uncertain
More informationRapid Introduction to Machine Learning/ Deep Learning
Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian
More informationSIMULATION APPROACHES TO GENERAL PROBABILISTIC INFERENCE ON BELIEF NETWORKS
SIMULATION APPROACHES TO GENERAL PROBABILISTIC INFERENCE ON BELIEF NETWORKS Ross D. Shachter Department of Engineering-Economic Systems, Stanford University Terman Engineering Center, Stanford, CA 94305-4025
More informationBayesian Networks. Motivation
Bayesian Networks Computer Sciences 760 Spring 2014 http://pages.cs.wisc.edu/~dpage/cs760/ Motivation Assume we have five Boolean variables,,,, The joint probability is,,,, How many state configurations
More informationIntelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks
Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2016/2017 Lesson 13 24 march 2017 Reasoning with Bayesian Networks Naïve Bayesian Systems...2 Example
More informationCS 188: Artificial Intelligence. Bayes Nets
CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew
More informationA Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004
A Brief Introduction to Graphical Models Presenter: Yijuan Lu November 12,2004 References Introduction to Graphical Models, Kevin Murphy, Technical Report, May 2001 Learning in Graphical Models, Michael
More informationDirected Graphical Models or Bayesian Networks
Directed Graphical Models or Bayesian Networks Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Bayesian Networks One of the most exciting recent advancements in statistical AI Compact
More informationArtificial Intelligence
ICS461 Fall 2010 Nancy E. Reed nreed@hawaii.edu 1 Lecture #14B Outline Inference in Bayesian Networks Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic
More informationInference in Bayesian networks
Inference in Bayesian networks AIMA2e hapter 14.4 5 1 Outline Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic simulation Approximate inference
More informationProbabilistic Graphical Networks: Definitions and Basic Results
This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical
More informationLecture 8: Bayesian Networks
Lecture 8: Bayesian Networks Bayesian Networks Inference in Bayesian Networks COMP-652 and ECSE 608, Lecture 8 - January 31, 2017 1 Bayes nets P(E) E=1 E=0 0.005 0.995 E B P(B) B=1 B=0 0.01 0.99 E=0 E=1
More informationCS 343: Artificial Intelligence
CS 343: Artificial Intelligence Bayes Nets: Sampling Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.
More informationImproving Search-Based Inference in Bayesian Networks
From: Proceedings of the Eleventh International FLAIRS Conference. Copyright 1998, AAAI (www.aaai.org). All rights reserved. Improving Search-Based Inference in Bayesian Networks E. Castillo, J. M. Guti~rrez
More informationSampling Algorithms for Probabilistic Graphical models
Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir
More informationDirected Graphical Models
CS 2750: Machine Learning Directed Graphical Models Prof. Adriana Kovashka University of Pittsburgh March 28, 2017 Graphical Models If no assumption of independence is made, must estimate an exponential
More informationRecall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.
ecall from last time Lecture 3: onditional independence and graph structure onditional independencies implied by a belief network Independence maps (I-maps) Factorization theorem The Bayes ball algorithm
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September
More informationProbabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech
Probabilistic Graphical Models and Bayesian Networks Artificial Intelligence Bert Huang Virginia Tech Concept Map for Segment Probabilistic Graphical Models Probabilistic Time Series Models Particle Filters
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 3, 2016 CPSC 422, Lecture 11 Slide 1 422 big picture: Where are we? Query Planning Deterministic Logics First Order Logics Ontologies
More informationStochastic Simulation
Stochastic Simulation Idea: probabilities samples Get probabilities from samples: X count x 1 n 1. x k total. n k m X probability x 1. n 1 /m. x k n k /m If we could sample from a variable s (posterior)
More informationBayesian Networks: Belief Propagation in Singly Connected Networks
Bayesian Networks: Belief Propagation in Singly Connected Networks Huizhen u janey.yu@cs.helsinki.fi Dept. Computer Science, Univ. of Helsinki Probabilistic Models, Spring, 2010 Huizhen u (U.H.) Bayesian
More informationCS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 Plan for today and next week Today and next time: Bayesian networks (Bishop Sec. 8.1) Conditional
More informationDecision-Theoretic Specification of Credal Networks: A Unified Language for Uncertain Modeling with Sets of Bayesian Networks
Decision-Theoretic Specification of Credal Networks: A Unified Language for Uncertain Modeling with Sets of Bayesian Networks Alessandro Antonucci Marco Zaffalon IDSIA, Istituto Dalle Molle di Studi sull
More informationQuantifying Uncertainty
Quantifying Uncertainty CSL 302 ARTIFICIAL INTELLIGENCE SPRING 2014 Outline Probability theory - basics Probabilistic Inference oinference by enumeration Conditional Independence Bayesian Networks Learning
More informationPart I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS
Part I C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Probabilistic Graphical Models Graphical representation of a probabilistic model Each variable corresponds to a
More informationRecall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem
Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)
More informationImplementing Machine Reasoning using Bayesian Network in Big Data Analytics
Implementing Machine Reasoning using Bayesian Network in Big Data Analytics Steve Cheng, Ph.D. Guest Speaker for EECS 6893 Big Data Analytics Columbia University October 26, 2017 Outline Introduction Probability
More informationInference in Bayesian networks
Inference in Bayesian networks hapter 14.4 5 hapter 14.4 5 1 Exact inference by enumeration Outline Approximate inference by stochastic simulation hapter 14.4 5 2 Inference tasks Simple queries: compute
More informationLecture 6: Graphical Models: Learning
Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)
More informationTópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863
Tópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863 Daniel, Edmundo, Rosa Terceiro trimestre de 2012 UFRJ - COPPE Programa de Engenharia de Sistemas e Computação Bayesian Networks
More informationECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4
ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian
More informationIntroduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013
Introduction to Bayes Nets CS 486/686: Introduction to Artificial Intelligence Fall 2013 1 Introduction Review probabilistic inference, independence and conditional independence Bayesian Networks - - What
More informationQuantifying uncertainty & Bayesian networks
Quantifying uncertainty & Bayesian networks CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2016 Soleymani Artificial Intelligence: A Modern Approach, 3 rd Edition,
More informationProbability. CS 3793/5233 Artificial Intelligence Probability 1
CS 3793/5233 Artificial Intelligence 1 Motivation Motivation Random Variables Semantics Dice Example Joint Dist. Ex. Axioms Agents don t have complete knowledge about the world. Agents need to make decisions
More informationOutline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car
CSE 573: Artificial Intelligence Autumn 2012 Bayesian Networks Dan Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer Outline Probabilistic models (and inference)
More informationProbabilistic Reasoning. (Mostly using Bayesian Networks)
Probabilistic Reasoning (Mostly using Bayesian Networks) Introduction: Why probabilistic reasoning? The world is not deterministic. (Usually because information is limited.) Ways of coping with uncertainty
More informationDirected and Undirected Graphical Models
Directed and Undirected Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Last Lecture Refresher Lecture Plan Directed
More informationMachine Learning Summer School
Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,
More informationCS 5522: Artificial Intelligence II
CS 5522: Artificial Intelligence II Bayes Nets: Independence Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]
More informationOutline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination
Probabilistic Graphical Models COMP 790-90 Seminar Spring 2011 The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Outline It Introduction ti Representation Bayesian network Conditional Independence Inference:
More informationIntroduction to Artificial Intelligence. Unit # 11
Introduction to Artificial Intelligence Unit # 11 1 Course Outline Overview of Artificial Intelligence State Space Representation Search Techniques Machine Learning Logic Probabilistic Reasoning/Bayesian
More informationTDT70: Uncertainty in Artificial Intelligence. Chapter 1 and 2
TDT70: Uncertainty in Artificial Intelligence Chapter 1 and 2 Fundamentals of probability theory The sample space is the set of possible outcomes of an experiment. A subset of a sample space is called
More informationTractable Inference in Hybrid Bayesian Networks with Deterministic Conditionals using Re-approximations
Tractable Inference in Hybrid Bayesian Networks with Deterministic Conditionals using Re-approximations Rafael Rumí, Antonio Salmerón Department of Statistics and Applied Mathematics University of Almería,
More informationInference in Graphical Models Variable Elimination and Message Passing Algorithm
Inference in Graphical Models Variable Elimination and Message Passing lgorithm Le Song Machine Learning II: dvanced Topics SE 8803ML, Spring 2012 onditional Independence ssumptions Local Markov ssumption
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationDynamic Importance Sampling in Bayesian Networks Based on Probability Trees
Dynamic Importance Sampling in Bayesian Networks Based on Probability Trees Serafín Moral Dpt. Computer Science and Artificial Intelligence University of Granada Avda. Andalucía 38, 807 Granada, Spain
More informationAn Introduction to Bayesian Machine Learning
1 An Introduction to Bayesian Machine Learning José Miguel Hernández-Lobato Department of Engineering, Cambridge University April 8, 2013 2 What is Machine Learning? The design of computational systems
More informationLecture 8: PGM Inference
15 September 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I 1 Variable elimination Max-product Sum-product 2 LP Relaxations QP Relaxations 3 Marginal and MAP X1 X2 X3 X4
More informationOutline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012
CSE 573: Artificial Intelligence Autumn 2012 Reasoning about Uncertainty & Hidden Markov Models Daniel Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 Outline
More information{ p if x = 1 1 p if x = 0
Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =
More informationLearning Bayesian Networks (part 1) Goals for the lecture
Learning Bayesian Networks (part 1) Mark Craven and David Page Computer Scices 760 Spring 2018 www.biostat.wisc.edu/~craven/cs760/ Some ohe slides in these lectures have been adapted/borrowed from materials
More informationArtificial Intelligence Bayes Nets: Independence
Artificial Intelligence Bayes Nets: Independence Instructors: David Suter and Qince Li Course Delivered @ Harbin Institute of Technology [Many slides adapted from those created by Dan Klein and Pieter
More informationBayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018
Bayesian networks CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018 Soleymani Slides have been adopted from Klein and Abdeel, CS188, UC Berkeley. Outline Probability
More informationBayesian belief networks
CS 2001 Lecture 1 Bayesian belief networks Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square 4-8845 Milos research interests Artificial Intelligence Planning, reasoning and optimization in the presence
More informationBayesian networks. Chapter 14, Sections 1 4
Bayesian networks Chapter 14, Sections 1 4 Artificial Intelligence, spring 2013, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 14, Sections 1 4 1 Bayesian networks
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationInference in Bayesian Networks
Andrea Passerini passerini@disi.unitn.it Machine Learning Inference in graphical models Description Assume we have evidence e on the state of a subset of variables E in the model (i.e. Bayesian Network)
More informationModeling and Inference with Relational Dynamic Bayesian Networks Cristina Manfredotti
Modeling and Inference with Relational Dynamic Bayesian Networks Cristina Manfredotti cristina.manfredotti@gmail.com The importance of the context Cristina Manfredotti 2 Complex activity recognition Cristina
More informationFinal Examination CS 540-2: Introduction to Artificial Intelligence
Final Examination CS 540-2: Introduction to Artificial Intelligence May 7, 2017 LAST NAME: SOLUTIONS FIRST NAME: Problem Score Max Score 1 14 2 10 3 6 4 10 5 11 6 9 7 8 9 10 8 12 12 8 Total 100 1 of 11
More informationQuantifying Uncertainty & Probabilistic Reasoning. Abdulla AlKhenji Khaled AlEmadi Mohammed AlAnsari
Quantifying Uncertainty & Probabilistic Reasoning Abdulla AlKhenji Khaled AlEmadi Mohammed AlAnsari Outline Previous Implementations What is Uncertainty? Acting Under Uncertainty Rational Decisions Basic
More informationBayes Nets: Independence
Bayes Nets: Independence [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Bayes Nets A Bayes
More informationPlanning and Optimal Control
Planning and Optimal Control 1. Markov Models Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University of Hildesheim, Germany 1 / 22 Syllabus Tue.
More informationInference in Bayesian Networks
Lecture 7 Inference in Bayesian Networks Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Slides by Stuart Russell and Peter Norvig Course Overview Introduction
More informationArtificial Intelligence Methods. Inference in Bayesian networks
Artificial Intelligence Methods Inference in Bayesian networks In which we explain how to build network models to reason under uncertainty according to the laws of probability theory. Dr. Igor rajkovski
More informationProbabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference
Probabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference David Poole University of British Columbia 1 Overview Belief Networks Variable Elimination Algorithm Parent Contexts
More informationCOS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference
COS402- Artificial Intelligence Fall 2015 Lecture 10: Bayesian Networks & Exact Inference Outline Logical inference and probabilistic inference Independence and conditional independence Bayes Nets Semantics
More informationReasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule
Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule Alan Mackworth UBC CS 322 Uncertainty 2 March 13, 2013 Textbook 6.1.3 Lecture Overview Recap: Probability & Possible World Semantics
More informationJunction Tree, BP and Variational Methods
Junction Tree, BP and Variational Methods Adrian Weller MLSALT4 Lecture Feb 21, 2018 With thanks to David Sontag (MIT) and Tony Jebara (Columbia) for use of many slides and illustrations For more information,
More informationGeneralized Evidence Pre-propagated Importance Sampling for Hybrid Bayesian Networks
Generalized Evidence Pre-propagated Importance Sampling for Hybrid Bayesian Networks Changhe Yuan Department of Computer Science and Engineering Mississippi State University Mississippi State, MS 39762
More informationCSEP 573: Artificial Intelligence
CSEP 573: Artificial Intelligence Hidden Markov Models Luke Zettlemoyer Many slides over the course adapted from either Dan Klein, Stuart Russell, Andrew Moore, Ali Farhadi, or Dan Weld 1 Outline Probabilistic
More informationGraphical Models - Part I
Graphical Models - Part I Oliver Schulte - CMPT 726 Bishop PRML Ch. 8, some slides from Russell and Norvig AIMA2e Outline Probabilistic Models Bayesian Networks Markov Random Fields Inference Outline Probabilistic
More informationUncertainty. Logic and Uncertainty. Russell & Norvig. Readings: Chapter 13. One problem with logical-agent approaches: C:145 Artificial
C:145 Artificial Intelligence@ Uncertainty Readings: Chapter 13 Russell & Norvig. Artificial Intelligence p.1/43 Logic and Uncertainty One problem with logical-agent approaches: Agents almost never have
More informationBayesian Networks. Characteristics of Learning BN Models. Bayesian Learning. An Example
Bayesian Networks Characteristics of Learning BN Models (All hail Judea Pearl) (some hail Greg Cooper) Benefits Handle incomplete data Can model causal chains of relationships Combine domain knowledge
More informationReview: Bayesian learning and inference
Review: Bayesian learning and inference Suppose the agent has to make decisions about the value of an unobserved query variable X based on the values of an observed evidence variable E Inference problem:
More informationPublished in: Tenth Tbilisi Symposium on Language, Logic and Computation: Gudauri, Georgia, September 2013
UvA-DARE (Digital Academic Repository) Estimating the Impact of Variables in Bayesian Belief Networks van Gosliga, S.P.; Groen, F.C.A. Published in: Tenth Tbilisi Symposium on Language, Logic and Computation:
More informationY. Xiang, Inference with Uncertain Knowledge 1
Inference with Uncertain Knowledge Objectives Why must agent use uncertain knowledge? Fundamentals of Bayesian probability Inference with full joint distributions Inference with Bayes rule Bayesian networks
More informationBayesian networks. Chapter Chapter
Bayesian networks Chapter 14.1 3 Chapter 14.1 3 1 Outline Syntax Semantics Parameterized distributions Chapter 14.1 3 2 Bayesian networks A simple, graphical notation for conditional independence assertions
More informationApproximate Inference
Approximate Inference Simulation has a name: sampling Sampling is a hot topic in machine learning, and it s really simple Basic idea: Draw N samples from a sampling distribution S Compute an approximate
More information