Artificial Intelligence

Similar documents
Stochastic inference in Bayesian networks, Markov chain Monte Carlo methods

Inference in Bayesian networks

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

Inference in Bayesian Networks

Inference in Bayesian networks

Artificial Intelligence Methods. Inference in Bayesian networks

Objectives. Probabilistic Reasoning Systems. Outline. Independence. Conditional independence. Conditional independence II.

Probabilistic Reasoning Systems

Bayesian Network. Outline. Bayesian Network. Syntax Semantics Exact inference by enumeration Exact inference by variable elimination

Bayesian networks. Chapter 14, Sections 1 4

Bayesian networks. Chapter Chapter

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

CS 343: Artificial Intelligence

PROBABILISTIC REASONING SYSTEMS

Outline } Exact inference by enumeration } Exact inference by variable elimination } Approximate inference by stochastic simulation } Approximate infe

CS 188: Artificial Intelligence. Bayes Nets

Bayesian Networks. Philipp Koehn. 6 April 2017

Bayesian Networks. Philipp Koehn. 29 October 2015

Bayes Nets III: Inference

Bayesian Networks BY: MOHAMAD ALSABBAGH

Uncertainty and Bayesian Networks

Informatics 2D Reasoning and Agents Semester 2,

Artificial Intelligence Bayesian Networks

Lecture 10: Bayesian Networks and Inference

Bayes Nets: Sampling

Sampling from Bayes Nets

Approximate Inference

qfundamental Assumption qbeyond the Bayes net Chain rule conditional independence assumptions,

Machine Learning for Data Science (CS4786) Lecture 24

Intelligent Systems (AI-2)

14 PROBABILISTIC REASONING

Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo. Sampling Methods. Oliver Schulte - CMPT 419/726. Bishop PRML Ch.

CS 188: Artificial Intelligence Spring Announcements

Belief Networks for Probabilistic Inference

COS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference

Bayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling

School of EECS Washington State University. Artificial Intelligence

Markov Networks.

Quantifying uncertainty & Bayesian networks

Bayes Networks 6.872/HST.950

Sampling Algorithms for Probabilistic Graphical models

Introduction to Artificial Intelligence Belief networks

CS 188, Fall 2005 Solutions for Assignment 4

Probabilistic Reasoning. Kee-Eung Kim KAIST Computer Science

Introduction to Artificial Intelligence (AI)

Outline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

Bayes Networks. CS540 Bryan R Gibson University of Wisconsin-Madison. Slides adapted from those used by Prof. Jerry Zhu, CS540-1

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

COMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma

Markov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials)

last two digits of your SID

Markov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials)

Directed Graphical Models

Artificial Intelligence Bayes Nets: Independence

MCMC and Gibbs Sampling. Kayhan Batmanghelich

Markov Chains and MCMC

Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo. Sampling Methods. Machine Learning. Torsten Möller.

15-780: Graduate Artificial Intelligence. Bayesian networks: Construction and inference

Graphical Models and Kernel Methods

PROBABILISTIC REASONING Outline

Bayesian networks: approximate inference

Sampling Methods. Bishop PRML Ch. 11. Alireza Ghane. Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech

4 : Exact Inference: Variable Elimination

CS 5522: Artificial Intelligence II

Course Overview. Summary. Outline

Bayesian networks. Independence. Bayesian networks. Markov conditions Inference. by enumeration rejection sampling Gibbs sampler

Bayes Nets: Independence

Inference in Bayesian networks

9 Forward-backward algorithm, sum-product on factor graphs

Exact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = =

Informatics 2D Reasoning and Agents Semester 2,

CS 730/730W/830: Intro AI

Bayesian Methods in Artificial Intelligence

Uncertainty. Introduction to Artificial Intelligence CS 151 Lecture 2 April 1, CS151, Spring 2004

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Probability. CS 3793/5233 Artificial Intelligence Probability 1

Statistical Approaches to Learning and Discovery

An Introduction to Bayesian Networks: Representation and Approximate Inference

Quantifying Uncertainty

Soft Computing. Lecture Notes on Machine Learning. Matteo Matteucci.

Introduction to Probabilistic Reasoning. Image credit: NASA. Assignment

STA 4273H: Statistical Machine Learning

Bayesian belief networks. Inference.

Graphical Models - Part I

Learning With Bayesian Networks. Markus Kalisch ETH Zürich

Product rule. Chain rule

Probabilistic Reasoning. (Mostly using Bayesian Networks)

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

A Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004

Bayesian Networks. Machine Learning, Fall Slides based on material from the Russell and Norvig AI Book, Ch. 14

Bayes Net Representation. CS 188: Artificial Intelligence. Approximate Inference: Sampling. Variable Elimination. Sampling.

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012

PROBABILITY. Inference: constraint propagation

Hidden Markov Models. Vibhav Gogate The University of Texas at Dallas

Uncertainty. 22c:145 Artificial Intelligence. Problem of Logic Agents. Foundations of Probability. Axioms of Probability

Bayesian Networks. Motivation

Reasoning Under Uncertainty

Introduction to Bayesian Learning

Transcription:

ICS461 Fall 2010 Nancy E. Reed nreed@hawaii.edu 1 Lecture #14B Outline Inference in Bayesian Networks Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic simulation Approximate inference by Markov Chain Monte Carlo (MCMC) Ref: Ch 14, sec 4-5 Inference Tasks Simple queries: compute posterior marginal P(X i E=e) e.g., P(NoGas Gauge = empty; Lights=on; Starts=false) Conjunctive queries: P(X i,x j E=e) = P(X i E=e) * P(X j X i,e=e) Optimal decisions: decision networks include utility information; probabilistic inference required for P(outcome action; evidence) Value of information: which evidence to seek next? Sensitivity analysis: which probability values are most critical? Explanation: why do I need a new starter motor? 3 Inference by Enumeration Slightly intelligent way to sum out variables from the joint without actually constructing its explicit representation Simple query on the burglary network: Rewrite full joint entries using product of CPT entries: Recursive depth-first enumeration: O(n) space, O(d n ) time 4 Enumeration Algorithm Evaluation Tree Enumeration is inefficient: repeated computation e.g., computes P(j a)p(m a) for each value of e 1

Inference by Variable Elimination Variable Elimination: Basic Operations Variable Elimination Algorithm Irrelevant Variables Irrelevant Variables, cont. Complexity of Exact Inference Singly connected networks (or polytrees): - any two nodes are connected by at most one (undirected) path - time and space cost of variable elimination are O(d k n) Multiply connected networks: - can reduce 3SAT to exact inference NP-hard - equivalent to counting 3SAT models NP-complete 2

Inference by Stochastic Simulation Basic idea: 1) Draw N samples from a sampling distribution S 2) Compute an approximate posterior probability P 3) Show this converges to the true probability P Outline: Sampling from an empty network Rejection sampling: reject samples disagreeing with evidence Likelihood weighting: use evidence to weight samples Markov Chain Monte Carlo (MCMC): sample from a stochastic process whose stationary distribution is the true posterior Sampling From an Empty Network Example Example, cont. 3

Sampling From an Empty Network, cont. Rejection Sampling Analysis of Rejection Sampling 4

Likelihood Weighting Likelyhood Weighting Example ω = 1.0 ω = 1.0 ω = 1.0 X 0.1 ω = 1.0 X 0.1 ω = 1.0 X 0.1 5

ω = 1.0 X 0.1 ω = 1.0 x 0.1 x 0.99 = 0.099 Likelihood Weighting Analysis Approximate Inference Using MCMC State of network = current assignment to all variables. Generate next state by sampling one variable given Markov blanket Sample each variable in turn, keeping evidence fixed. Hence likelihood weighting returns consistent estimates but performance still degrades with many evidence variables because a few samples have nearly all the total weight Can also choose a variable to sample at random each time The Markov Chain With Sprinkler = true, WetGrass = true, there are four states: Wander about for a while, average what you see MCMC Example, cont. Estimate P(Rain Sprinkler=true,WetGrass=true) Sample Cloudy or Rain given its Markov blanket, repeat. Count number of times Rain is true and false in the samples E.g., visit 100 states 31 have Rain=true, 69 have Rain=false P(Rain Sprinkler=true,WetGrass=true) = Normalize(h31,69i) = (0:31,0:69) Theorem: chain approaches stationary distribution: long-run fraction of time spent in each state is exactly proportional to its posterior probability 6

Markov Blanket Sampling Summary Exact inference by variable elimination: polytime on polytrees, NP-hard on general graphs space = time, very sensitive to topology Approximate inference by LW, MCMC: LW does poorly when there is lots of (downstream) evidence LW, MCMC generally insensitive to topology Convergence can be very slow with probabilities close to 1 or 0 Can handle arbitrary combinations of discrete and continuous variables 7