Artificial Intelligence Bayes Nets

Similar documents
Outline. CSE 473: Artificial Intelligence Spring Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

Bayes Nets. CS 188: Artificial Intelligence Fall Example: Alarm Network. Bayes Net Semantics. Building the (Entire) Joint. Size of a Bayes Net

CS 5522: Artificial Intelligence II

Product rule. Chain rule

15-780: Graduate Artificial Intelligence. Bayesian networks: Construction and inference

Bayes Nets: Independence

Artificial Intelligence Bayes Nets: Independence

Directed Graphical Models

Bayesian Networks. Example of Inference. Another Bayes Net Example (from Andrew Moore) Example Bayes Net. Want to know P(A)? Proceed as follows:

Announcements. CS 188: Artificial Intelligence Spring Bayes Net Semantics. Probabilities in BNs. All Conditional Independences

Objectives. Probabilistic Reasoning Systems. Outline. Independence. Conditional independence. Conditional independence II.

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation

Probabilistic Reasoning Systems

Bayesian Networks. Axioms of Probability Theory. Conditional Probability. Inference by Enumeration. Inference by Enumeration CSE 473

CS 188: Artificial Intelligence Spring Announcements

CSE 473: Artificial Intelligence Autumn 2011

Bayesian belief networks

Bayesian networks. Chapter 14, Sections 1 4

Bayesian belief networks. Inference.

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

Bayesian networks. Chapter Chapter

CS 5522: Artificial Intelligence II

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

CS 188: Artificial Intelligence Fall 2009

Y. Xiang, Inference with Uncertain Knowledge 1

CS 343: Artificial Intelligence

Bayesian Networks Practice

Bayesian Networks BY: MOHAMAD ALSABBAGH

Be able to define the following terms and answer basic questions about them:

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018

Probabilistic Models. Models describe how (a portion of) the world works

Lecture 8: Bayesian Networks

PROBABILISTIC REASONING SYSTEMS

CS 484 Data Mining. Classification 7. Some slides are from Professor Padhraic Smyth at UC Irvine

COS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference

Recap: Bayes Nets. CS 473: Artificial Intelligence Bayes Nets: Independence. Conditional Independence. Bayes Nets. Independence in a BN

Artificial Intelligence Methods. Inference in Bayesian networks

CSEP 573: Artificial Intelligence

Announcements. CS 188: Artificial Intelligence Fall Example Bayes Net. Bayes Nets. Example: Traffic. Bayes Net Semantics

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.

Inference in Bayesian networks

CS 188: Artificial Intelligence Fall 2008

Bayesian Network. Outline. Bayesian Network. Syntax Semantics Exact inference by enumeration Exact inference by variable elimination

Introduction to Artificial Intelligence Belief networks

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Bayes Nets III: Inference

Bayesian Network Representation

Intro to AI: Lecture 8. Volker Sorge. Introduction. A Bayesian Network. Inference in. Bayesian Networks. Bayesian Networks.

Quantifying uncertainty & Bayesian networks

Biointelligence Lab School of Computer Sci. & Eng. Seoul National University

Artificial Intelligence Bayesian Networks

Probabilistic Graphical Networks: Definitions and Basic Results

Probabilistic Reasoning. (Mostly using Bayesian Networks)

CS 188: Artificial Intelligence Spring Announcements

Bayesian Networks. 22c:145 Artificial Intelligence. Bayesian Networks. Review of Probability Theory. Bayesian (Belief) Networks.

Graphical Models - Part I

Outline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

Bayesian Networks 3 D-separation. D-separation

Stochastic Methods. 5.0 Introduction 5.1 The Elements of Counting 5.2 Elements of Probability Theory

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Course Introduction. Probabilistic Modelling and Reasoning. Relationships between courses. Dealing with Uncertainty. Chris Williams.

last two digits of your SID

COMP5211 Lecture Note on Reasoning under Uncertainty

Probabilistic Representation and Reasoning

Probabilistic Classification

Probabilistic Graphical Models

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

Graphical Models and Kernel Methods

Probabilistic Models

Implementing Machine Reasoning using Bayesian Network in Big Data Analytics

Directed Graphical Models

Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

Probabilistic Models Bayesian Networks Markov Random Fields Inference. Graphical Models. Foundations of Data Analysis

Lecture 17: May 29, 2002

4 : Exact Inference: Variable Elimination

Artificial Intelligence

Definition: A "system" of equations is a set or collection of equations that you deal with all together at once.

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008

Be able to define the following terms and answer basic questions about them:

Probabilistic Graphical Models (I)

STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas. Bayesian networks: Representation

p L yi z n m x N n xi

Introduction to Artificial Intelligence. Unit # 11

CS 188: Artificial Intelligence. Bayes Nets

Bayesian Networks. Alan Ri2er

Bayesian networks (1) Lirong Xia

Bayesian Networks. Machine Learning, Fall Slides based on material from the Russell and Norvig AI Book, Ch. 14

p(x) p(x Z) = y p(y X, Z) = αp(x Y, Z)p(Y Z)

Introduction to Bayesian Networks. Probabilistic Models, Spring 2009 Petri Myllymäki, University of Helsinki 1

Bayesian Networks. Semantics of Bayes Nets. Example (Binary valued Variables) CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-III

An Introduction to Bayesian Networks in Systems and Control

Probabilistic Machine Learning

Introduction to Bayesian Networks

Tópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863

Directed Graphical Models or Bayesian Networks

Uncertainty and Bayesian Networks

Transcription:

rtificial Intelligence ayes Nets print troubleshooter (part of Windows 95) Nilsson - hapter 19 Russell and Norvig - hapter 14 ayes Nets; page 1 of 21 ayes Nets; page 2 of 21 joint probability distributions E1 E2 E3 E4 E5 D1 D2 D3 0.05 0.03 P(D1 E1, NO E3) - inefficient for reasoning - hard to acquire the probabilities ayes nets = belief nets make use of independence inherent in the domain expert systems: medicine, Microsoft P() ayes net = directed acyclic graph with conditional probability tables urglary Johnalls larm P(J).90.05 E P().95.94.29 Maryalls P(E).002 this really means P(= =,E=) P(M).70.01 nodes = random variables, links = direct influences ayes Nets; page 3 of 21 ayes Nets; page 4 of 21

conditional probability tables from ayes nets to joint probability distributions Y Y P( ND, Y) urglary P() P(E).002 ND 1.0 0.0 0.0 0.0 larm E P().95.94.29 OIN P(OIN) 0.5 Johnalls P(J).90.05 Maryalls P(M).70.01 urglary larm Johnalls Maryalls 0.000001197 P(, E,, J, M) = P() P(E) P(,E) P( J ) P(M ) ayes Nets; page 5 of 21 ayes Nets; page 6 of 21 ayes Nets; page 7 of 21 from joint probability distributions to ayes nets (1) repeatedly: - pick a variable - condition it on the smallest possible set of variables picked previously P(,, ) = P() P( ) P(,) 0.054 0.126 0.002 0.018 0.432 0.288 0.032 0.048 order:,, P() 0.2 P() 0.3 0.1 0.6 0.4 P() 0.9 0.9 ayes Nets; page 8 of 21 from joint probability distributions to ayes nets (2) 1 2 urglary 3 larm Johnalls Maryalls ordering does matter 4 5 urglary 3 larm 4 5 2 1 Johnalls Maryalls urglary Johnalls sizes of the conditional probability tables 10 13 put causes before effects - smaller network - easier to make probability judgements 2 4 5 3 31 larm 1 Maryalls

from joint probability distributions to ayes nets (3) independence example warning attery ayes nets merely represent joint probability distributions ayes nets have nothing to do with causality Radio Ignition Gas it is smart, but not necessary, to make the edges go from causes to effects Starts all six of these ayes nets are fine! Moves gas and radio - independent given ignition - independent given battery - independent given nothing - dependent given starts - dependent given moves ayes Nets; page 9 of 21 ayes Nets; page 10 of 21 direction-dependent separation if every undirected path from a node in to a node in Y is blocked by E, then and Y are conditionally independent given E three ways in which a path from to Y can be blocked by evidence E example (1) re Radio and Gas are guaranteed to be independent (not knowing anything)? (1) Z E Y Radio attery Ignition Gas Starts (2) (3) Z Z Yes, the structure guarantees it. here is only one undirected path from Radio to Gas. his path is blocked because Ignition -> Starts <- Gas is blocked. Moves ayes Nets; page 11 of 21 ayes Nets; page 12 of 21

example (2) wo astonomers, in different parts of the world, make measurements M1 and M2 of the number of stars N in some small region of the sky, using their telescopes. Normally, there is a small possibility of error by up to one star. Each telescope can also (with a slightly smaller probability) be badly out of focus (event 1 and 2), in which case the scientist will undercount by three or more stars. Does the following network correctly reflect these facts? example (2) wo astonomers, in different parts of the world, make measurements M1 and M2 of the number of stars N in some small region of the sky, using their telescopes. Normally, there is a small possibility of error by up to one star. Each telescope can also (with a slightly smaller probability) be badly out of focus (event 1 and 2), in which case the scientist will undercount by three or more stars. Does the following network correctly reflect these facts? 1 2 1 N 2 M1 M2 M1 M2 N ayes Nets; page 13 of 21 ayes Nets; page 14 of 21 example (3) some simple inferences ayes Nets; page 15 of 21 re urglary and Johnalls are guaranteed to be conditionally independent given larm? urglary urglary larm larm Johnalls Maryalls Johnalls Maryalls some of the unblocked undirected paths Yes, the structure guarantees it. No, the structure does not guarantee it. D P() P() P() P(D) 0.8 0.4 0.8 0.8 0.3 0.3 0.3 P( ) = 0.8 these do NO need P(NO ) = 1 - P( ) = 0.2 to sum to one P( NO ) = 0.3 P(NO NO ) = 1 - P( NO ) = 0.7 P() = P() P( ) + P(NO ) P( NO ) = 0.4 0.8 + 0.6 0.3 = 0.5 P( ) = P( ) P() / P() = 0.8 0.4 / 0.5 = 0.64 P(, ) = P() P( ) P( ) + P(NO ) P( NO ) P( NO ) = 0.4 0.8 0.8 + 0.6 0.3 0.3 = 0.31 P(D ) = P(D ) P( ) + P(D NO ) P(NO ) = 0.8 0.8 + 0.3 0.2 = 0.7 ayes Nets; page 16 of 21

more complex inferences complexity attery probabilistic inference on polytrees can be done in polynomial time Radio Ignition Gas polytrees are DGs where there is at most one path between any two nodes Starts Moves observe evidence; for example, symptoms calculate the probability of the various diseases given the evidence in general, probabilistic inference is NP hard 1a) 1b) + D D D clustering conditioning 2) stochastic simulation ayes Nets; page 17 of 21 ayes Nets; page 18 of 21 E 1 U 1 here: algorithms for causal chains E 1 U 1 here: algorithms for causal chains U 2 P( E 1, E 2 ) U 2 P( E 1 ) U 3 U 3 Y 3 Y 2 Y 1 P( E 2, E 1 ) = P(E 2, E 1 ) P( E 1 ) / P(E 2 E 1 ) P( E 2, E 1 ) is proportional to P(E 2 ) P( E 1 ) Y 3 Y 2 Y 1 for each node U i from U 1 to : P(U i E 1 ) = P(U i-1, U i E 1 ) + P(NO U i-1, U i E 1 ) = P(U i-1 E 1 ) P(U i U i-1, E 1 ) + P(NO U i-1 E 1 ) P(U i NO U i-1, E 1 ) = P(U i-1 E 1 ) P(U i U i-1 ) + (1 - P(U i-1 E 1 )) P(U i NO U i-1 ) E 2 E 2 ayes Nets; page 19 of 21 ayes Nets; page 20 of 21

E 1 U 1 U 2 here: algorithms for causal chains P(E 2 ) U 3 Y 3 Y 2 Y 1 for each node Y i from to Y 1 : P(E 2 Y i ) = P(Y i-1, E 2 Y i ) + P(NO Y i-1, E 2 Y i ) = P(Y i-1 Y i ) P(E 2 Y i, Y i-1 ) + P(NO Y i-1 Y i ) P(E 2 Y i, NO Y i-1 ) = P(Y i-1 Y i ) P(E 2 Y i-1 ) + (1 - P(Y i-1 Y i )) P(E 2 NO Y i-1 ) E 2 ayes Nets; page 21 of 21