Summary of the Bayes Net Formalism. David Danks Institute for Human & Machine Cognition

Similar documents
Learning in Bayesian Networks

Directed Graphical Models

Directed Graphical Models or Bayesian Networks

Causality in Econometrics (3)

Rapid Introduction to Machine Learning/ Deep Learning

Part I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS

Probabilistic Models

Introduction to Causal Calculus

STAT 598L Probabilistic Graphical Models. Instructor: Sergey Kirshner. Bayesian Networks

Bayes Nets: Independence

Directed and Undirected Graphical Models

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008

2 : Directed GMs: Bayesian Networks

CS Lecture 3. More Bayesian Networks

Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence

Bayesian Approaches Data Mining Selected Technique

Probabilistic Graphical Networks: Definitions and Basic Results

1. what conditional independencies are implied by the graph. 2. whether these independecies correspond to the probability distribution

CS 5522: Artificial Intelligence II

Tópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Today. Why do we need Bayesian networks (Bayes nets)? Factoring the joint probability. Conditional independence

CS 5522: Artificial Intelligence II

Challenges (& Some Solutions) and Making Connections

Bayesian networks (1) Lirong Xia

An Introduction to Reversible Jump MCMC for Bayesian Networks, with Application

CMPT Machine Learning. Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th

Bayesian Inference. Definitions from Probability: Naive Bayes Classifiers: Advantages and Disadvantages of Naive Bayes Classifiers:

Bayesian Networks. Machine Learning, Fall Slides based on material from the Russell and Norvig AI Book, Ch. 14

Artificial Intelligence Bayes Nets: Independence

Representation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models

Y. Xiang, Inference with Uncertain Knowledge 1

Introduction to Artificial Intelligence. Unit # 11

Based on slides by Richard Zemel

CS 343: Artificial Intelligence

Abstract. Three Methods and Their Limitations. N-1 Experiments Suffice to Determine the Causal Relations Among N Variables

2 : Directed GMs: Bayesian Networks

Probabilistic Models. Models describe how (a portion of) the world works

STA 4273H: Statistical Machine Learning

K. Nishijima. Definition and use of Bayesian probabilistic networks 1/32

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

Bayesian Machine Learning

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech

Learning P-maps Param. Learning

An Introduction to Bayesian Machine Learning

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

CS 188: Artificial Intelligence Spring Announcements

Chris Bishop s PRML Ch. 8: Graphical Models

CS 188: Artificial Intelligence Fall 2009

Bayesian Graphical Models for Structural Vector AutoregressiveMarch Processes 21, / 1

CSE 473: Artificial Intelligence Autumn 2011

Review: Bayesian learning and inference

COMP538: Introduction to Bayesian Networks

Chapter 16. Structured Probabilistic Models for Deep Learning

CS 188: Artificial Intelligence Fall 2008

Probabilistic Graphical Models

The Origin of Deep Learning. Lili Mou Jan, 2015

1 : Introduction. 1 Course Overview. 2 Notation. 3 Representing Multivariate Distributions : Probabilistic Graphical Models , Spring 2014

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Bayesian Networks Basic and simple graphs

Lecture 6: Graphical Models: Learning

A Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004

An Empirical-Bayes Score for Discrete Bayesian Networks

Statistical Approaches to Learning and Discovery

CS839: Probabilistic Graphical Models. Lecture 2: Directed Graphical Models. Theo Rekatsinas

Probabilistic Graphical Models

Lecture 4 October 18th

Machine Learning Summer School

Bayesian Networks. Motivation

Bayesian Networks Representation

Bayesian Networks Inference with Probabilistic Graphical Models

Learning Bayesian Networks (part 1) Goals for the lecture

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Outline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

Being Bayesian About Network Structure:

Artificial Intelligence: Cognitive Agents

Uncertainty and Bayesian Networks

Graphical Representation of Causal Effects. November 10, 2016

Probabilistic Graphical Models (I)

Lecture 9: PGM Learning

Probabilistic Causal Models

Robust Monte Carlo Methods for Sequential Planning and Decision Making

3 : Representation of Undirected GM

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

BN Semantics 3 Now it s personal! Parameter Learning 1

Causal Inference & Reasoning with Causal Bayesian Networks

Causal Bayesian networks. Peter Antal

STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas. Bayesian networks: Representation

Motivation. Bayesian Networks in Epistemology and Philosophy of Science Lecture. Overview. Organizational Issues

Introduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013

COMP5211 Lecture Note on Reasoning under Uncertainty

LEARNING WITH BAYESIAN NETWORKS

Undirected Graphical Models

Bayesian Networks to design optimal experiments. Davide De March

A brief introduction to Conditional Random Fields

Transcription:

Summary of the Bayes Net Formalism David Danks Institute for Human & Machine Cognition

Bayesian Networks Two components: 1. Directed Acyclic Graph (DAG) G: There is a node for every variable D: Some nodes have arrows btw. them (X Y) A: Can t get from one node back to itself by following arrows 2. Joint Probability Distribution For all {X,Y,,Z}, P(X=x,Y=y,,Z=z) is defined

Example of a Bayes Net Directed Acyclic Graph: Air Pressure Barometer Storm Tomorrow Joint Probability Distribution: P(AP=High, B=High, ST=Yes) = 0.2 P(AP=High, B=Low, ST=Yes) = 0.005

Connecting the Graph & JPD Markov assumption: X is (probabilistically) independent of its (graphical) non-descendants conditional on its (graphical) parents. No edge Conditional independence

Connecting the Graph & JPD Markov assumption: X is (probabilistically) independent of its (graphical) non-descendants conditional on its (graphical) parents. No edge Conditional independence Example: X Y Z X Z Y

Connecting the Graph & JPD The Markov assumption implies a factorization of the JPD based on the graph: P(X=x,Y=y,,Z=z) = Π P(v parents(v)) The Markov assumption allows us to move from graph to probability distribution

Connecting the Graph & JPD Faithfulness assumption: The (probabilistic) effects of (graphical) paths never exactly offset. Conditional independence No edge The Faithfulness assumption allows us to move from probability distribution to graph

Bayesian Network Example 1. Running causes you to eat more 2. Eating more causes you to gain weight 3. Running increases your metabolism 4. Increased metabolism leads to weight loss + Running Food Eaten + Weight + Metabolism Note: Faithfulness rules out Running Weight

Learning Bayes Nets Given some data from the world, why would we want to learn a Bayes net? 1. Compact representation of the data There are fast algorithms for prediction/inference given observations of the environment 2. Causal knowledge There are fast algorithms for prediction/inference given interventions in the environment

Bayesian Updating Start with a probability distribution over possible graphs P(G)

Bayesian Updating Start with a probability distribution over possible graphs, then figure out which graph is most likely given the observed data P(G)

Features of Bayesian Updating Advantages: 1. Output is fine-grained probabilistic information 2. Rational basis for the learning algorithm 3. Robust to data errors Disadvantages: 1. Number of possible graphs is super-exponential, & likelihood functions not always solvable analytically almost always use heuristics 2. Cannot easily incorporate unobserved variables

Constraint-Based Learning 1. Determine the (un)conditional associations and independencies in the data 2. Determine the set of Bayes nets that could have produced that data

Constraint-Based Learning 1. Determine the (un)conditional associations and independencies in the data 2. Determine the set of Bayes nets that could have produced that data X W (all) Y Z (all) X Y,Z W Z X,W Y

Constraint-Based Learning 1. Determine the (un)conditional associations and independencies in the data 2. Determine the set of Bayes nets that could have produced that data X W (all) Y Z (all) X U Z X Y,Z W Z X,W Y W Y

Features of C-B Learning Advantages: 1. Feasible, asymptotically correct algorithms (though worst case requires exponential comp.) 2. Easily incorporates unobserved variables 3. Gives exactly the information in the data Disadvantages: 1. Susceptible to mistaken independence judgments (big problem for small datasets) 2. Cannot use fine-grained prior knowledge

Layers of Bayes Nets Start with simple association graph

Layers of Bayes Nets Then we can add a causal interpretation:

Layers of Bayes Nets Or we can provide parameter information: P(X)= P(Y X)= A C B

Layers of Bayes Nets Or we can combine causation & parameters P(X)= P(Y X)= A C B P(X)= P(Y X)= A C B

Useful Websites Tutorials: http://www.cs.berkeley.edu/~murphyk/bayes/bayes.html Constraint-based learning software: http://www.phil.cmu.edu/tetrad/index.html (free) Bayesian learning software: http://www.hugin.com (commercial) http://research.microsoft.com/~dmax/winmine/tooldoc.htm (free)