Bayesian Networks to design optimal experiments. Davide De March

Similar documents
Directed and Undirected Graphical Models

Probabilistic Graphical Models (I)

Machine Learning 4771

Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence

Learning in Bayesian Networks

Tópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863

A Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004

CS Lecture 3. More Bayesian Networks

Markov Independence (Continued)

Directed Graphical Models

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Part I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008

Machine Learning Lecture 14

10708 Graphical Models: Homework 2

CSC 412 (Lecture 4): Undirected Graphical Models

Inference in Bayesian Networks

Rapid Introduction to Machine Learning/ Deep Learning

CS839: Probabilistic Graphical Models. Lecture 2: Directed Graphical Models. Theo Rekatsinas

COMP538: Introduction to Bayesian Networks

CPSC 540: Machine Learning

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech

Graphical Models and Kernel Methods

Lecture 17: May 29, 2002

Statistical Approaches to Learning and Discovery

Bayes Nets: Independence

Lecture 8: Bayesian Networks

Undirected Graphical Models: Markov Random Fields

p L yi z n m x N n xi

Junction Tree, BP and Variational Methods

Outline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

Bayesian Networks: Representation, Variable Elimination

1. what conditional independencies are implied by the graph. 2. whether these independecies correspond to the probability distribution

Exercises, II part Exercises, II part

Probabilistic Graphical Networks: Definitions and Basic Results

Decision-Theoretic Specification of Credal Networks: A Unified Language for Uncertain Modeling with Sets of Bayesian Networks

Graphical models. Sunita Sarawagi IIT Bombay

Directed Graphical Models or Bayesian Networks

CS 5522: Artificial Intelligence II

Chapter 16. Structured Probabilistic Models for Deep Learning

A Tutorial on Bayesian Belief Networks

COS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference

Representation of undirected GM. Kayhan Batmanghelich

Chris Bishop s PRML Ch. 8: Graphical Models

CS6220: DATA MINING TECHNIQUES

Bayesian Learning in Undirected Graphical Models

{ p if x = 1 1 p if x = 0

Introduction to Probabilistic Graphical Models

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.

Using Graphs to Describe Model Structure. Sargur N. Srihari

Inference as Optimization

Final Examination CS540-2: Introduction to Artificial Intelligence

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

Undirected Graphical Models

Lecture 6: Graphical Models

STAT 598L Probabilistic Graphical Models. Instructor: Sergey Kirshner. Bayesian Networks

Introduction to Artificial Intelligence. Unit # 11

Probabilistic Graphical Models. Rudolf Kruse, Alexander Dockhorn Bayesian Networks 153

Informatics 2D Reasoning and Agents Semester 2,

Lecture 6: Graphical Models: Learning

Directed and Undirected Graphical Models

TDT70: Uncertainty in Artificial Intelligence. Chapter 1 and 2

The Monte Carlo Method: Bayesian Networks

3 : Representation of Undirected GM

CSE 473: Artificial Intelligence Autumn 2011

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Tópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863

Lecture 4 October 18th

CS Lecture 4. Markov Random Fields

Towards an extension of the PC algorithm to local context-specific independencies detection

Implementing Machine Reasoning using Bayesian Network in Big Data Analytics

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

Based on slides by Richard Zemel

Machine Learning Summer School

COMP538: Introduction to Bayesian Networks

Outline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

Belief Update in CLG Bayesian Networks With Lazy Propagation

Preliminaries Bayesian Networks Graphoid Axioms d-separation Wrap-up. Bayesian Networks. Brandon Malone

An Introduction to Bayesian Machine Learning

Summary of the Bayes Net Formalism. David Danks Institute for Human & Machine Cognition

Conditional Independence

Artificial Intelligence Bayes Nets: Independence

Bayesian Network Modelling

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Bayesian Inference. Definitions from Probability: Naive Bayes Classifiers: Advantages and Disadvantages of Naive Bayes Classifiers:

Introduction to Bayesian Networks

Probabilistic Graphical Models

Message Passing and Junction Tree Algorithms. Kayhan Batmanghelich

Introduction to Bayesian Learning

Bayesian Networks. Motivation

Probabilistic Graphical Models

Review: Bayesian learning and inference

Events A and B are independent P(A) = P(A B) = P(A B) / P(B)

Markov properties for directed graphs

Bayesian Networks Inference with Probabilistic Graphical Models

Bayes Networks 6.872/HST.950

Bayesian Networks for Causal Analysis

Lecture 5: Bayesian Network

Transcription:

Bayesian Networks to design optimal experiments Davide De March davidedemarch@gmail.com 1

Outline evolutionary experimental design in high-dimensional space and costly experimentation the microwell mixture experiments for vesicle formation Bayesian Network models for designing experiments 2

The experimentation how can we generate vesicles with a mixture of amphiphilic molecules? 3

The biochemical technology microwell plates one experiment in each well (mixture exp.) a large set of molecules process variables (defined a priori by the experimentalists) 4

The mixture experiment each well contains a mixture of 5 molecules, selected from an initial set of 16 X= {X 1,,X 16 } the constraints in these experimentation are: 0 X i 1 16 i=1 X i =1 the system response is the Turbidity measured by a fluorescent microscope 5

The optimal design find the best subset of variables that leads to a maximum system response, defining: which factors are relevant for the response their levels their interactions 6

classical approaches (simplex lattice design, simplex fractional design, screening) cannot deal with high-dimensional problems. in our case the simplex lattice design defines 15504 points to be tested (too many and too costly) 7

The design as a dynamical process the iterative process in designing experiments allows to reach an improved or optimal system response DESIGN EXPERIMENT MODELING The use of statistical models helps experimentalists to select the most informative variables, the significant interactions and to define 8 a new experimental design

Experimental design by model we model the experimental data we identify the most significant variables and the most relevant interactions for the system response we use the Bayesian network to define an optimal experimental design 9

Bayesian Network (BN) BNs are probabilistic graphical models that specify the multivariate probability distributions on a set of variables. They are described by a triple {V,A,Θ} where \ 10

{V,A} is the topological part of the BN and represents the structure of the network. V is the set of nodes in the structure (each node is a variable) and A is the set of arches between nodes and describes the dependence between nodes (no arc between 2 nodes means independence) Θ is the quantitative part of the network, and represents the conditional probability distribution (BN's parameters) 11

Connections in the BN all the links of a BN can be summarised in three types of connections: Serial connection A B C in this connection A & C are conditionally independent if we have knowledge about the state of B 12

Diverging connection: A B C... E without knowledge about the state of A, BC&E are dependent. B,C&E are conditionally independent given the knowledge of the state of A (A block the information if it is observed) 13

Converging connection: B C... E A B,C&E are independent without knowledge about the state of A (or any of its descendant). Given evidence of the state of A, B,C&E become conditionally dependent 14

Conditional independence & Markov property Given three variables X, Y & Z, X &Y are conditionally independent given Z if: X (Y Z) P( X=x Y=y,Z=z)=P( X=x Z=z) x,y,z the Markov property: X i Nd ( X i ) Pa( X i ) X i X X is conditionally independent of its non descendants (nd) given its parents (Pa). It simplifies the computation of the probability distribution among nodes 15

Example In this network F ( S,X ) ( B,L ) L ( B S ) X (S L ) 16

The joint distribution the joint distribution of the example can be represented as: P (s,b,l,f,x )=P ( s ) P (b s ) P (l b,s ) P ( f b,s,l ) P ( x b,s,l,f ) due to the Markov property we have for x: P (x b,s,l,f ) =P ( x l) consequently the joint distribution can be expressed as: P (s,b,l,f,x )=P ( s ) P (b s ) P (l s) P ( f b,l ) P ( x l) the joint distribution of the network is expressed by: n P( x 1,,x n )= i=1 P ( x i Pa ( x i ) ) 17

Inference in BNs Given some evidences about the state of a variable, you can infer about the state of the other variables in the network the two binary variables case P(B A=a j )= 2 i=1 P( A=a j B ) P ( B) P ( A=a j B=b i ) P ( B=b i ) 18

Inference in a BN is a NP-hard problem (Cooper, 1990) there are many algorithms that make exact or approximate inferences Junction Tree propagation is an efficient algorithm with local procedure for BN inference (Jensen, 2001) 19

Junction Tree Junction Tree is an indirect acyclic graph its nodes are clusters of variables given two clusters, C 1 and C 2, every node on the path between them contains their intersection C 1 C 2 a Separator, S, is associated with each edge and contains the variables in the intersection between neighbouring nodes 20

A A C B A C a) Direct Acyclic Graph (DAG) D D E a) b) E b) Moral Graph with the marriage of parents B D A c) C E BC BCD ABC d) CD CDE c) Triangulated Graph no chordless circle of length 4. d)junction Tree 21

Learning the structure learning the structure means to create the best network for the process three approaches to model structure and parameters: expert knowledge modelling data (Search & Score) a mix of the two previous methods 22

Search & Score algorithms The Search & Score techniques are regarded as efficient approaches to learn the network structure (some examples GA, K2, Hill Climbing ) this approach maximises a score function considering all the possible DAGs (i.e it maximises the fitness of the model to the data) 23

The BN (S&S algorithm) to design the biochemical experiments the final structure: GeNIe Software with search and score (HC) and BIC information 24

Inference in the BN MeanT, is discretised in 6 levels. the aim of the analysis is to understand the system variations and variable relation in the biochemical process 25

Inference on the BN Software for inference - Netica (www.norsys.com) 26

Evidence on MeanT at high-level values X13 drastically increases its presence in the state b other variables decrease their presence (e.g X 15 X 10 ) 27

Results With the dependence/independence structure of the Bayesian network we highlight some important results: X 13 is the most important molecule to define high-values of the MeanT, but only at a certain level (0.2) some interactions, especially involving X 13, help to obtain higher value of the response The identification of the model and of the variable dependencies allows to provide the experimentalist with the next sets of experiment. 28

Thanks 29