Cell Biology. Lecture 1: 10-Oct-12. Marco Grzegorczyk. (Gen-)Regulatory Network. Microarray Chips. (Gen-)Regulatory Network. (Gen-)Regulatory Network

Similar documents
Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

Hidden Markov Models

Speech and Language Processing

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

JAB Chain. Long-tail claims development. ASTIN - September 2005 B.Verdier A. Klinger

EM and Structure Learning

Expectation Maximization Mixture Models HMMs

Bayesian Learning. Smart Home Health Analytics Spring Nirmalya Roy Department of Information Systems University of Maryland Baltimore County

Machine learning: Density estimation

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

Negative Binomial Regression

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Linear Regression Analysis: Terminology and Notation

Linear Approximation with Regularization and Moving Least Squares

Bayesian belief networks

Generative and Discriminative Models. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

Hidden Markov Models

Artificial Intelligence Bayesian Networks

Structure Learning. Instructor: Su-In Lee University of Washington, Seattle. Score-based structure learning

An adaptive SMC scheme for ABC. Bayesian Computation (ABC)

Generative classification models

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Web Appendix B Estimation. We base our sampling procedure on the method of data augmentation (e.g., Tanner and Wong,

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

Markov Chain Monte Carlo Lecture 6

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

CIS587 - Artificial Intellgence. Bayesian Networks CIS587 - AI. KB for medical diagnosis. Example.

Machine Learning for Signal Processing Linear Gaussian Models

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

Statistical analysis using matlab. HY 439 Presented by: George Fortetsanakis

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong

Basically, if you have a dummy dependent variable you will be estimating a probability.

I529: Machine Learning in Bioinformatics (Spring 2017) Markov Models

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

3.1 ML and Empirical Distribution

Machine Learning for Signal Processing Linear Gaussian Models

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1

Classification as a Regression Problem

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Interpolated Markov Models for Gene Finding

Comparison of Regression Lines

A New Method for Estimating Overdispersion. David Fletcher and Peter Green Department of Mathematics and Statistics

Mean Field / Variational Approximations

Linear regression. Regression Models. Chapter 11 Student Lecture Notes Regression Analysis is the

Introduction to Vapor/Liquid Equilibrium, part 2. Raoult s Law:

1/10/18. Definitions. Probabilistic models. Why probabilistic models. Example: a fair 6-sided dice. Probability

Statistics for Economics & Business

Probability Theory. The nth coefficient of the Taylor series of f(k), expanded around k = 0, gives the nth moment of x as ( ik) n n!

Maximum Likelihood Estimation (MLE)

Statistical inference for generalized Pareto distribution based on progressive Type-II censored data with random removals

9.913 Pattern Recognition for Vision. Class IV Part I Bayesian Decision Theory Yuri Ivanov

Learning Objectives for Chapter 11

Chapter 9: Statistical Inference and the Relationship between Two Variables

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

Lecture 3 Stat102, Spring 2007

CSC321 Tutorial 9: Review of Boltzmann machines and simulated annealing

Dynamical Systems and Information Theory

TCOM 501: Networking Theory & Fundamentals. Lecture 7 February 25, 2003 Prof. Yannis A. Korilis

Feature Selection & Dynamic Tracking F&P Textbook New: Ch 11, Old: Ch 17 Guido Gerig CS 6320, Spring 2013

Finite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Limited Dependent Variables and Panel Data. Tibor Hanappi

e i is a random error

A quantum-statistical-mechanical extension of Gaussian mixture model

6. Stochastic processes (2)

Empirical Methods for Corporate Finance. Identification

Chapter 11: Simple Linear Regression and Correlation

6. Stochastic processes (2)

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Marginal Models for categorical data.

Learning with Partially Observed Data

Evaluation for sets of classes

A Robust Method for Calculating the Correlation Coefficient

An Introduction to Bayesian Networks: Concepts and Learning from Data

Boise State University Department of Electrical and Computer Engineering ECE 212L Circuit Analysis and Design Lab

Global Sensitivity. Tuesday 20 th February, 2018

Why Bayesian? 3. Bayes and Normal Models. State of nature: class. Decision rule. Rev. Thomas Bayes ( ) Bayes Theorem (yes, the famous one)

Homework Assignment 3 Due in class, Thursday October 15

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore

Lecture Notes on Linear Regression

Statistical learning

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

NETWORK IDENTIFICATION AND TIME SERIES ANALYIS USING S-SYSTEMS

4.3 Poisson Regression

THEOREMS OF QUANTUM MECHANICS

Learning undirected Models. Instructor: Su-In Lee University of Washington, Seattle. Mean Field Approximation

Hopfield networks and Boltzmann machines. Geoffrey Hinton et al. Presented by Tambet Matiisen

Bayesian decision theory. Nuno Vasconcelos ECE Department, UCSD

Solution of Equilibrium Equation in Dynamic Analysis. Mode Superposition. Dominik Hauswirth Method of Finite Elements II Page 1

Aging model for a 40 V Nch MOS, based on an innovative approach F. Alagi, R. Stella, E. Viganò

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Course 395: Machine Learning - Lectures

11 Bayesian Methods. p(w D) = p(d w)p(w) (1)

Chapter 13: Multiple Regression

Methods Lunch Talk: Causal Mediation Analysis

L1 Variable Selection Based Causal Structure Learning

Discriminative classifier: Logistic Regression. CS534-Machine Learning

Transcription:

5.0.202 Genetsche Netzwerke Wntersemester 202/203 ell ology Lecture : 0-Oct-2 Marco Grzegorczyk Gen-Regulatory Network Mcroarray hps G G 2 G 3 2 3 metabolte metabolte Gen-Regulatory Network Gen-Regulatory Network G G 2 G 3 2 3 G G 3 G 2 metabolte metabolte

5.0.202 Regulatory networks and sgnallng pathways low cytometry technology Receptor molecules ell membrane ctvaton Interacton n sgnallng pathway hosphorylated proten Inhbton rom Sachs et al Scence 2005 roten actvaton cascade roten actvaton cascade ell membran ell membran phosphorylaton nucleus phosphorylaton nucleus -> cell response -> cell response roten actvaton cascade 2

5.0.202 possbly completely unknown possbly completely unknown.g.: low cytometry eperments Here: oncentratons of phosphorylated protens possbly completely unknown Statstcal Task tract a network from a matr.g.: low cytometry eperments varables... n genes/ protens n 2 cells 2 22 n2 m 2m nm Machne Learnng statstcal methods ther m ndependent steady-state observatons of the system N Or tme seres of the system of length m: n t= m lementary molecular bologcal processes escrpton wth dfferental equatons oncentratons Knetc parameters q Rates 3

5.0.202 Gven: Gene epresson tme seres arameters q known: Numercally ntegrate the dfferental equatons for dfferent hypothetcal networks an we nfer the correct gene regulatory network? Model selecton for known parameters q Model selecton for unknown parameters q Measured gene epresson tme seres Gene epresson tme seres predcted wth dfferent models Measured gene epresson tme seres Gene epresson tme seres predcted wth dfferent models ompare Hghest lkelhood: best model Hghest lkelhood: over-fttng Model selecton: fnd the best pathway Select the model wth the hghest posteror probablty: Ths requres an ntegraton of the whole parameter space: Statc ayesan networks NOS GS Marrage between theory and probablty theory. rected acyclc G represents condtonal ndependence relatons. Markov assumpton leads to a factorzaton of the ont probablty dstrbuton: Ths ntegral s usually ntractable especally for systems of non-lnear dfferental equatons 4

5.0.202 5 ayesan networks versus causal networks ayesan networks represent condtonal ndependency relatons - not necessarly causal nteractons. quvalence classes of Ns completed partally drected s Gs v-structure = = Y Z ynamc ayesan networks for tme seres Y Y Z Z t t+ Interpretaton: No need for the acyclcty constrant!!! Unfoldng n tme ayesan networks n b N 2 ~ b 0 arametersaton: Gaussan Ge scorng metrc: ~NμΣ wth normal-wshart dstrbuton of the unknown parameters.e.: μ~nμ*vw - and W~WshartT 0 d ayesan networks d Ge metrc: closed form soluton score Ge unform dstrbuton

5.0.202 6 Learnng the network structure n 4 6 8 0 #Gs 543 37 0 6 78 0 42 0 8 Idea: Heurstcally searchng for the M* that s most supported by the M*> e.g.: greedy search algorthm score Ge Learnng the network structure strbuton of MM samplng of ayesan networks etter dea: ayesan model averagng va Markov han Monte arlo MM smulatons onstruct and smulate a Markov han M t t n the space of Gs whose dstrbuton converges to the posteror dstrbuton as statonary dstrbuton.e.: M t = t to generate a G sample: G G 2 G 3 G T Structure MM samplng scheme based on sngle edge operatons quvalence classes of Ns completed partally drected s Gs v-structure = =

5.0.202 Gs G representatons Gs Utlse the G G sample for estmatng the posteror probablty of edge relaton features: T ˆ I G T nterpretaton superposton where IG s f the G of G contans the drected edge and 0 otherwse MM samplng of ayesan networks The G sample G G 2 G 3 G T s generated va Markov han Monte arlo MM smulatons That s va smulaton of a Markov han M t t n the space of Gs whose dstrbuton converges to the posteror dstrbuton: M t = t MM samplng of ayesan networks The G sample G G 2 G 3 G T s generated va Markov han Monte arlo MM smulatons That s va smulaton of a Markov han M t t n the space of Gs whose dstrbuton converges to the posteror dstrbuton: M t = t In practce: t s not nfnte!!! robablstc nference true regulatory network edge posteror probabltes hgh Thresholdng low T:/2 :0/4 concrete network predctons T:2/2 :/4 7