Optimal State Estimation for Boolean Dynamical Systems using a Boolean Kalman Smoother

Similar documents
State-Feedback Control of Partially-Observed Boolean Dynamical Systems Using RNA-Seq Time Series Data

Multiple Model Adaptive Controller for Partially-Observed Boolean Dynamical Systems

The Kalman Filter ImPr Talk

GLOBEX Bioinformatics (Summer 2015) Genetic networks and gene expression data

Completion Time of Fuzzy GERT-type Networks with Loops

Why do we care? Measurements. Handling uncertainty over time: predicting, estimating, recognizing, learning. Dealing with time

Introduction to Bioinformatics

Intrinsic Noise in Nonlinear Gene Regulation Inference

Basic modeling approaches for biological systems. Mahesh Bule

Chapter 05: Hidden Markov Models

Why do we care? Examples. Bayes Rule. What room am I in? Handling uncertainty over time: predicting, estimating, recognizing, learning

Linear Dynamical Systems

Bioinformatics 2 - Lecture 4

Human-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Hidden Markov Models. AIMA Chapter 15, Sections 1 5. AIMA Chapter 15, Sections 1 5 1

Introduction to Bioinformatics

Human Pose Tracking I: Basics. David Fleet University of Toronto

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Cybergenetics: Control theory for living cells

Lecture 5: Logistic Regression. Neural Networks

Multiscale Systems Engineering Research Group

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models

Computer Intensive Methods in Mathematical Statistics

Synchronous state transition graph

STA 4273H: Statistical Machine Learning

Simulation of Gene Regulatory Networks

A Higher-Order Interactive Hidden Markov Model and Its Applications Wai-Ki Ching Department of Mathematics The University of Hong Kong

ON SCALABLE CODING OF HIDDEN MARKOV SOURCES. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose

Bayesian Networks BY: MOHAMAD ALSABBAGH

Sensitivity Analysis for Discrete-Time Randomized Service Priority Queues

arxiv: v1 [q-bio.mn] 7 Nov 2018

Note Set 5: Hidden Markov Models

Networks in systems biology

5.3 METABOLIC NETWORKS 193. P (x i P a (x i )) (5.30) i=1

Probabilistic reconstruction of the tumor progression process in gene regulatory networks in the presence of uncertainty

Probabilistic Graphical Models

Inferring biological dynamics Iterated filtering (IF)

Modeling with Itô Stochastic Differential Equations

Introduction to Artificial Intelligence (AI)

Machine Learning for OR & FE

State Space and Hidden Markov Models

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Linear Discrete-time State Space Realization of a Modified Quadruple Tank System with State Estimation using Kalman Filter

On Optimal Coding of Hidden Markov Sources

Cellular Systems Biology or Biological Network Analysis

Hidden Markov models 1

Lecture 4: Hidden Markov Models: An Introduction to Dynamic Decision Making. November 11, 2010

A NONLINEARITY MEASURE FOR ESTIMATION SYSTEMS

EM-algorithm for Training of State-space Models with Application to Time Series Prediction

Towards a Bayesian model for Cyber Security

PROBABILISTIC REASONING OVER TIME

Household Energy Disaggregation based on Difference Hidden Markov Model

Modeling and state estimation Examples State estimation Probabilities Bayes filter Particle filter. Modeling. CSC752 Autonomous Robotic Systems

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

1 What is a hidden Markov model?

Enhancing a Model-Free Adaptive Controller through Evolutionary Computation

Sig2GRN: A Software Tool Linking Signaling Pathway with Gene Regulatory Network for Dynamic Simulation

LEARNING DYNAMIC SYSTEMS: MARKOV MODELS

Hidden Markov models for time series of counts with excess zeros

CS 5522: Artificial Intelligence II

Modelling residual wind farm variability using HMMs

Written Exam 15 December Course name: Introduction to Systems Biology Course no

Course 495: Advanced Statistical Machine Learning/Pattern Recognition

Stochastic Models, Estimation and Control Peter S. Maybeck Volumes 1, 2 & 3 Tables of Contents

Lecture 7: Simple genetic circuits I

Measures for information propagation in Boolean networks

CS 5522: Artificial Intelligence II

Hidden Markov Models. By Parisa Abedi. Slides courtesy: Eric Xing

Chaotic Dynamics in an Electronic Model of a Genetic Network

CSEP 573: Artificial Intelligence

Discovering molecular pathways from protein interaction and ge

MACHINE LEARNING 2 UGM,HMMS Lecture 7

DNA regulatory circuits can be often described by

Page 1. References. Hidden Markov models and multiple sequence alignment. Markov chains. Probability review. Example. Markovian sequence

Fuzzy Clustering of Gene Expression Data

CSE 473: Artificial Intelligence

Kalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q

Condensed Table of Contents for Introduction to Stochastic Search and Optimization: Estimation, Simulation, and Control by J. C.

Miscellaneous. Regarding reading materials. Again, ask questions (if you have) and ask them earlier

An Brief Overview of Particle Filtering

Computer Intensive Methods in Mathematical Statistics

Lecture 6: Bayesian Inference in SDE Models

Undirected Graphical Models

CS 343: Artificial Intelligence

Dynamic System Identification using HDMR-Bayesian Technique

RAO-BLACKWELLISED PARTICLE FILTERS: EXAMPLES OF APPLICATIONS

Probability and Time: Hidden Markov Models (HMMs)

9 Multi-Model State Estimation

arxiv: v1 [q-bio.mn] 2 Aug 2018

arxiv: v1 [cs.sy] 25 Oct 2017

Sensor Fusion: Particle Filter

Hidden Markov Models Hamid R. Rabiee

Regulation of Gene Expression

Convolutional networks. Sebastian Seung

L06. LINEAR KALMAN FILTERS. NA568 Mobile Robotics: Methods & Algorithms

Using the Kalman Filter to Estimate the State of a Maneuvering Aircraft

The Ensemble Kalman Filter:

A Probabilistic Relational Model for Characterizing Situations in Dynamic Multi-Agent Systems

New Statistical Model for the Enhancement of Noisy Speech

Transcription:

Optimal State Estimation for Boolean Dynamical Systems using a Boolean Kalman Smoother Mahdi Imani and Ulisses Braga-Neto Department of Electrical and Computer Engineering Texas A&M University College Station, TX, USA Email: m.imani88@tamu.edu, ulisses@ece.tamu.edu Abstract This paper is concerned with state estimation at a fixed time point in a given time series of observations of a Boolean dynamical system. Towards this end, we introduce the Boolean Kalman Smoother, which provides an efficient algorithm to compute the optimal MMSE state estimator for this problem. Performance is investigated using a Boolean network model of the p53-mdm2 negative feedback loop gene regulatory network observed through time series of Next-Generation Sequencing (NGS) data. Index Terms Boolean Dynamical System, Boolean Kalman Smoother, Gene Regulatory Network I. INTRODUCTION Gene regulatory networks govern the functioning of key cellular processes, such as the cell cycle, stress response, and DNA repair. Boolean networks [1], [2] have emerged as an effective model of the dynamical behavior of regulatory network consisting of genes in activated/inactivated states, the relationship among which is governed by networks of logical gates updated at discrete time intervals. The Boolean Kalman Filter (BKF) [3] [8] is an online, recursive algorithm to compute the optimal state estimator for a signal model consisting of a Boolean state process measured through a noisy observation process. The state process corresponds to a Boolean network with noisy state transitions, while the observation process is quite general. In this paper, we present an extension of the BKF to the smoothing problem, when a fixed interval of data is acquired and available for processing offline. This algorithm, called the Boolean Kalman Smoother (BKS), bears similarities to the forward-backward algorithm used in the inference of hidden markov models [9]. Performance of the BKS in the inference of Boolean networks is investigated using a model of the p53- MDM2 negative feedback loop network observed through next-generation sequencing data. The results indicate that on average, the BKS has lower MSE and lower error rates than the BKF, as expected. II. STOCHASTIC SIGNAL MODEL Deterministic Boolean network models are unable to cope with (1) uncertainty in state transition due to system noise and the effect of unmodeled variables; the fact that (2) the Boolean states of a system are never observed directly, but only indirectly through expression-based technologies, such as RNA-seq. This calls for a stochastic approach. We describe below the stochastic signal model for Boolean dynamical systems, first proposed in [3], which will be employed here. A. State Model Assume that the system is described by a state process {X k ; k = 0, 1,...}, where X k {0, 1} d is a Boolean vector of size d in the case of a gene regulatory network, the components of X k represent the activation/inactivation state, at discrete time k, of the genes comprising the network. The state is assumed to be updated at each discrete time through the following nonlinear signal model: X k = f (X k 1, u k 1 ) n k (state model) (1) for k = 1, 2,...; where u k 1 {0, 1} p is an input vector of dimension p at time k 1, {n k ; k = 1, 2,...} is a white noise process with n k = (N k1,..., N kd ) {0, 1} d, f {0, 1} d+p {0, 1} d is a network function, and indicates component-wise modulo-2 addition. In this paper, we employ a specific model for the network function that is motivated by gene pathway diagrams commonly encountered in biomedical research. We assume that the state vector is updated according to

the following equation (where the input u k 1 is omitted): 1, d j=1 a ij (X k 1 ) j + b i > 0, (X k ) i = f i (X k 1 ) = 0, d j=1 a ij (X k 1 ) j + b i < 0, (2) for i = 1, 2,..., d, where the parameter a ij can take three values: +1 if there is positive regulation (activation) from gene j to gene i; 1 if there is negative regulation (inhibition) from gene j to gene i; and 0 if gene i is not an input to gene j. The second set of parameters, biases (b i ), can take two values: +1/2 if gene i is positively biased, in the sense that an equal number of activation and inhibition inputs activate the gene; 1/2 if the reverse is true. Each element of state at time step k can be obtained by adding noise to the output of equation 2. The proposed model is depicted in figure 1 (where the input u k 1 is omitted). f where λ ki is the mean read count of transcript i at time k. Recall that, according to the Boolean state model, there are two possible states for the abundance of transcript i at time k: high (X ki = 1, active gene) and low (X ki = 0, inactive gene). Accordingly, the parameter λ ki is modeled as follows: log(λ ki ) = log(s) + µ b + θ i, if X ki = 0, log(λ ki ) = log(s) + µ b + δ i + θ i, if X ki = 1. where the parameter s is the sequencing depth, µ b > 0 is the baseline level of expression in the inactivated transcriptional state, δ i > 0 expresses the effect on the observed RNA-seq read count as gene i goes from the inactivated to the activated state, and θ i is a noise parameter that models unknown and unwanted technical effects that may occur during the experiment. We assume δ i to be Gaussian with mean µ δ > 0 and variance σδ 2, common to all transcripts, where σ δ is assumed to be small enough to keep δ i positive. Furthermore, we assume that θ i is zero-mean Gaussian noise with small variance σθ 2, common to all transcripts. Typical values for all these parameters are given in section IV. (5) III. BOOLEAN KALMAN SMOOTHER Fig. 1: Proposed network model (without inputs). B. Observation Model In most real-world applications, the system state is only partially observable, and distortion is introduced in the observations by sensor noise this is certainly the case with RNA-seq transcriptomic data. Let Y k be the observation corresponding to the state X k at time k. The observation Y k is formed from the state X k through the equation: Y k = h (X k 1, v k ) (observation model) (3) for k = 1, 2,...; where v k is observation noise. Assume Y k = (Y k1,..., Y kd ) is a vector containing the RNAseq data at time k, for k = 1, 2,... A single-lane NGS platform is considered here, in which Y ki is the read count corresponding to transcript i in the single lane, for i = 1, 2,... In this study, we choose to use a Poisson model for the number of reads for each transcript: P (Y ki = m λ ki ) = e λ λm ki ki, m = 0, 1,... (4) m! The optimal smoothing problem consists of, given a time point 1 < S < T and data in the interval {Y 1,..., Y T }, finding an estimator ˆX S = g(y 1,..., Y T ) of the Boolean state X S at time S that minimizes the mean-square error (MSE): MSE(Y 1,..., Y T ) = E [ ˆX S X S 2 Y 1,..., Y T ] (6) at each value of {Y 1,..., Y T }. A recursive algorithm to solve this problem, called the (fixed-interval) Boolean Kalman Smoother (BKS), is described next. Let (x 1,..., x 2d ) be an arbitrary enumeration of the possible state vectors. Define the following distribution vectors of length 2 d : Π k k (i) = P (X k = x i Y k,..., Y 1 ), Π k k 1 (i) = P (X k = x i Y k 1,..., Y 1 ), k k (i) = P (Y k+1,..., Y T X k = x i ), k k 1 (i) = P (Y k,..., Y T X k = x i ), (7) for i = 1,..., 2 d and k = 1,..., T, where T T = 1 d 1, by definition. Also define Π 0 0 to be the initial (prior) distribution of the states at time zero. Let the prediction matrix M k of size 2 d 2 d be the transition matrix of the Markov chain defined by the state

model: (M k ) ij = P (X k = x i X k 1 = x j ) = P (n k = x i f(x j, u k 1 )), (8) for i, j = 1,..., 2 d. Additionally, given a value of the observation vector Y k at time k, the update matrix T k, also of size 2 d 2 d, is a diagonal matrix defined by: (T k ) jj = p (Y k X k = x j ) (9) for j = 1,..., 2 d. In the specific case of RNA-seq data, considered in the previous section, we have: (T k ) jj = e ( d i=1 λji) d λ Y ki ji i=1 Y ki! d = s( i=1 Y ki) d d i=1 Y ki! exp s exp(µ b + θ i + δ i (x j ) i ) i=1 + (µ b + θ i + δ i (x j ) i )Y ki ), (10) for j = 1,..., 2 d. Finally, define the matrix A of size d 2 d via A = [x 1 x 2d ]. The following result, given here without proof, provides a recursive algorithm to compute the optimal MMSE state estimator. Theorem 1. (Boolean Kalman Smoother.) The optimal minimum MSE estimator ˆX S of the state X S given the observations Y 1,..., Y T, where 1 < S < T, is given by: ˆX S = E [X S Y 1,..., Y T ], (11) where v(i) = I v(i)>1/2 for i = 1,..., d. This estimator and its optimal MSE can be computed by the following procedure. Forward Estimator: 1) Initialization Step: Π 1 0 = M 1 Π 0 0. For k = 1, 2,..., S 1, do: 2) Update Step: β k = T k (y k ) Π k k 1. 3) Normalization Step: Π k k = β k / β k 1. 4) Prediction Step: Π k+1 k = M k+1 Π k k. Backward Estimator: 1) Initialization Step: T T 1 = T T (y T )1 d 1. For k = T 1, T 2,..., S, do: 2) Prediction Step: k k = M k+1 T k+1 k. 3) Update Step: k k 1 = T k (y k ) k k. Smoothed Distribution Vector: Π S S 1 S S 1 Π S T =, Π S S 1 S S 1 1 where denotes componentwise vector multiplication. MMSE Estimator: The MMSE estimator is given by: with optimal conditional MSE ˆX S = AΠ S T (12) MSE(Y 1,..., Y T ) = min{aπ S T, (AΠ S T ) c } 1, (13) where the minimum is applied component-wise, and v c (i) = 1 v(i), for i = 1,..., d. Estimation of the state at time k = T requires only the forward estimation step, in which case the Boolean Kalman Smoother reduces to the Boolean Kalman Filter (BKF), introduced in [3]. The normalization step in the forward estimator is not strictly necessary and can be skipped, by letting Π k k = β k (though in this case the meaning of the vectors Π k k and Π k+1 k change, of course). In addition, the matrix T k can be scaled at will. In particular, the constant s ( d i=1 Y ki) / d i=1 Y ki! in (10) can be dropped, which results in significant computational savings. IV. NUMERICAL EXPERIMENT In this section, we conduct a numerical experiment using a Boolean network based on the well-known pathway for the p53 MDM2 negative feedback system, shown in figure 2. We consider the input to be either no stress, dna dsb = 0 or DNA damage, dna dsb = 1, separately. The process noise is assumed to have independent Wip1 dna dsb ATM p53 Mdm2 Fig. 2: Activation/inactivation diagram for the p53 MDM2 negative feedback loop. components distributed as Bernoulli(p), where the noise parameter p gives the amount of perturbation to the Boolean state process; the process noise is categorized into two levels, p = 0.01 (small noise) and p = 0.1 (large noise). On the other hand, σθ 2 (see Section II-B) is the technical effect variance, which specifies the uncertainty of the observations. This value is likewise categorized into two levels: σθ 2 = 0.01 (clean observations) and

Average MSE σθ 2 = 0.1 (noisy observations). Two sequencing depths are considered for the observations in the simulations, s = 1.0175 and s = 2.875 which correspond to 1K 50K and 50K 100K reads respectively. The parameters δ i are generated from a Gaussian distribution with mean (µ δ ) 3 and variance (σδ 2 ) 0.5. The baseline expression µ b is set to 0.01. as the number of reads or the noise increases. In addition, when the input dna dsb is 0 (no DNA damage), the performance of both methods is better than when the input is 1 (DNA damage). The reason can be found by looking at the state transitions of p53 MDM2 network for both inputs. When the input is 0, the network has one singleton attractor ( 0000 ), while for input 1, there are two cyclic attractors, the first of which contains two states, while the other contains six states. In the presence of cyclic or multiple attractors, due to the changes of state trajectories, the estimation process is more difficult. TABLE I: Performance of the BKF and the BKS. Noise Parameters dna dsb = 0 dna dsb = 1 Reads BKF BKS BKF BKS p = 0.01 1K-50K 0.94 0.95 0.60 0.63 σθ 2 = 0.01 5K-100K 0.95 0.96 0.62 0.68 p = 0.1 1K-50K 0.74 0.76 0.59 0.61 σθ 2 = 0.01 5K-100K 0.75 0.76 0.59 0.62 p = 0.01 1K-50K 0.93 0.94 0.55 0.58 σθ 2 = 0.1 5K-100K 0.94 0.95 0.56 0.62 p = 0.1 1K-50K 0.71 0.72 0.54 0.56 σθ 2 = 0.1 5K-100K 0.71 0.73 0.54 0.57 V. CONCLUSION Time Fig. 3: Average MSE of BKF and BKS over 1000 runs. Figure 3 displays the average MSE achieved by the BKS at a fixed time point, as well as the BKF, for T = 100 observations, over 1000 independent runs. It is seen that BKS has smaller MSE on average in comparison to BKF as expected, since the BKS uses future observations, but the BKF uses only the observations up to the present time. Furthermore, we can see that the average MSE of both methods is higher in the presence of large noise. Next, the performance of the BKS and the BKF with different values of noise and different number of reads is examined. The average performance of the methods is defined here as the estimation error rate, i.e., the average number of correctly estimated states (over the length T = 100 of the signal and 1000 runs), which is presented in table I. The results show that the average performance of the BKS is higher than that of the BKF. Furthermore, the performance of both methods decreases This paper introduced a method for the inference of gene regulatory networks that is based on a novel algorithm, called the Boolean Kalman Smoother, which efficiently computes the optimal state estimator for discretetime Boolean dynamical systems given the entire history of observations. The smoothing process at each time step contains two estimators: the forward estimator, in which the previous observations are involved in the process, and the backward estimator, in which the future observations are employed in the process of estimation, in a process that bears similarities to the forward-backward algorithm commonly applied to the inference of hidden markov models [9]. The method was illustrated by application to the p53 MDM2 negative feedback network observed through next-generation sequencing data. The results indicate that on average, the BKS has lower MSE and lower error rates than the BKF. ACKNOWLEDGMENT The authors acknowledge the support of the National Science Foundation, through NSF award CCF-1320884.

REFERENCES [1] S. Kauffman, Metabolic stability and epigenesis in randomly constructed genetic nets, Journal of Theoretical Biology, vol. 22, pp. 437 467, 1969. [2] S. Kauffman, The Origins of Order: Self-Organization and Selection in Evolution. Oxford University Press, 1993. [3] U. Braga-Neto, Optimal state estimation for Boolean dynamical systems, 2011. Proceedings of 45th Annual Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA. [4] U. Braga-Neto, Joint state and parameter estimation for Boolean dynamical systems, 2012. Proceedings of the IEEE Statistical Signal Processing Workshop (SSP 12), Ann Arbor, MI. [5] U. Braga-Neto, Particle filtering approach to state estimation in boolean dynamical systems, 2013. Proceedings of the IEEE Global Conference on Signal and Image Processing (Global- SIP 13), Austin, TX. [6] A. Bahadorinejad and U. Braga-Neto, Optimal fault detection in stochastic boolean regulatory networks, 2014. Proceedings of the IEEE International Workshop on Genomic Signal Processing and Statistics (GENSIPS 2014), Atlanta, GA. [7] A. Bahadorinejad and U. Braga-Neto, Optimal fault detection and diagnosis in transcriptional circuits using next-generation sequencing, IEEE/ACM Transactions on Computational Biology and Bioinformatics, 2015. Preprint. [8] M. Imani and U. Braga-Neto, Optimal gene regulatory network inference using the boolean kalman filter and multiple model adaptive estimation, 2015. Proceedings of the 49th Annual Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA. [9] L. Rabiner, A tutorial on hidden markov models and selected applications in speech recognition, Proceedings of the IEEE, vol. 77, no. 2, pp. 257 286, 1989.