AMARKOV modulated Markov process (MMMP) is a

Size: px
Start display at page:

Download "AMARKOV modulated Markov process (MMMP) is a"

Transcription

1 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 57, NO. 2, FEBRUARY An EM Algorithm for Markov Modulated Markov Processes Yariv Ephraim, Fellow, IEEE, and William J. J. Roberts, Senior Member, IEEE Abstract An expectation maximization (EM) algorithm for estimating the parameter of a Markov modulated Markov process in the maximum likelihood sense is developed. This is a doubly stochastic random process with an underlying continuous-time finite-state homogeneous Markov chain. Conditioned on that chain, the observable process is a continuous-time finite-state nonhomogeneous Markov chain. The generator of the observable process at any given time is determined by the state of the underlying Markov chain at that time. The parameter of the process comprises the set of generators for the underlying and conditional Markov chains. The proposed approach generalizes an earlier approach by Rydén for estimating the parameter of a Markov modulated Poisson process. Index Terms Expectation-maximization (EM) algorithm, Markov modulated Markov processes, Markov modulated Poisson processes. I. INTRODUCTION AMARKOV modulated Markov process (MMMP) is a doubly stochastic random process comprising an observable process and an underlying hidden process. The underlying process is a continuous-time finite-state homogeneous Markov chain. Conditioned on that chain, the observable process is a continuous-time finite-state nonhomogeneous Markov chain. The generator of the observable process at any given time is determined by the state of the underlying Markov chain at that time. Thus, the number of generators characterizing the observable process equals the number of states of the underlying Markov chain. The parameter of the MMMP comprises the set of generators of the underlying and conditional Markov chains. The goal of this paper is to develop an expectation-maximization (EM) algorithm for maximum-likelihood (ML) estimation of the parameter of an MMMP from a given sample path of the observable process. MMMPs generalize the more familiar Markov modulated Poisson processes. The latter is a nonhomogeneous conditional Poisson process whose rate at any given time is determined by the state of an underlying continuous-time finite-state homogeneous Markov chain [11]. The generalization is in the sense that Manuscript received November 21, 2007; revised September 16, First published October 31, 2008; current version published January 30, The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Subhrakanti Dey. Y. Ephraim is with the Department of Electrical and Computer Engineering, George Mason University, Fairfax, VA USA ( yephraim@gmu. edu). W. J. J. Roberts is with Atlantic Coast Technologies, Inc., Silver Spring, MD USA ( robertsb@acoast.com). Digital Object Identifier /TSP increments of the conditional Markov chain are not independent as is the case with increments of the conditional Poisson process. Similarly to Markov modulated Poisson processes [22], [23], an MMMP may also be interpreted as a hidden Markov process comprising a discrete-time Markov chain and a sequence of conditionally independent observations. A survey of hidden Markov processes may be found in [10]. An MMMP may be viewed as a Markov renewal process whose dwell time in each state is not necessarily exponentially distributed [6]. In fact, that dwell time is a finite mixture of exponential densities, as is demonstrated in Section II by a simple example. Such finite mixtures may be used to approximate arbitrarily closely, in the Kullback Leibler sense, any desirable dwell time density in the family of continuous mixture densities [17]. Hence, MMMPs may be useful in hidden Markov modeling where explicit durational models are necessary. Examples include modeling of speech signals in automatic speech recognition applications [16], and hidden Markov modeling of hydrothermal flow episodes [27]. See also Yu and Kobayashi [31] and the references therein. MMMPs have been used in ion channel modeling as early as 1994, see, e.g., Ball, Milne and Yeo, [3], and subsequent papers by some of these authors. MMMPs have also been proposed for evaluating the likelihood of DNA sequences in molecular phylogenetics (see, e.g., [5] and the references therein). In both cases, no efficient algorithm for parameter estimation was proposed. We hope that the approach presented here would be useful for these important applications. Our approach for estimating the parameter of an MMMP was inspired by an EM algorithm developed by Rydén [24] for Markov modulated Poisson processes, and by an earlier EM approach developed by Asmussen and Nerman for estimating the parameter of a phase-type distribution, see Asmussen, Nerman and Olsson [2]. This distribution is that of a Markov chain with transient states and a single absorbing state. Of interest is the absorption time of the chain. The approach developed here uses the classical EM formulation of Dempster, Laird, and Rubin [8] for estimating the parameter of a process from a sequence of observations. This becomes possible, for example, when a continuous-time Markov chain is seen as a sequence of conditionally independent exponentially distributed random variables representing the dwell times of the chain in each state, together with a discrete time Markov chain representing state transitions at the jump times of the continuous-time chain. The likelihood function of a Markov chain represented in this way was derived independently by Billingsley [4] and Albert [1]. The complete data in this approach includes the entire sample path of the hidden chain, and the resulting -step and -step of the algorithm are explicit. The -step is implemented using forward-backward X/$ IEEE

2 464 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 57, NO. 2, FEBRUARY 2009 recursions similar to those associated with Markov modulated Poisson processes [24], [20]. An alternative approach to the same parameter estimation problem was taken by Elliott, Aggoun, and Moore [9]. They used the EM approach for continuous-time processes of Dembo and Zeitouni [7], [32], and developed the recursive estimators required to implement the -step. These estimators were derived using transformation of measures and the generalized Bayes rule or the Kallianpur Striebel formula [18, Lemma 7.4]. The recursions are given in terms of stochastic integrals as is usually the case with nonlinear estimation of continuous-time signals. Contrary to the EM approach of Rydén [24], the approach in [9] uses only forward recursions. Implementation of the approach in [9], however, requires solution of stochastic differential equations. We are not aware of any implementation of the approach in [9]. In Ledoux [15], the two approaches were developed and compared for the related Markovian arrival processes (MAPs) with hidden states. It was shown that the filter-based approach of [9] requires computations while the forward backward approach of [24] requires only computations. The application of Rydén s EM approach for MAPs and their batch extension was originally developed by Klemm, Lindemann, and Lohmann [14]. In yet another related work [29], the forward backward EM algorithm was applied to estimate the generator of a vector of two continuous-time Markov chains of which one is hidden. In [29], the transition matrix of the joint process was first estimated from sampled data. Then, the generator was estimated as the principal logarithm of the transition matrix estimate normalized by the sampling period. A thorough discussion of this approach was recently given in [21]. Due to the close relationship between Markov modulated Poisson processes and MMMPs, it is worthwhile mentioning that randomly modulated Poisson processes have attracted significant research interests since the 1970s (see, for example, Snyder [26] and Segall, Davis, and Kailath [25] and the references therein). In these works, the underlying process is rather general and need not be a finite state Markov or even a diffusion process. For this very general case, the modulated Poisson process can be represented as an increasing predictable process and an additive martingale noise process w.r.t. an appropriately defined sigma field. This representation is reminiscent of the classic representation of a signal observed in additive white noise. General finite-dimensional recursive filtering and smoothing formulas were developed, see, e.g., [25], for estimating the underlying process from the observed counts of the conditionally Poisson process. The plan for the remainder of this paper is as follows. In Section II, we derive some likelihood functions associated with the MMMP. In Section III, we develop the EM algorithm. In Section IV, we provide a numerical example. II. DENSITIES OF MMMP In this section, we set our notation, define the MMMP, and derive some densities associated with this process. In general, we use upper case letters to denote random variables and lower case letters to denote their realizations. We will not make this distinction when referring to sample paths of the process. Thus, for example, we may use to refer to the process itself or to a particular sample path of that process. We also use to denote either a probability measure or a density, as appropriate. These conventions greatly simplify the notation. We trust that the exact meaning of these symbols will always be clear from the context. Let denote the underlying process of the MMMP. It is assumed to be a separable continuous-time finite-state homogeneous irreducible Markov chain with state space of say. The order of the chain is assumed known. With probability one, all sample paths are right-continuous step functions with a finite number of jumps in each finite interval [1, Theorem 2.1]. A sample path of the chain in a given interval is characterized by the number of jumps, the sequence of visited states, and the time spent in each state within that interval. Assume that the chain jumps times in at, and define and. Let the initial state at time be denoted by, and let the state visited during be denoted by for. The time spent in state is. Let denote the initial distribution of the chain. Let the generator of the chain, where for denote Let denote the rate at which leaves state.we have When then is the stationary distribution of the chain [13, p. 261]. Similarly, let denote the observable process of the MMMP. Conditioned on, it is a separable continuous-time finite-state nonhomogeneous irreducible Markov chain with state space of say. The order is also assumed known. Assume that the process starts from some state at time, and that it jumps times in at. When, we denote.for, let denote the state of the process in the interval, and let denote the dwell time of the process in that state. The last state reached by the process is. When, the dwell time of the process in is. We use, and to denote realizations of, and, respectively. The initial distribution and generator of are controlled by in the following manner. When, the distribution of is given by. Similarly, when, the observable process transits at time according to the generator. Let denote the rate that leaves state while the underlying chain is in state. The set of generators of the two Markov chains,, constitutes the parameter of interest of the MMMP. Note that is a two-dimensional Markov process. The density of the observable process on the space of sample functions may be evaluated as follows. Assume that the process (1) (2)

3 EPHRAIM AND ROBERTS: AN EM ALGORITHM FOR MARKOV MODULATED MARKOV PROCESSES 465 jumps times in. Suppose first that the last jump occurs at. The desired density can then be obtained from the joint density of the random variables This density, as well as other densities of interest, will follow from the joint density of and the sequence of random variables ς. The latter random variables signify the states of the underlying chain at and at the jump times of the observable process. Let the augmented collection of random variables be denoted as ς (3) ς ς ς (4) Let and. The density of follows from the Markov property which implies conditional independence of and ς ς given ς. This density is therefore given by where for denotes a realization of ς, and. Let ς for. Rewriting (5) in terms of the Markov chain shows that is a Markov renewal process [6]. A. Transition Density Matrix An expression for the density in (5) can be obtained from a similar argument to that used for Markov modulated Poisson processes, see, e.g., Freed and Shepp [12]. Specifically, suppose that at time the underling Markov chain is in state and the observable process is in state. Recall that denotes the time of the first jump of the observable process. We are interested in the conditional probability of the event given for, and we denote the corresponding density by. Following an argument similar to that given in [12, (2.5)], we obtain where denotes the Kronecker delta. To interpret this expression, note that the first jump of the observable process from to at time may occur while the underlying Markov chain first jumps to some intermediate state at some random time, before it eventually reaches its final state (which may be equal to ) at time. The jump of the observable process may also occur while the underlying chain remains in during. In that case,. The density corresponding to (5) (6) the first alternative is given by the second term on the right-hand side of (6). This term comprises the product of the probability that the observable process does not jump in while the underlying chain remains in state, the exponential density for the dwell time of the underlying chain in state during, the probability of the jump from to at time, and the density corresponding to the jump time of the observable process from to and the jump of the underlying chain from to. This product is summed over all possible and integrated over. The first term on the right-hand side of (6) is relevant only when the underlying chain does not jump in. This term comprises the product of the probability that the underlying chain remains in state in, the exponential density of whose parameter is, and the probability of the jump of the observable process from to at time under the generator. Next, for, define the transition density matrix and the matrix The transition density matrix (7) is obtained by differentiating (6) w.r.t.. We find that Since, we have from (9) (7) (8) (9) (10) This expression is similar to [12, (2.6)] for Markov modulated Poisson processes. The density of can now be obtained from (5) and (10). Define.Wehave (11) where denotes an vector whose components are all equal to one. Note that evaluation of requires products of matrices. Furthermore, this approach does not provide an explicit form for since we only have the exponential matrix form (10). When, the density in (5) should be multiplied by an additional transition matrix which is derived as follows. Let (12) denote the transition probability of the underlying chain from state at time 0 to state at time while the observable process remains in state for at least seconds. Let denote the corresponding transition matrix. It follows that (5) must be multiplied by. The transition probability satisfies an equation similar to (6) with the exception that the first term on the right-hand side is replaced by. This change

4 466 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 57, NO. 2, FEBRUARY 2009 only affects the initial condition of the differential equation. Since,wehave (13) which is analogous to a result by Neuts [19] that was used for Markov modulated Poisson processes in [24]. Finally, the transition matrix of ς for given can be obtained from (10) and is given by (14) B. An Example To illustrate the evaluation and form of the transition density matrix (10), we consider a relatively simple MMMP with. This example is similar to that given by Rydén [22, Sec. 2.3] for switched Poisson processes, which are Markov modulated Poisson processes for which the order of the underlying Markov chain is equal to two. For this case, let (15), and for. To evaluate the transition density matrix (10) we note that is a simple matrix if its eigenvalues are distinct. As such, it can be written as, where and denote, respectively, a diagonal matrix of the eigenvalues and a nonsingular matrix of corresponding eigenvectors of. We express the eigenvalues in terms of (16) We have that and. The eigenvectors are given by for, and the determinant of is given by. Using these results in (10) we obtain (17), shown at the bottom of the page, which is analogous to the result of Rydén [22]. This example demonstrates nicely the nonexponential nature of the densities of the dwell time of the observable process in each state for a given transition of the underlying chain. In this example, each density is a mixture of two exponential densities. In general, we would have a mixture of exponential densities. III. THE EM ALGORITHM In this section, we describe an EM algorithm for ML estimation of the true parameter of the MMMP, which we denote by, from a sample path of the observable process. In the EM approach, a new parameter estimate, say, is obtained from a given parameter estimate, say,as (18) where is related to the underlying Markov chain, and the expectation is performed over given and the current parameter estimate. Starting from an initial estimate, say, the procedure generates a sequence of estimates with nondecreasing likelihood values. It is well known that every fixed point of the algorithm is a stationary point of the likelihood function [30]. Under certain conditions, there exists a subsequence of which converges to a stationary point of the likelihood function [30]. The expectation and maximization in (18) represent the - and -steps of the algorithm, respectively. The exact form of is not specified by the EM algorithm. This quantity can be conveniently chosen to simplify the estimation procedure. One choice could be the random variables ς used in (4). Unfortunately, this choice does not lead to explicit reestimation formula in each iteration for two reasons. First, we only have the transition density matrix in its exponential form rather than the individual densities for all possible values of. Second, the latter densities are mixtures of exponentials as was demonstrated in the Example in Section II. The logarithm of a mixture density does not simplify as in the case with densities from the exponential family. An alternative form of, proposed by Asmussen and Nerman (see [2]) and used by Rydén in the context of estimating the parameter of a Markov modulated Poisson processes [24], is that of being the entire sample path of the underlying chain in. This choice is useful in our case as well. Consider where is the entire sample path of the underlying chain. The density of was derived by Albert [1]. Since it also forms the basis for the density of, we begin with its description. A. Parameter of Underlying Chain Let be a random variable which represents the number of jumps of in with the last jump occurring prior to. Let denote a natural number which represents a value that may take. Recall that for denotes the time of the th jump of the chain. Using and, the state of the chain during is for. The process is characterized by (19) (17)

5 EPHRAIM AND ROBERTS: AN EM ALGORITHM FOR MARKOV MODULATED MARKOV PROCESSES 467 Albert [1] showed that the density corresponding to the probability measure is given by (20) (21) where for, and. Hence,.Anexpression for was given in [1] and [24], using the indicator function and the notation (22) Maximization of over and, using, results in the following intuitive estimates in the st iteration of the EM algorithm. (29) Thus, is the ratio of the conditional mean estimates of the number of jumps from to and the dwell time of the chain in state. Note that is not necessarily the stationary distribution of the chain. Albert [1] provided a similar estimate for when the chain is observable. The conditional probability in (25) (26), and the conditional density in (27), may be evaluated using forward-backward recursions in a similar way to that proposed by Rydén [24] for Markov modulated Poisson processes. To simplify the notation, we suppress the dependency on. We begin with (27). From Bayes rule (23) # (24) The reestimation formula for the parameter of the underlying chain is obtained from maximization of over. Let, and denote, respectively, the conditional mean estimates of, and given and. We have (25) (26) (30) Using the Markov property of conditional independence (31) Now, if, where is the time of the th jump of the observable process, and, then the last term of (31) multiplied by follows from the forward density given by (27) where denotes the conditional density of a jump from to at time. The expression for follows from a limit argument similar to that developed by Asmussen, Nerman and Olsson [2, Appendix] for phase-type distributions. The idea behind the proof is to consider a uniform partition of the time interval between consecutive jumps of the observable process into subintervals of length each. Then (28) Now, can be evaluated by i) applying Lebesgue s dominated convergence theorem which turns the indicator function in (28) into the conditional probability ; ii) using Bayes rule along with (31) (33) given below; and iii) using the identity. The proof also clarifies that, a result which will be required for evaluating (30) below. (32) where is a vector whose th component equals one and the remaining components are all equal to zero, and the transition density matrices in the right-hand side of (32) are given in (10) and (13), respectively. The expression in the right-hand side of (32) corresponds to rather than. As indicated in [24, p. 437] for Markov-modulated Poisson processes, this replacement has no effect on the density since the underlying Markov chain is continuous in probability. Next, when the last jump of the observable process, the first term of (31) follows from the backward density which is given by (33) When or when, these two expressions can be modified in a straightforward manner.

6 468 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 57, NO. 2, FEBRUARY 2009 The forward recursion is used to evaluate the bracketed term in (32). This recursion is given by Using the density of a Markov chain on the space of sample functions derived by Albert [1], it is straightforward to verify that the conditional density of the observable process in the th subinterval is given by (34) The backward recursion is used to evaluate the bracketed term in (33). This recursion is given by (38) where (35) A scaling procedure similar to that developed in [20] for Markov modulated Poisson processes may be applied here. Recursion (34) can be used to evaluate the likelihood of a sample path of the observable process when as. The conditional probability in (25) and (26) may be similarly evaluated as follows. From Bayes rule, the Markov property, and the continuity in probability of,wehave (36) denotes the time that the observable process spends in state, and (39) # (40) is the number of jumps of the observable process from to in. The last term in (38) is due to the inclusion of the probability of a possible jump of the observable process at from to say. Substituting (38) (40) into (37) and rearranging terms, we obtain in The two last terms can be evaluated using (32) and (33), respectively, and. B. Parameter of Observable Process We next derive the conditional density and the reestimation formula for the parameter of the observable process. We express the conditional density of the observable process for a sample path with jumps at. We denote and. Suppose that the underlying Markov chain jumped at, and let and. For, let and let. During the th subinterval, the underlying chain resides in state, and is a Markov chain with a fixed generator. Using the Markov property of,wehave where (41) (42) represents the total time that the observable process spends in state while the underlying chain resides in state in, (37) where for notational convenience we have used. The conditional density of given and is essentially given by an expression similar to (21). Here, however, the observable process may jump simultaneously with the underlying Markov chain at the beginning of each subinterval. Hence, we cannot exclude such jumps. (43)

7 EPHRAIM AND ROBERTS: AN EM ALGORITHM FOR MARKOV MODULATED MARKOV PROCESSES 469 where is the Dirac function, is the total number of jumps of the observable process from state to state while the underlying chain is in state in, and TABLE I MMMP EXAMPLE. TRUE PARAMETER ; INITIAL ESTIMATE ;FINAL ESTIMATE ^ (44) where mean over the generators of the observable process, taking into account that, we obtain (45) is the total number of jumps of the observable process from state to any other state while the underlying chain jumps from any state to state in. The conditional mean estimate is determined by the conditional mean estimates of the quantities defined in (42) (45). From (42) we have Similarly, from (43), we have (46) (47) (50) which can be interpreted similarly to in (29). The conditional probabilities in (46) (47) may be efficiently calculated as in (36) using the forward-backward recursions (34) (35). The algorithm presented here generalizes the EM algorithm for Markov modulated Poisson processes of Rydén [24] when the conditional Poisson process is seen as a conditional Markov process. To see that, it suffices to show that the likelihood of the conditional Markov process coincides with the likelihood of the conditional Poisson process in each subinterval. The result then follows from the fact that the conditional Poisson process has independent increments. Let denote the generator of the conditional Markov process at the given subinterval, and assume that the rate of the conditional Poisson process in the same subinterval is.wehave. Now, for every jump of the conditional Poisson process, it holds for the representing conditional Markov process that (51) From (45), we have (48) where contrary to (27), now represents the conditional probability of a jump from to at the known time, and that probability equals zero. Hence, (49) The conditional mean is now readily available from (41) and (46) (49). Maximizing this conditional independently of and for all. Suppose that there are Poisson events in. Substituting (51) in a density similar to that given in (21) but for the observable process, we obtain. This likelihood coincides with the corresponding th term in the second line of [24, (6)]. IV. NUMERICAL EXAMPLE The EM algorithm presented in Section III was implemented using Matlab. The forward-backward equations, (34) (35), were scaled to provide numerical stability using the approach developed in Roberts, Ephraim and Dieguez [20] for EM estimation of the parameter of a Markov modulated Poisson process. The integrals of matrix exponentials that appear in (26) (27) and in (46) were evaluated using the approach due to Van Loan [28] as was done in [20]. Matrix exponentials were calculated using the Matlab function expm which relies upon the Padé approximation with repeated squaring. The

8 470 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 57, NO. 2, FEBRUARY 2009 algorithm was terminated when the relative difference between log-likelihood values of successive iterations was. Our setup was motivated by Rydén s experiments on Markov modulated Poisson processes estimation [24]. Two different MMMPs were considered using the order of and for the underlying Markov chain, and the order of for the observable process. For the case, the generator of the underlying Markov chain and its initial estimate were identical to those of [24, case 2A]. As in [24, case 2A], 5000 inter-event times were used. The rates of the observable process and their initial estimates were similar to those used for the Poisson intensities in [24, case 2A]. For the case, the MMMP setup was similarly fashioned on [24, case 3A]. All Markov chains had uniform initial state distributions, which were assumed known by the algorithm. The MMMP parameter estimates were consistent with the Markov modulated Poisson process estimates obtained in [24]. We demonstrate the results in Table I for only. ACKNOWLEDGMENT The authors would like to thank the perceptive referees for their useful comments which greatly improved the presentation. REFERENCES [1] A. Albert, Estimating the infinitesimal generator of a continuous time, finite state Markov process, Ann. Math. Statist., vol. 23, no. 2, pp , Jun [2] S. Asmussen, O. Nerman, and M. Olsson, Fitting phase-type distributions via the EM algorithm., Scand. J. Statist., vol. 23, no. 4, pp , [3] F. Ball, R. K. Milne, and G. F. Yeo, Continuous-time Markov chains in a random environment, with applications to ion channel modelling, Adv. Appl. Probab., vol. 26, no. 4, pp , Dec [4] P. Billingsley, Statistical Inference for Markov Processes. Chicago, IL: Univ. of Chicago Press, [5] D. Bryant, N. Galtier, and M. A. Poursat, Likelihood calculation in molecular phylogenetics, in Mathematics of Evolution and Phylogeny, O. Gascuel, Ed. New York: Oxford, [6] E. Çinlar, Markov renewal theory: A survey, Manage. Sci., ser. Theory Series, vol. 21, no. 7, pp , Mar [7] A. Dembo and O. Zeitouni, Parameter estimation of partially observed continuous time stochastic processes via the EM algorithm, Stoch. Process. Appl., vol. 23, no. 1, pp , [8] A. P. Dempster, N. M. Laird, and D. B. Rubin, Maximum likelihood from incomplete data via the EM algorithm, J. Royal Statist. Soc. B, vol. 39, pp. 1 38, [9] R. J. Elliott, L. Aggoun, and J. B. Moore, Hidden Markov Models: Estimation and Control. New York: Springer-Verlag, [10] Y. Ephraim and N. Merhav, Hidden Markov processes, IEEE Trans. Inf. Theory, vol. 48, pp , Jun [11] W. Fischer and K. Meier-Hellstern, The Markov-modulated Poisson process (MMPP) cookbook, Perform. Eval., vol. 18, pp , [12] D. S. Freed and L. A. Shepp, A Poisson process whose rate is a hidden Markov chain, Adv. Appl. Probab., vol. 14, pp , [13] G. R. Grimmett and D. R. Stirzaker, Probability and Random Processes. Oxford, U.K.: Oxford Science, [14] A. Klemm, C. Lindemann, and M. Lohmann, Modeling IP traffic using the batch Markovian arrival process, Perform. Eval., vol. 54, pp , [15] J. Ledoux, Filtering and the EM-algorithm for the Markovian arrival process, Commun. Statist. Theory Methods, vol. 36, no. 14, pp , Jan [16] S. E. Levinson, Continuously variable duration hidden Markov models for automatic speech recognition, Comput. Speech Lang., vol. 1, pp , [17] J. Q. Li and A. R. Barron, Mixture density estimation, in Advances in Neural Information Processing Systems, S. A. Solla, T. K. Leen, and K.-R. Mueller, Eds. Cambridge, MA: MIT Press, 2000, vol. 12, pp [18] R. S. Liptser and A. N. Shiryaev, Statistics of Random Processes. New York: Springer-Verlag, 2004, pt. I. [19] M. F. Neuts, Structured Stochastic Matrices of M/G/1 Type and Their Applications. New York: Marcel Dekker, [20] W. J. J. Roberts, Y. Ephraim, and E. Dieguez, On Rydén EM algorithm for estimating MMPPs, IEEE Signal Process. Lett., vol. 13, no. 6, pp , Jun [21] W. J. J. Roberts and Y. Ephraim, An EM algorithm for ion-channel current estimation, IEEE Trans. Signal Process., vol. 56, no. 1, pp , Jan [22] T. Rydén, Parameter estimation for Markov modulated Poisson processes, Commun. Statist. Stoch. Models, vol. 10, no. 4, pp , [23] T. Rydén, Consistent and asymptotically normal parameter estimates for Markov modulated Poisson processes, Scand. J. Statist., vol. 22, pp , [24] T. Rydén, An EM algorithm for estimation in Markov-modulated Poisson processes, Comput. Statist. Data Anal., vol. 21, pp , [25] A. Segall, M. H. A. Davis, and T. Kailath, Nonlinear filtering with counting observations, IEEE Trans. Inf. Theory, vol. IT-21, pp , Mar [26] D. L. Snyder, Filtering and detection for doubly stochastic Poisson processes, IEEE Trans. Inf. Theory, vol. IT-18, pp , Jan [27] R. A. Sohn, Stochastic analysis of exit fluid temperature records from the active TAG hydrothermal mound (Mid-Atlantic Ridge 26 N): 2. Hidden Markov models of flow episodes, J. Geophys. Res., vol. 112, 2007, B [28] C. F. Van Loan, Computing integrals involving the matrix exponential, IEEE Trans. Autom. Control, vol. 23, no. 3, pp , [29] W. Wei, B. Wang, and D. Towsley, Continuous-time hidden Markov models for network performance evaluation, Perform. Eval., vol. 49, pp , [30] C. F. J. Wu, On the convergence properties of the EM algorithm, Ann. Statist., vol. 11, no. 1, pp , [31] S.-Z. Yu and H. Kobayashi, Practical implementation of an efficient forward-backward algorithm for an explicit-duration hidden Markov model, IEEE Trans. Signal Process., vol. 54, no. 5, pp , May [32] O. Zeitouni and A. Dembo, Exact filters for the estimation of the number of transitions of finite-state continuous-time Markov processes, IEEE Trans. Inf. Theory, vol. 34, pp , Jul Yariv Ephraim (S 82 M 84 SM 90 F 94) received the D.Sc. degree in electrical engineering from the Technion Israel Institute of Technology, Haifa, Israel, in During , he was a Rothschild Postdoctoral Fellow at the Information Systems Laboratory of Stanford University, Stanford, CA. From 1985 to 1993, he was a Member of Technical Staff at the Information Principles Research Laboratory of AT&T Bell Laboratories, Murray Hill, NJ. In 1991, he joined George Mason University, Fairfax, VA, where he currently is Professor of Electrical and Computer Engineering. His research interest are in statistical signal processing. William J. J. Roberts (S 89 M 90 SM 06) received the Ph.D. degree in information technology from George Mason University, Fairfax, VA, in From 1990 to 2000, he was with the Defence Science Technology Organization, Salisbury, South Australia. From 1998 to 1999, he held a postdoctoral position at the Tokyo Institute of Technology, Tokyo, Japan. Since 2000, he has been with Atlantic Coast Technologies, Inc., Silver Spring, MD. His interests are in statistical signal processing.

On modeling network congestion using continuous-time bivariate Markov chains

On modeling network congestion using continuous-time bivariate Markov chains On modeling networ congestion using continuous-time bivariate Marov chains Brian L. Mar and Yariv Ephraim Dept. of Electrical and Computer Engineering George Mason University, MS 1G5 44 University Drive,

More information

An EM algorithm for Batch Markovian Arrival Processes and its comparison to a simpler estimation procedure

An EM algorithm for Batch Markovian Arrival Processes and its comparison to a simpler estimation procedure An EM algorithm for Batch Markovian Arrival Processes and its comparison to a simpler estimation procedure Lothar Breuer (breuer@info04.uni-trier.de) University of Trier, Germany Abstract. Although the

More information

A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models

A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes (bilmes@cs.berkeley.edu) International Computer Science Institute

More information

HIDDEN MARKOV CHANGE POINT ESTIMATION

HIDDEN MARKOV CHANGE POINT ESTIMATION Communications on Stochastic Analysis Vol. 9, No. 3 (2015 367-374 Serials Publications www.serialspublications.com HIDDEN MARKOV CHANGE POINT ESTIMATION ROBERT J. ELLIOTT* AND SEBASTIAN ELLIOTT Abstract.

More information

MANY papers and books are devoted to modeling fading

MANY papers and books are devoted to modeling fading IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 16, NO. 9, DECEMBER 1998 1809 Hidden Markov Modeling of Flat Fading Channels William Turin, Senior Member, IEEE, Robert van Nobelen Abstract Hidden

More information

N.G.Bean, D.A.Green and P.G.Taylor. University of Adelaide. Adelaide. Abstract. process of an MMPP/M/1 queue is not a MAP unless the queue is a

N.G.Bean, D.A.Green and P.G.Taylor. University of Adelaide. Adelaide. Abstract. process of an MMPP/M/1 queue is not a MAP unless the queue is a WHEN IS A MAP POISSON N.G.Bean, D.A.Green and P.G.Taylor Department of Applied Mathematics University of Adelaide Adelaide 55 Abstract In a recent paper, Olivier and Walrand (994) claimed that the departure

More information

An Alternative Proof for the Capacity Region of the Degraded Gaussian MIMO Broadcast Channel

An Alternative Proof for the Capacity Region of the Degraded Gaussian MIMO Broadcast Channel IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 58, NO. 4, APRIL 2012 2427 An Alternative Proof for the Capacity Region of the Degraded Gaussian MIMO Broadcast Channel Ersen Ekrem, Student Member, IEEE,

More information

THERE has been a growing interest in using switching diffusion

THERE has been a growing interest in using switching diffusion 1706 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 53, NO. 5, MAY 2007 Two-Time-Scale Approximation for Wonham Filters Qing Zhang, Senior Member, IEEE, G. George Yin, Fellow, IEEE, Johb B. Moore, Life

More information

THE VARIANCE CONSTANT FOR THE ACTUAL WAITING TIME OF THE PH/PH/1 QUEUE. By Mogens Bladt National University of Mexico

THE VARIANCE CONSTANT FOR THE ACTUAL WAITING TIME OF THE PH/PH/1 QUEUE. By Mogens Bladt National University of Mexico The Annals of Applied Probability 1996, Vol. 6, No. 3, 766 777 THE VARIANCE CONSTANT FOR THE ACTUAL WAITING TIME OF THE PH/PH/1 QUEUE By Mogens Bladt National University of Mexico In this paper we consider

More information

Bisection Ideas in End-Point Conditioned Markov Process Simulation

Bisection Ideas in End-Point Conditioned Markov Process Simulation Bisection Ideas in End-Point Conditioned Markov Process Simulation Søren Asmussen and Asger Hobolth Department of Mathematical Sciences, Aarhus University Ny Munkegade, 8000 Aarhus C, Denmark {asmus,asger}@imf.au.dk

More information

Optimal Decentralized Control of Coupled Subsystems With Control Sharing

Optimal Decentralized Control of Coupled Subsystems With Control Sharing IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 58, NO. 9, SEPTEMBER 2013 2377 Optimal Decentralized Control of Coupled Subsystems With Control Sharing Aditya Mahajan, Member, IEEE Abstract Subsystems that

More information

HOPFIELD neural networks (HNNs) are a class of nonlinear

HOPFIELD neural networks (HNNs) are a class of nonlinear IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II: EXPRESS BRIEFS, VOL. 52, NO. 4, APRIL 2005 213 Stochastic Noise Process Enhancement of Hopfield Neural Networks Vladimir Pavlović, Member, IEEE, Dan Schonfeld,

More information

STA 414/2104: Machine Learning

STA 414/2104: Machine Learning STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far

More information

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 55, NO. 3, MARCH

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 55, NO. 3, MARCH IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL 55, NO 3, MARCH 2007 809 Probability of Error Analysis for Hidden Markov Model Filtering With Random Packet Loss Alex S C Leong, Member, IEEE, Subhrakanti Dey,

More information

Markov modulated Poisson process associated with state-dependent marks and its applications to the deep earthquakes

Markov modulated Poisson process associated with state-dependent marks and its applications to the deep earthquakes Ann Inst Stat Math (2012) 64:87 106 DOI 10.1007/s10463-010-0302-9 Markov modulated Poisson process associated with state-dependent marks and its applications to the deep earthquakes Shaochuan Lu Received:

More information

ABC methods for phase-type distributions with applications in insurance risk problems

ABC methods for phase-type distributions with applications in insurance risk problems ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Finding a Needle in a Haystack: Conditions for Reliable. Detection in the Presence of Clutter

Finding a Needle in a Haystack: Conditions for Reliable. Detection in the Presence of Clutter Finding a eedle in a Haystack: Conditions for Reliable Detection in the Presence of Clutter Bruno Jedynak and Damianos Karakos October 23, 2006 Abstract We study conditions for the detection of an -length

More information

THE capacity of the exponential server timing channel

THE capacity of the exponential server timing channel IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006 2697 Capacity of Queues Via Point-Process Channels Rajesh Sundaresan, Senior Member, IEEE, and Sergio Verdú, Fellow, IEEE Abstract A conceptually

More information

Reading: Karlin and Taylor Ch. 5 Resnick Ch. 3. A renewal process is a generalization of the Poisson point process.

Reading: Karlin and Taylor Ch. 5 Resnick Ch. 3. A renewal process is a generalization of the Poisson point process. Renewal Processes Wednesday, December 16, 2015 1:02 PM Reading: Karlin and Taylor Ch. 5 Resnick Ch. 3 A renewal process is a generalization of the Poisson point process. The Poisson point process is completely

More information

WE study the capacity of peak-power limited, single-antenna,

WE study the capacity of peak-power limited, single-antenna, 1158 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 3, MARCH 2010 Gaussian Fading Is the Worst Fading Tobias Koch, Member, IEEE, and Amos Lapidoth, Fellow, IEEE Abstract The capacity of peak-power

More information

SOME fundamental relationships between input output mutual

SOME fundamental relationships between input output mutual IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 54, NO. 5, MAY 2008 1837 Mutual Information and Conditional Mean Estimation in Poisson Channels Dongning Guo, Member, IEEE, Shlomo Shamai (Shitz), Fellow,

More information

Using Expectation-Maximization for Reinforcement Learning

Using Expectation-Maximization for Reinforcement Learning NOTE Communicated by Andrew Barto and Michael Jordan Using Expectation-Maximization for Reinforcement Learning Peter Dayan Department of Brain and Cognitive Sciences, Center for Biological and Computational

More information

A Modified Baum Welch Algorithm for Hidden Markov Models with Multiple Observation Spaces

A Modified Baum Welch Algorithm for Hidden Markov Models with Multiple Observation Spaces IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 9, NO. 4, MAY 2001 411 A Modified Baum Welch Algorithm for Hidden Markov Models with Multiple Observation Spaces Paul M. Baggenstoss, Member, IEEE

More information

An Efficient Approach to Multivariate Nakagami-m Distribution Using Green s Matrix Approximation

An Efficient Approach to Multivariate Nakagami-m Distribution Using Green s Matrix Approximation IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, VOL 2, NO 5, SEPTEMBER 2003 883 An Efficient Approach to Multivariate Nakagami-m Distribution Using Green s Matrix Approximation George K Karagiannidis, Member,

More information

Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE

Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 55, NO. 9, SEPTEMBER 2010 1987 Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE Abstract

More information

9 Multi-Model State Estimation

9 Multi-Model State Estimation Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 9 Multi-Model State

More information

ring structure Abstract Optical Grid networks allow many computing sites to share their resources by connecting

ring structure Abstract Optical Grid networks allow many computing sites to share their resources by connecting Markovian approximations for a grid computing network with a ring structure J. F. Pérez and B. Van Houdt Performance Analysis of Telecommunication Systems Research Group, Department of Mathematics and

More information

THE HEAVY-TRAFFIC BOTTLENECK PHENOMENON IN OPEN QUEUEING NETWORKS. S. Suresh and W. Whitt AT&T Bell Laboratories Murray Hill, New Jersey 07974

THE HEAVY-TRAFFIC BOTTLENECK PHENOMENON IN OPEN QUEUEING NETWORKS. S. Suresh and W. Whitt AT&T Bell Laboratories Murray Hill, New Jersey 07974 THE HEAVY-TRAFFIC BOTTLENECK PHENOMENON IN OPEN QUEUEING NETWORKS by S. Suresh and W. Whitt AT&T Bell Laboratories Murray Hill, New Jersey 07974 ABSTRACT This note describes a simulation experiment involving

More information

Multiscale Systems Engineering Research Group

Multiscale Systems Engineering Research Group Hidden Markov Model Prof. Yan Wang Woodruff School of Mechanical Engineering Georgia Institute of echnology Atlanta, GA 30332, U.S.A. yan.wang@me.gatech.edu Learning Objectives o familiarize the hidden

More information

ABSTRACT INTRODUCTION

ABSTRACT INTRODUCTION ABSTRACT Presented in this paper is an approach to fault diagnosis based on a unifying review of linear Gaussian models. The unifying review draws together different algorithms such as PCA, factor analysis,

More information

IN THIS PAPER, we consider a class of continuous-time recurrent

IN THIS PAPER, we consider a class of continuous-time recurrent IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II: EXPRESS BRIEFS, VOL. 51, NO. 4, APRIL 2004 161 Global Output Convergence of a Class of Continuous-Time Recurrent Neural Networks With Time-Varying Thresholds

More information

HEAVY-TRAFFIC EXTREME-VALUE LIMITS FOR QUEUES

HEAVY-TRAFFIC EXTREME-VALUE LIMITS FOR QUEUES HEAVY-TRAFFIC EXTREME-VALUE LIMITS FOR QUEUES by Peter W. Glynn Department of Operations Research Stanford University Stanford, CA 94305-4022 and Ward Whitt AT&T Bell Laboratories Murray Hill, NJ 07974-0636

More information

IN this paper, we consider the capacity of sticky channels, a

IN this paper, we consider the capacity of sticky channels, a 72 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 54, NO. 1, JANUARY 2008 Capacity Bounds for Sticky Channels Michael Mitzenmacher, Member, IEEE Abstract The capacity of sticky channels, a subclass of insertion

More information

Stochastic Processes

Stochastic Processes Stochastic Processes 8.445 MIT, fall 20 Mid Term Exam Solutions October 27, 20 Your Name: Alberto De Sole Exercise Max Grade Grade 5 5 2 5 5 3 5 5 4 5 5 5 5 5 6 5 5 Total 30 30 Problem :. True / False

More information

PARAMETER CONVERGENCE FOR EM AND MM ALGORITHMS

PARAMETER CONVERGENCE FOR EM AND MM ALGORITHMS Statistica Sinica 15(2005), 831-840 PARAMETER CONVERGENCE FOR EM AND MM ALGORITHMS Florin Vaida University of California at San Diego Abstract: It is well known that the likelihood sequence of the EM algorithm

More information

Optimum Sampling Vectors for Wiener Filter Noise Reduction

Optimum Sampling Vectors for Wiener Filter Noise Reduction 58 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 1, JANUARY 2002 Optimum Sampling Vectors for Wiener Filter Noise Reduction Yukihiko Yamashita, Member, IEEE Absact Sampling is a very important and

More information

Filter Design for Linear Time Delay Systems

Filter Design for Linear Time Delay Systems IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 11, NOVEMBER 2001 2839 ANewH Filter Design for Linear Time Delay Systems E. Fridman Uri Shaked, Fellow, IEEE Abstract A new delay-dependent filtering

More information

Markov Chains and Hidden Markov Models

Markov Chains and Hidden Markov Models Chapter 1 Markov Chains and Hidden Markov Models In this chapter, we will introduce the concept of Markov chains, and show how Markov chains can be used to model signals using structures such as hidden

More information

Computing Consecutive-Type Reliabilities Non-Recursively

Computing Consecutive-Type Reliabilities Non-Recursively IEEE TRANSACTIONS ON RELIABILITY, VOL. 52, NO. 3, SEPTEMBER 2003 367 Computing Consecutive-Type Reliabilities Non-Recursively Galit Shmueli Abstract The reliability of consecutive-type systems has been

More information

Variational Principal Components

Variational Principal Components Variational Principal Components Christopher M. Bishop Microsoft Research 7 J. J. Thomson Avenue, Cambridge, CB3 0FB, U.K. cmbishop@microsoft.com http://research.microsoft.com/ cmbishop In Proceedings

More information

The Expectation Maximization Algorithm

The Expectation Maximization Algorithm The Expectation Maximization Algorithm Frank Dellaert College of Computing, Georgia Institute of Technology Technical Report number GIT-GVU-- February Abstract This note represents my attempt at explaining

More information

output dimension input dimension Gaussian evidence Gaussian Gaussian evidence evidence from t +1 inputs and outputs at time t x t+2 x t-1 x t+1

output dimension input dimension Gaussian evidence Gaussian Gaussian evidence evidence from t +1 inputs and outputs at time t x t+2 x t-1 x t+1 To appear in M. S. Kearns, S. A. Solla, D. A. Cohn, (eds.) Advances in Neural Information Processing Systems. Cambridge, MA: MIT Press, 999. Learning Nonlinear Dynamical Systems using an EM Algorithm Zoubin

More information

Operations Research Letters. Instability of FIFO in a simple queueing system with arbitrarily low loads

Operations Research Letters. Instability of FIFO in a simple queueing system with arbitrarily low loads Operations Research Letters 37 (2009) 312 316 Contents lists available at ScienceDirect Operations Research Letters journal homepage: www.elsevier.com/locate/orl Instability of FIFO in a simple queueing

More information

Unsupervised Learning with Permuted Data

Unsupervised Learning with Permuted Data Unsupervised Learning with Permuted Data Sergey Kirshner skirshne@ics.uci.edu Sridevi Parise sparise@ics.uci.edu Padhraic Smyth smyth@ics.uci.edu School of Information and Computer Science, University

More information

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing Advances in Dynamical Systems and Applications ISSN 0973-5321, Volume 8, Number 2, pp. 401 412 (2013) http://campus.mst.edu/adsa Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes

More information

Basic math for biology

Basic math for biology Basic math for biology Lei Li Florida State University, Feb 6, 2002 The EM algorithm: setup Parametric models: {P θ }. Data: full data (Y, X); partial data Y. Missing data: X. Likelihood and maximum likelihood

More information

CEPSTRAL analysis has been widely used in signal processing

CEPSTRAL analysis has been widely used in signal processing 162 IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 7, NO. 2, MARCH 1999 On Second-Order Statistics and Linear Estimation of Cepstral Coefficients Yariv Ephraim, Fellow, IEEE, and Mazin Rahim, Senior

More information

STOCHASTIC MODELING OF ENVIRONMENTAL TIME SERIES. Richard W. Katz LECTURE 5

STOCHASTIC MODELING OF ENVIRONMENTAL TIME SERIES. Richard W. Katz LECTURE 5 STOCHASTIC MODELING OF ENVIRONMENTAL TIME SERIES Richard W Katz LECTURE 5 (1) Hidden Markov Models: Applications (2) Hidden Markov Models: Viterbi Algorithm (3) Non-Homogeneous Hidden Markov Model (1)

More information

Stabilization and Controllability for the Transmission Wave Equation

Stabilization and Controllability for the Transmission Wave Equation 1900 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 46, NO. 12, DECEMBER 2001 Stabilization Controllability for the Transmission Wave Equation Weijiu Liu Abstract In this paper, we address the problem of

More information

On Lossless Coding With Coded Side Information Daniel Marco, Member, IEEE, and Michelle Effros, Fellow, IEEE

On Lossless Coding With Coded Side Information Daniel Marco, Member, IEEE, and Michelle Effros, Fellow, IEEE 3284 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 7, JULY 2009 On Lossless Coding With Coded Side Information Daniel Marco, Member, IEEE, Michelle Effros, Fellow, IEEE Abstract This paper considers

More information

Statistics 992 Continuous-time Markov Chains Spring 2004

Statistics 992 Continuous-time Markov Chains Spring 2004 Summary Continuous-time finite-state-space Markov chains are stochastic processes that are widely used to model the process of nucleotide substitution. This chapter aims to present much of the mathematics

More information

O 3 O 4 O 5. q 3. q 4. Transition

O 3 O 4 O 5. q 3. q 4. Transition Hidden Markov Models Hidden Markov models (HMM) were developed in the early part of the 1970 s and at that time mostly applied in the area of computerized speech recognition. They are first described in

More information

Hidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010

Hidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010 Hidden Markov Models Aarti Singh Slides courtesy: Eric Xing Machine Learning 10-701/15-781 Nov 8, 2010 i.i.d to sequential data So far we assumed independent, identically distributed data Sequential data

More information

On the Multivariate Nakagami-m Distribution With Exponential Correlation

On the Multivariate Nakagami-m Distribution With Exponential Correlation 1240 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 51, NO. 8, AUGUST 2003 On the Multivariate Nakagami-m Distribution With Exponential Correlation George K. Karagiannidis, Member, IEEE, Dimitris A. Zogas,

More information

Stochastic Proximal Gradient Algorithm

Stochastic Proximal Gradient Algorithm Stochastic Institut Mines-Télécom / Telecom ParisTech / Laboratoire Traitement et Communication de l Information Joint work with: Y. Atchade, Ann Arbor, USA, G. Fort LTCI/Télécom Paristech and the kind

More information

Recursive Parameter Estimation for Continuous-time Bivariate Markov Chains

Recursive Parameter Estimation for Continuous-time Bivariate Markov Chains Recursive Parameter Estimation for Continuous-time Bivariate Markov Chains A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science at George Mason University By

More information

Towards a Bayesian model for Cyber Security

Towards a Bayesian model for Cyber Security Towards a Bayesian model for Cyber Security Mark Briers (mbriers@turing.ac.uk) Joint work with Henry Clausen and Prof. Niall Adams (Imperial College London) 27 September 2017 The Alan Turing Institute

More information

Convergence Rate of Expectation-Maximization

Convergence Rate of Expectation-Maximization Convergence Rate of Expectation-Maximiation Raunak Kumar University of British Columbia Mark Schmidt University of British Columbia Abstract raunakkumar17@outlookcom schmidtm@csubcca Expectation-maximiation

More information

Riccati difference equations to non linear extended Kalman filter constraints

Riccati difference equations to non linear extended Kalman filter constraints International Journal of Scientific & Engineering Research Volume 3, Issue 12, December-2012 1 Riccati difference equations to non linear extended Kalman filter constraints Abstract Elizabeth.S 1 & Jothilakshmi.R

More information

Title without the persistently exciting c. works must be obtained from the IEE

Title without the persistently exciting c.   works must be obtained from the IEE Title Exact convergence analysis of adapt without the persistently exciting c Author(s) Sakai, H; Yang, JM; Oka, T Citation IEEE TRANSACTIONS ON SIGNAL 55(5): 2077-2083 PROCESS Issue Date 2007-05 URL http://hdl.handle.net/2433/50544

More information

PROBABILISTIC METHODS FOR A LINEAR REACTION-HYPERBOLIC SYSTEM WITH CONSTANT COEFFICIENTS

PROBABILISTIC METHODS FOR A LINEAR REACTION-HYPERBOLIC SYSTEM WITH CONSTANT COEFFICIENTS The Annals of Applied Probability 1999, Vol. 9, No. 3, 719 731 PROBABILISTIC METHODS FOR A LINEAR REACTION-HYPERBOLIC SYSTEM WITH CONSTANT COEFFICIENTS By Elizabeth A. Brooks National Institute of Environmental

More information

Semi-Markov Decision Problems and Performance Sensitivity Analysis

Semi-Markov Decision Problems and Performance Sensitivity Analysis 758 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 48, NO. 5, MAY 2003 Semi-Markov Decision Problems Performance Sensitivity Analysis Xi-Ren Cao, Fellow, IEEE Abstract Recent research indicates that Markov

More information

EXPLICIT DISTRIBUTIONAL RESULTS IN PATTERN FORMATION. By V. T. Stefanov 1 and A. G. Pakes University of Western Australia

EXPLICIT DISTRIBUTIONAL RESULTS IN PATTERN FORMATION. By V. T. Stefanov 1 and A. G. Pakes University of Western Australia The Annals of Applied Probability 1997, Vol. 7, No. 3, 666 678 EXPLICIT DISTRIBUTIONAL RESULTS IN PATTERN FORMATION By V. T. Stefanov 1 and A. G. Pakes University of Western Australia A new and unified

More information

Chapter 2 Event-Triggered Sampling

Chapter 2 Event-Triggered Sampling Chapter Event-Triggered Sampling In this chapter, some general ideas and basic results on event-triggered sampling are introduced. The process considered is described by a first-order stochastic differential

More information

CHUN-HUA GUO. Key words. matrix equations, minimal nonnegative solution, Markov chains, cyclic reduction, iterative methods, convergence rate

CHUN-HUA GUO. Key words. matrix equations, minimal nonnegative solution, Markov chains, cyclic reduction, iterative methods, convergence rate CONVERGENCE ANALYSIS OF THE LATOUCHE-RAMASWAMI ALGORITHM FOR NULL RECURRENT QUASI-BIRTH-DEATH PROCESSES CHUN-HUA GUO Abstract The minimal nonnegative solution G of the matrix equation G = A 0 + A 1 G +

More information

SOLVING an optimization problem over two variables in a

SOLVING an optimization problem over two variables in a IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 3, MARCH 2009 1423 Adaptive Alternating Minimization Algorithms Urs Niesen, Student Member, IEEE, Devavrat Shah, and Gregory W. Wornell Abstract The

More information

Phase-Type Models of Channel-Holding Times in Cellular Communication Systems

Phase-Type Models of Channel-Holding Times in Cellular Communication Systems Downloaded from orbit.dtu.dk on: Dec 24, 2017 Phase-Type Models of Channel-Holding Times in Cellular Communication Systems Christensen, Thomas Kaare; Nielsen, Bo Friis; Iversen, Villy Bæk Published in:

More information

Learning Binary Classifiers for Multi-Class Problem

Learning Binary Classifiers for Multi-Class Problem Research Memorandum No. 1010 September 28, 2006 Learning Binary Classifiers for Multi-Class Problem Shiro Ikeda The Institute of Statistical Mathematics 4-6-7 Minami-Azabu, Minato-ku, Tokyo, 106-8569,

More information

Maximum Process Problems in Optimal Control Theory

Maximum Process Problems in Optimal Control Theory J. Appl. Math. Stochastic Anal. Vol. 25, No., 25, (77-88) Research Report No. 423, 2, Dept. Theoret. Statist. Aarhus (2 pp) Maximum Process Problems in Optimal Control Theory GORAN PESKIR 3 Given a standard

More information

Asymptotic Filtering and Entropy Rate of a Hidden Markov Process in the Rare Transitions Regime

Asymptotic Filtering and Entropy Rate of a Hidden Markov Process in the Rare Transitions Regime Asymptotic Filtering and Entropy Rate of a Hidden Marov Process in the Rare Transitions Regime Chandra Nair Dept. of Elect. Engg. Stanford University Stanford CA 94305, USA mchandra@stanford.edu Eri Ordentlich

More information

stochnotes Page 1

stochnotes Page 1 stochnotes110308 Page 1 Kolmogorov forward and backward equations and Poisson process Monday, November 03, 2008 11:58 AM How can we apply the Kolmogorov equations to calculate various statistics of interest?

More information

144 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 17, NO. 1, FEBRUARY A PDF f (x) is completely monotone if derivatives f of all orders exist

144 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 17, NO. 1, FEBRUARY A PDF f (x) is completely monotone if derivatives f of all orders exist 144 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 17, NO. 1, FEBRUARY 2009 Node Isolation Model and Age-Based Neighbor Selection in Unstructured P2P Networks Zhongmei Yao, Student Member, IEEE, Xiaoming Wang,

More information

Hidden Markov models for time series of counts with excess zeros

Hidden Markov models for time series of counts with excess zeros Hidden Markov models for time series of counts with excess zeros Madalina Olteanu and James Ridgway University Paris 1 Pantheon-Sorbonne - SAMM, EA4543 90 Rue de Tolbiac, 75013 Paris - France Abstract.

More information

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 50, NO. 5, MAY Bo Yang, Student Member, IEEE, and Wei Lin, Senior Member, IEEE (1.

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 50, NO. 5, MAY Bo Yang, Student Member, IEEE, and Wei Lin, Senior Member, IEEE (1. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL 50, NO 5, MAY 2005 619 Robust Output Feedback Stabilization of Uncertain Nonlinear Systems With Uncontrollable and Unobservable Linearization Bo Yang, Student

More information

Hidden Markov Models. By Parisa Abedi. Slides courtesy: Eric Xing

Hidden Markov Models. By Parisa Abedi. Slides courtesy: Eric Xing Hidden Markov Models By Parisa Abedi Slides courtesy: Eric Xing i.i.d to sequential data So far we assumed independent, identically distributed data Sequential (non i.i.d.) data Time-series data E.g. Speech

More information

Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach

Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach By Shiqing Ling Department of Mathematics Hong Kong University of Science and Technology Let {y t : t = 0, ±1, ±2,

More information

Expressions for the covariance matrix of covariance data

Expressions for the covariance matrix of covariance data Expressions for the covariance matrix of covariance data Torsten Söderström Division of Systems and Control, Department of Information Technology, Uppsala University, P O Box 337, SE-7505 Uppsala, Sweden

More information

DURING THE last two decades, many authors have

DURING THE last two decades, many authors have 614 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I: FUNDAMENTAL THEORY AND APPLICATIONS, VOL 44, NO 7, JULY 1997 Stability the Lyapunov Equation for -Dimensional Digital Systems Chengshan Xiao, David J Hill,

More information

Modelling Complex Queuing Situations with Markov Processes

Modelling Complex Queuing Situations with Markov Processes Modelling Complex Queuing Situations with Markov Processes Jason Randal Thorne, School of IT, Charles Sturt Uni, NSW 2795, Australia Abstract This article comments upon some new developments in the field

More information

A Complete Stability Analysis of Planar Discrete-Time Linear Systems Under Saturation

A Complete Stability Analysis of Planar Discrete-Time Linear Systems Under Saturation 710 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I: FUNDAMENTAL THEORY AND APPLICATIONS, VOL 48, NO 6, JUNE 2001 A Complete Stability Analysis of Planar Discrete-Time Linear Systems Under Saturation Tingshu

More information

Comments and Corrections

Comments and Corrections 1386 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL 59, NO 5, MAY 2014 Comments and Corrections Corrections to Stochastic Barbalat s Lemma and Its Applications Xin Yu and Zhaojing Wu Abstract The proof of

More information

Stochastic Sampling Algorithms for State Estimation of Jump Markov Linear Systems

Stochastic Sampling Algorithms for State Estimation of Jump Markov Linear Systems 188 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 45, NO. 1, JANUARY 2000 Stochastic Sampling Algorithms for State Estimation of Jump Markov Linear Systems Arnaud Doucet, Andrew Logothetis, Vikram Krishnamurthy

More information

WAITING-TIME DISTRIBUTION FOR THE r th OCCURRENCE OF A COMPOUND PATTERN IN HIGHER-ORDER MARKOVIAN SEQUENCES

WAITING-TIME DISTRIBUTION FOR THE r th OCCURRENCE OF A COMPOUND PATTERN IN HIGHER-ORDER MARKOVIAN SEQUENCES WAITING-TIME DISTRIBUTION FOR THE r th OCCURRENCE OF A COMPOUND PATTERN IN HIGHER-ORDER MARKOVIAN SEQUENCES Donald E. K. Martin 1 and John A. D. Aston 2 1 Mathematics Department, Howard University, Washington,

More information

Linear Dynamical Systems

Linear Dynamical Systems Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations

More information

Characterization of Convex and Concave Resource Allocation Problems in Interference Coupled Wireless Systems

Characterization of Convex and Concave Resource Allocation Problems in Interference Coupled Wireless Systems 2382 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL 59, NO 5, MAY 2011 Characterization of Convex and Concave Resource Allocation Problems in Interference Coupled Wireless Systems Holger Boche, Fellow, IEEE,

More information

USING multiple antennas has been shown to increase the

USING multiple antennas has been shown to increase the IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 55, NO. 1, JANUARY 2007 11 A Comparison of Time-Sharing, DPC, and Beamforming for MIMO Broadcast Channels With Many Users Masoud Sharif, Member, IEEE, and Babak

More information

LIST OF MATHEMATICAL PAPERS

LIST OF MATHEMATICAL PAPERS LIST OF MATHEMATICAL PAPERS 1961 1999 [1] K. Sato (1961) Integration of the generalized Kolmogorov-Feller backward equations. J. Fac. Sci. Univ. Tokyo, Sect. I, Vol. 9, 13 27. [2] K. Sato, H. Tanaka (1962)

More information

(implicitly assuming time-homogeneity from here on)

(implicitly assuming time-homogeneity from here on) Continuous-Time Markov Chains Models Tuesday, November 15, 2011 2:02 PM The fundamental object describing the dynamics of a CTMC (continuous-time Markov chain) is the probability transition (matrix) function:

More information

Statistical NLP: Hidden Markov Models. Updated 12/15

Statistical NLP: Hidden Markov Models. Updated 12/15 Statistical NLP: Hidden Markov Models Updated 12/15 Markov Models Markov models are statistical tools that are useful for NLP because they can be used for part-of-speech-tagging applications Their first

More information

MIXTURE OF EXPERTS ARCHITECTURES FOR NEURAL NETWORKS AS A SPECIAL CASE OF CONDITIONAL EXPECTATION FORMULA

MIXTURE OF EXPERTS ARCHITECTURES FOR NEURAL NETWORKS AS A SPECIAL CASE OF CONDITIONAL EXPECTATION FORMULA MIXTURE OF EXPERTS ARCHITECTURES FOR NEURAL NETWORKS AS A SPECIAL CASE OF CONDITIONAL EXPECTATION FORMULA Jiří Grim Department of Pattern Recognition Institute of Information Theory and Automation Academy

More information

Stochastic process. X, a series of random variables indexed by t

Stochastic process. X, a series of random variables indexed by t Stochastic process X, a series of random variables indexed by t X={X(t), t 0} is a continuous time stochastic process X={X(t), t=0,1, } is a discrete time stochastic process X(t) is the state at time t,

More information

IEOR 6711: Stochastic Models I Professor Whitt, Thursday, November 29, Weirdness in CTMC s

IEOR 6711: Stochastic Models I Professor Whitt, Thursday, November 29, Weirdness in CTMC s IEOR 6711: Stochastic Models I Professor Whitt, Thursday, November 29, 2012 Weirdness in CTMC s Where s your will to be weird? Jim Morrison, The Doors We are all a little weird. And life is a little weird.

More information

Least Mean Square Algorithms With Markov Regime-Switching Limit

Least Mean Square Algorithms With Markov Regime-Switching Limit IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 50, NO. 5, MAY 2005 577 Least Mean Square Algorithms With Markov Regime-Switching Limit G. George Yin, Fellow, IEEE, and Vikram Krishnamurthy, Fellow, IEEE

More information

A Recurrent Neural Network for Solving Sylvester Equation With Time-Varying Coefficients

A Recurrent Neural Network for Solving Sylvester Equation With Time-Varying Coefficients IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL 13, NO 5, SEPTEMBER 2002 1053 A Recurrent Neural Network for Solving Sylvester Equation With Time-Varying Coefficients Yunong Zhang, Danchi Jiang, Jun Wang, Senior

More information

WE start with a general discussion. Suppose we have

WE start with a general discussion. Suppose we have 646 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 43, NO. 2, MARCH 1997 Minimax Redundancy for the Class of Memoryless Sources Qun Xie and Andrew R. Barron, Member, IEEE Abstract Let X n = (X 1 ; 111;Xn)be

More information

PITMAN S 2M X THEOREM FOR SKIP-FREE RANDOM WALKS WITH MARKOVIAN INCREMENTS

PITMAN S 2M X THEOREM FOR SKIP-FREE RANDOM WALKS WITH MARKOVIAN INCREMENTS Elect. Comm. in Probab. 6 (2001) 73 77 ELECTRONIC COMMUNICATIONS in PROBABILITY PITMAN S 2M X THEOREM FOR SKIP-FREE RANDOM WALKS WITH MARKOVIAN INCREMENTS B.M. HAMBLY Mathematical Institute, University

More information

Received: 20 December 2011; in revised form: 4 February 2012 / Accepted: 7 February 2012 / Published: 2 March 2012

Received: 20 December 2011; in revised form: 4 February 2012 / Accepted: 7 February 2012 / Published: 2 March 2012 Entropy 2012, 14, 480-490; doi:10.3390/e14030480 Article OPEN ACCESS entropy ISSN 1099-4300 www.mdpi.com/journal/entropy Interval Entropy and Informative Distance Fakhroddin Misagh 1, * and Gholamhossein

More information

EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER

EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER Zhen Zhen 1, Jun Young Lee 2, and Abdus Saboor 3 1 Mingde College, Guizhou University, China zhenz2000@21cn.com 2 Department

More information

Hybrid HMM/MLP models for time series prediction

Hybrid HMM/MLP models for time series prediction Bruges (Belgium), 2-23 April 999, D-Facto public., ISBN 2-649-9-X, pp. 455-462 Hybrid HMM/MLP models for time series prediction Joseph Rynkiewicz SAMOS, Université Paris I - Panthéon Sorbonne Paris, France

More information