Exact Synchronization for Finite-State Sources

Size: px
Start display at page:

Download "Exact Synchronization for Finite-State Sources"

Transcription

1 Santa Fe Institute Woring Paper XXX arxiv.org: [nlin.cd] Exact Synchronization for Finite-State Sources Nicholas F. Travers 1, 2, 1, 2, 3, 4, and James P. Crutchfield 1 Complexity Sciences Center 2 Mathematics Department 3 Physics Department University of California at Davis, One Shields Avenue, Davis, CA Santa Fe Institute 1399 Hyde Par Road, Santa Fe, NM (Dated: October 25, 2010) We analyze how an observer synchronizes to the internal state of a finite-state information source, using the ɛ-machine causal representation. Here, we treat the case of exact synchronization, when only a finite number of observations are required. The more difficult case of strictly asymptotic synchronization is treated in a sequel. In both cases, we find that an observer synchronizes to the source exponentially fast and, as a result, the accuracy in an observer s predictions of future source output approaches its optimal level exponentially fast as well. We show how to analytically calculate the synchronization exponent. Additionally, we provide a characterization theorem for exact ɛ-machines and, from it, build a polynomial-time algorithm to test for exactness. PACS numbers: r c Tp Ey I. INTRODUCTION One of the ey questions in time series analysis is prediction: given a sequence of observed time series data X 0, X 1,..., X t 1, how well can one predict the next data point X t? If the X t s are independent, identically distributed and one nows their distributions, then, of course, no amount of past data can aid an observer in this tas. However, generally there are correlations between them and, in this case, the more data an observer has, the better its predictions become. Our wor is ultimately motivated by an interest in the question of prediction, but we approach it from a logically prior problem that of synchronization. Specifically, we study data sources from a class of finite-state hidden Marov models nown as ɛ-machines [1]. We tae the position of an observer that has a correct ɛ-machine model of the source and tries to predict the next symbol it generates using past observations of the source output. That is, we assume the observer sees past output symbols generated by the source, but does not have direct access to its hidden internal states (either current or past). Rather, the observer must infer the internal state (synchronize) only through observations of the output. Since future output is a function of the current state, better nowledge of the current state enables an observer to mae better predictions. The more readily an observer synchronizes, the better it can predict. Synchronization and state estimation for finite-state machines is a central interest in several disciplines. In information theory state estimation for hidden Marov Electronic address: ntravers@math.ucdavis.edu Electronic address: chaos@cse.ucdavis.edu models (HMMs) has been studied extensively; see, for example, Refs. [2] and [3]. However, the output sequences of generic HMMs irretrievably throw away internal state information. They lac, what is called, unifilarity. Without this property, an observer cannot (on average) synchronize completely and, if synchronization is gained, it can be repeatedly lost. And so, much of the analysis there is necessarily different, as we will consider only unifilar HMMs. There has been substantial wor, as well, in symbolic dynamics on synchronizing presentations of sofic shifts [4] and in automata theory on synchronizing sequences and homing sequences for Mealy machines [5]. Much of this wor, though, is of an entirely topological nature, since the machines are often defined with nonprobabilistic transitions. Although, in some cases probabilistic machines have been studied as well [6]. We will consider here, specifically, probabilistic, unifilar, finitestate machines. Our development proceeds as follows. Section II provides the necessary bacground. Definitions of ɛ-machines and synchronization are given, as well as those for the predictive quantities h µ and h µ (L) the entropy rate and its finite-time window approximates [7]. We identify two qualitatively distinct types of synchronization: exact (synchronization driven by finite sequences) and asymptotic (where there are no such finite sequences). The exact case is the subject here; the nonexact case is treated in a sequel [8]. Section III gives results on the synchronization rates for exact ɛ-machines and consequences for the convergence rates of the average state uncertainty H(L) and entropy-rate approximation h µ (L). Section IV characterizes exact ɛ-machines and introduces an algorithmic test for exactness. Finally, Sec. V summarizes our results and indicates several extensions.

2 2 II. BACKGROUND This section introduces the primary objects of study: finite-state ɛ-machines and the processes that they describe. Additionally, we give formal definitions for exact and asymptotic synchronization, the source entropy rate h µ, and its length-l approximation h µ (L). The latter two quantities will be used to monitor an observer s predictive capability. A. Stationary Information Sources Let A be a finite alphabet, and let X 0, X 1,... be the random variables (RVs) for a sequence of observed symbols x t A generated by an information source. We denote the RVs for the sequence of future symbols beginning at time t 0 as X X 0 X 1 X 2..., the bloc of L symbols beginning at time t 0 as X L X 0 X 1...X L 1, and the bloc of L symbols beginning at a given time t as X L t X t X t+1...x t+l 1. A stationary source is one for which Pr( X L t ) Pr( X L 0 ) for all t and all L > 0. We monitor our observer s performance at predicting a stationary source using information-theoretic measures [9]. The first step in this is to review several basic measures for stationary processes. Definition 1. The bloc entropy H(L) for a stationary source is H(L) H[ X L ] { x L } Pr( x L ) log 2 Pr( x L ). (1) The bloc entropy gives the average uncertainty in observing blocs X L. Definition 2. The entropy rate h µ is the asymptotic average entropy per symbol: H(L) h µ lim L L lim L H[X L X L ]. (2) It gives the intrinsic randomness of a source the irreducible uncertainty in future output once all of the correlations and structure of the source are nown. An observer cannot predict the source output any better than an error rate that is a function of h µ [9]. Definition 3. The entropy rate s length-l approximation is h µ (L) H(L) H(L 1) H[X L 1 X L 1 ]. (3) That is, h µ (L) is the observer s average uncertainty in the next symbol to be generated after observing the first L 1 symbols. For any stationary process, h µ (L) monotonically decreases to the limit h µ [9]. However, the form of convergence depends on the process. The lower the value of h µ a source has, the better an observer s predictions of the source output will be asymptotically. The faster h µ (L) converges to h µ, the faster an observer s predictions will reach this optimal asymptotic level. Since we are often interested in maing predictions after observing only finite data, the source s true entropy rate h µ, as well as the rate of convergence of h µ (L) to h µ, are both important properties. B. Hidden Marov Models In what follows we will restrict our attention to an important class of stationary information sources nown as hidden Marov models. For simplicity, we assume the number of states is finite. Definition 4. A finite-state edge-labeled hidden Marov machine (HMM) consists of 1. a finite set of states S {σ 1,..., σ N }, 2. a finite alphabet of symbols A, and 3. a set of N by N symbol-labeled transition matrices T (x), x A, where T (x) ij is the probability of transitioning from state σ i to state σ j on symbol x. The corresponding overall state-to-state transition matrix is denoted T x A T (x). A hidden Marov machine can be depicted as a directed graph with labeled edges. The nodes are the states {σ 1,..., σ N } and for all x, i, j with T (x) ij > 0 there is an edge from state σ i to state σ j labeled p x for the symbol x and transition probability p T (x) ij. We require that the transition matrices T (x) be such that this graph is strongly connected. A hidden Marov machine M generates a stationary process P as follows. Initially, M starts in some state σ i chosen according to the stationary distribution π over machine states the distribution satisfying πt π. It then pics an outgoing edge according to their relative transition probabilities T (x) i j, generates the symbol x labeling this edge, and follows the edge to a new state σ j. The next output symbol and state are consequently chosen in a similar fashion, and the process is repeated indefinitely. We denote S 0, S 1, S 2,... as the RVs for the sequence of machine states visited and X 0, X 1, X 2,... as the RVs for the associated sequence of output symbols generated. The sequence (S L ) L 0 is a Marov chain with transition ernel T. However, the stochastic process we consider is not the sequence of states, but rather the associated sequence of outputs (X L ) L 0, which is not (generally) Marovian. We assume the observer directly observes this sequence of outputs, but does not have direct access to the machine s hidden internal states.

3 3 C. Examples In the following it will be helpful to refer to several example hidden Marov machines that illustrate ey properties and definitions. We introduce four examples, all with a binary alphabet A {0, 1}. 1. Even Process Figure 1 gives a HMM for the Even Process. Its transitions matrices are: T (0) p 0, 0 0 T (1) 0 1 p. (4) 1 0 p 0 1 p 1 σ 1 σ FIG. 1: A hidden Marov machine (the ɛ-machine) for the Even Process. The transitions denote the probability p of generating symbol x as p x. The support for the Even Process consists of all binary sequences in which blocs of uninterrupted 1s are even in length, bounded by 0s. After each even length is reached, there is a probability p of breaing the bloc of 1s by inserting one or more 0s. The hidden Marov machine has two internal states, S {σ 1, σ 2 }, and a single parameter p [0, 1] that controls the transition probabilities. 2. Alternating Biased Coins Figure 2 shows a HMM for the Alternating Biased Coins (ABC) Process. The transitions matrices are: T (0) 0 1 p, 1 q 0 T (1) 0 p. (5) q 0 the biases of the two coins that are flipped alternatively in sequence to generate the process. 3. SNS Process Figure 3 depicts a two-state HMM for the SNS Process which generates long sequences of 1s broen by isolated 0s. Its matrices are: T (0) 0 0, 1 q 0 T (1) p 1 p. (6) 0 q 1 p 1 p 1 σ 1 σ 2 q 0 1 q 1 FIG. 3: An HMM for the SNS Process. Note that the two transitions leaving state σ 1 both emit x Noisy Period-2 Process Finally, Fig. 4 depicts a nonminimal HMM for the Noisy Period-2 (NP2) Process. The transition matrices are: T (0) T (1) p p p p 0 0 0, 1 1 σ 1 σ 2. (7) p 1, 1 p 0 σ 1 σ 2 q 1, 1 q 0 p 1, 1 p 0 p 1, 1 p 0 FIG. 2: A hidden Marov machine (the ɛ-machine) for the Alternating Biased Coins Process. It too has two internal states. There are also two transition probability parameters: p, q [0, 1]. These control σ σ 3 FIG. 4: An HMM for the Noisy Period-2 Process.

4 4 It is clear by inspection that the same process can be captured by a hidden Marov machine with fewer states. Specifically, the distribution over future sequences from states σ 1 and σ 3 are the same, so those two states are redundant and can be merged. The same is also true for states σ 2 and σ 4. D. ɛ-machines We now introduce a class of hidden Marov machines that has a number of desirable properties for analyzing synchronization. Definition 5. A finite-state ɛ-machine is a finite-state edge-labeled hidden Marov machine with the following properties: 1. Unifilarity: For each state σ S and each symbol x A there is at most one outgoing edge from state σ labeled with symbol x. 2. Probabilistically distinct states: For each pair of distinct states σ, σ j S there exists some finite word w x 0 x 1... x L 1 such that Pr( X L w S 0 σ ) Pr( X L w S 0 σ j ). The hidden Marov machines given above for the Even Process and ABC process are both ɛ-machines. The SNS machine of example 3 is not an ɛ-machine, however, since state σ 1 is not unifilar. The NP2 machine of example 4 is also not an ɛ-machine, since it does not have probabilistically distinct states, as noted before. ɛ-machines were originally defined in Ref. [1] as hidden Marov machines whose states, nown as causal states, were the equivalence classes of infinite pasts x with the same probability distribution over futures x. This history ɛ-machine definition is, in fact, equivalent to the generating ɛ-machine definition presented above in the finite-state case. Although, this is not immediately apparent. Formally, it follows from the synchronization results established here and in Ref. [8]. It can also be shown that an ɛ-machine M for a given process P is unique up to isomorphism [1]. That is, there cannot be two different finite-state edge-labeled hidden Marov models with unifilar transitions and probabilistically distinct states that both generate the same process P. Furthermore, ɛ-machines are minimal unifilar generators in the sense that any other unifilar machine M generating the same process P as an ɛ-machine M will have more states than M. Note that uniqueness does not hold if we remove either condition 1 or 2 in the ɛ-machine definition. E. Synchronization Assume now that an observer has a correct model M (ɛ-machine) for a process P, but is not able to directly observe M s hidden internal state. Rather, the observer must infer the internal state by observing the output data that M generates. For a word w of length L generated by M let φ(w) Pr(S w) be the observer s belief distribution as to the current state of the machine after observing the word w. That is, φ(w) Pr(S L σ X L w) Pr(S L σ X L w, S 0 π). (8) Let L(M) be the set of all finite words that M generates, L L (M) be the set of all words of length L it generates, and L (M) be the set of all infinite sequences x x 0 x 1... which it generates. Definition 6. A word w L(M) is a synchronizing word (simply, sync word) for M if H[φ(w)] 0, where H[µ] is the Shannon entropy of a discrete distribution µ. That is, a synchronizing word w is one for which the observer nows exactly which state M is in after seeing w. Let us denote: Sync M : The set of M s synchronizing sequences as SYN(M) { x L (M): there exists L N such that x L is a sync word for M}. Wea Sync M : The set of M s wealy synchronizing sequences as WSYN(M) { x L (M): H[φ( x L )] 0 as L }. Definition 7. An ɛ-machine M is exactly synchronizable (simply, exact) if Pr(SYN(M)) 1. That is, the observer synchronizes exactly to almost every sequence generated by M in finite time. Definition 8. An ɛ-machine M is asymptotically synchronizable if Pr(WSYN(M)) 1. That is, the observer s uncertainty in M s state vanishes asymptotically for almost every sequence generated. The Even Process ɛ-machine, Fig. 1, is an exact machine. Any word containing a 0 is a sync word for this machine, and almost every x generated by it contains at least one 0. The ABC Process ɛ-machine, Fig. 2, is not exactly synchronizable, but it is asymptotically synchronizable. Remar. If w L(M) is a sync word, then by unifilarity so is wu, for all u with wu L(M). Thus, once an observer synchronizes exactly, it remains synchronized exactly for all future times. It follows that any exactly synchronizable machine is also asymptotically synchronizable. Remar. If w L(M) is a sync word then so is uw, for all u with uw L(M). Since any finite word w L(M) will be contained in almost every infinite sequence x the machine generates, it follows that a machine is exactly synchronizable if (and only if) it has some sync word w of finite length.

5 5 Remar. It turns out all finite-state ɛ-machines are asymptotically synchronizable; see Ref. [8]. Hence, there are two disjoint classes to consider: exactly synchronizable machines and asymptotically synchronizable machines that are nonexact. The exact case is the subject of the remainder. Finally, one last important quantity for synchronization is the observer s average uncertainty in the machine state after seeing a length L bloc of data [7]. Definition 9. The observer s average state uncertainty at time L is: H(L) H[S L X L ] Pr( x L ) H[S L X L x L ]. (9) { x L } That is, H(L) is the expected value of an observer s uncertainty in the machine state after observing L symbols. Now, for an ɛ-machine, an observer s prediction of the next output symbol is a direct function of the probability distribution over machine states induced by the previously observed symbols. Specifically, A. Exact Machine Synchronization Theorem Our first theorem states that an observer will synchronize (exactly) to the internal state of any exact ɛ-machine exponentially fast. Theorem 1. For any exact ɛ-machine M, there are constants K > 0 and 0 < α < 1 such that for all L N. Pr(NSYN L ) Kα L, (11) Proof. Let M be an exact machine with sync word w L(M, σ ). Since the graph of M is strongly connected we now that for each state σ j, there is a word v j such that δ(σ j, v j ) σ. Let w j v j w, n max j w j, and p min j Pr(w j σ j ). Then, for all L 0, we have: Pr(w X n+l w X L ) Pr(w X n L w X L ) min j Pr(w X n L S L σ j ) p. (12) Pr(X L x X L x L ) Pr(x σ ) Pr(S L σ X L x L ). (10) Hence, Pr(w X n+l w X L ) 1 p, (13) Hence, the better an observer nows the machine state at the current time, the better it can predict the next symbol generated. And, on average, the closer H(L) is to 0, the closer h µ (L) is to h µ. Therefore, the rate of convergence of h µ (L) to h µ for an ɛ-machine is closely related to the average rate of synchronization. III. EXACT SYNCHRONIZATION RESULTS This section provides our main results on synchronization rates for exact machines and draws out the consequences for the convergence rates of H(L) and h µ (L). The following notation will be used throughout: 1. SYN L {w L L (M) : w is a sync word for M}. 2. NSYN L {w L L (M) : w is not a sync word for M}. 3. SYN L,σ {w L L (M) : w synchronizes the observer to state σ }. 4. L(M, σ ) {w : M generates w starting in state σ }. 5. For a word w L(M, σ ), δ(σ, w) is defined to be w the (unique) state in S such that σ δ(σ, w). 6. For a set of states S S, we define: δ(s, w) {σ j S : σ w σj from some σ S}. for all L 0. And, therefore, for all m N: Pr(NSYN mn ) P r(w X mn ) Pr(w X n ) Pr(W X 2n w X n ) Pr(w X mn w X (m 1)n ) (1 p) (1 p) (1 p) (1 p) m β m, (14) where β 1 p. Hence, for L mn (m N): Pr(NSYN L ) Pr(NSYN mn ) β m α mn α L, (15) where α β 1/n. Since Pr(NSYN L ) is monotonically decreasing, it follows that: Pr(NSYN L ) 1 α n αl Kα L, (16) for all L N (where K 1/α n ). Remar. In the above proof we have implicitly assumed β 0. If β 0, then the conclusion follows trivially.

6 6 B. Synchronization Rate Theorem 1 states that an observer synchronizes (exactly) to any exact ɛ-machine exponentially fast. However, the synchronization rate constant: α lim L Pr(NSYN L) 1/L (17) depends on the machine, and it may often be of practical interest to now what the constant α is. We now we provide a method for computing α analytically. It is based on the construction of an auxiliary machine M. Definition 10. Let M be an ɛ-machine with states S {σ 1,..., σ N }, alphabet A, and transition matrices T (x), x A. The possibility machine M is defined as follows: 1. The alphabet of M is A. 2. The states of M are pairs of the form (σ, S) where σ S and S is a subset of S that contains σ. 3. The transition probabilities are: Pr((σ, S) x (σ, S )) Pr(x σ)i(x, (σ, S), (σ, S )), where I(x, (σ, S), (σ, S )) is the indicator function: I(x,(σ, S), (σ, S )) { 1 if δ(σ, x) σ and δ(s, x) S 0 otherwise. A state of M is said to be initial if it is of the form (σ, S) for some σ S. For simplicity we restrict the M machine to consist of only those states that are accessible from initial states. The other states are irrelevant for the analysis below. The idea is that M states represent states of the joint (machine,observer) system. Causal state σ is the true machine state at the current time, and S is the set of states that the observer believes are currently possible for the machine to be in, after observing all previous symbols. Initially, all states are possible (to the observer), so initial states are those in which the set of possible states is the complete set S. If the current true machine state is σ and then the symbol x is generated, then the new true machine state must be δ(σ, x). Similarly, if the observer believes any of the states in S are currently possible and then the symbol x is generated, then the new (to the observer) set of possible states is δ(s, x). This accounts for transitions in M topologically. The probability of generating a given symbol x from (σ, S) is, of course, governed only by the true state σ of the machine: Pr(x (σ, S)) Pr(x σ). An example of this construction for a 3-state exact ɛ-machine is given in Appendix A. Note that the graph of the M machine there has a single recurrent strongly connected component, which is isomorphic to the original machine M. This is not an accident. It will always be the case, as long as the original machine M is exact. Remar. If M has more than 1 state, the graph of M itself is never strongly connected. So, M is not an ɛ-machine or even an HMM in the sense of Def. 4. However, we still refer to M as a machine. M s states are denoted S {q 1,..., qñ}, its symbol-labeled transition matrices T (x), and its overall state-to-state transition matrix T T x A (x). We assume the states are ordered in such a way that the initial states (σ 1, S),..., (σ N, S) are, respectively, q 1,..., q N. Similarly, the recurrent states (σ 1, {σ 1 }), (σ 2, {σ 2 }),..., (σ N, {σ N }) are, respectively, q n+1, q n+2,..., qñ, where n Ñ N. The ordering of the other states is irrelevant. In this case, the matrix T has the following bloc upper triangular form: B B T, (18) O T where B is a n n matrix with nonnegative entries, B is a n N matrix with non-negative entries, O is a N n matrix of all zeros, and T is the N N state-to-state transition matrix of the original ɛ-machine M. Let π (π 1,..., π N, 0,..., 0) denote the length-ñ row vector whose distribution over the initial states is the same as the stationary distribution π for the ɛ-machine M. Then, the initial probability distribution φ 0 over states of the joint (machine, observer) system is simply: φ 0 π, (19) and the distribution over states of the joint system after the first L symbols is: φ L π T L. (20) If the joint system is in a recurrent state of the form (σ, {σ }), then to the observer the only possible state of the ɛ-machine is the true state. The observer is synchronized. For all other states of M, the observer is not yet synchronized. Hence, the probability the observer is not synchronized after L symbols is simply the combined probability of all nonreccurent states q i in the distribution φ L. Mathematically, we have: n Pr(NSYN L ) ( φ L ) i i1 n ( π T L ) i i1 n (π B B L ) i i1 π B B L 1, (21)

7 7 where π B (π 1,..., π N, 0,..., 0) is the length-n row vector corresponding to the distribution over initial states π. (The third equality follows from the bloc uppertriangular form of T.) Appendix B shows that: lim ( π B B L ) 1/L 1 r, (22) L where r r(b) is the (left) spectral radius of B: r(b) max{ λ : λ is a (left) eigenvalue of B}. (23) Thus, we have, Theorem 2. For any exact ɛ-machine M, α r. C. Consequences We now apply Thm. 1 to show that an observer s average uncertainty H(L) in the machine state and average uncertainty h µ (L) in predictions of future symbols both decay exponentially fast to their respective limits: 0 and h µ. The decay constant α in both cases is essentially bounded by the sync rate constant α from Thm. 2. Proposition 1. For any exact ɛ-machine M, there are constants K > 0 and 0 < α < 1 such that H(L) Kα L, for all L N. Proof. Let M be any exact machine. Then by Thm. 1, there are constants C > 0 and 0 < α < 1 such that Pr(NSYN L ) Cα L, for all L N. Thus, we have: H(L) Pr(w)H[φ(w)] w L L (M) Pr(w)H[φ(w)] + Pr(w)H[φ(w)] w SYN L w NSYN L 0 + Pr(w) log(n) w NSYN L log(n) Cα L Kα L, (24) where N is the number of machine states and K C log(n). Proposition 2. For any exact ɛ-machine M: h µ π H[X next σ ], (25) where X next is the RV for the symbol following σ. And, there are constants K > 0 and 0 < α < 1 such that: for all L N. h µ (L) h µ Kα L, (26) Remar. The h µ formula Eq. (25) has been nown for some time, although in slightly different contexts. Shannon, for example, derived this formula in his original publication [10] for a type of hidden Marov machine that is similar (apparently unifilar) to an ɛ-machine. Proof. Let M be any exact machine. Since we now h µ (L) h µ it suffices to show there are constants K > 0 and 0 < α < 1 such that: h µ(l) π H[X next σ ] KαL, (27) for all L N. This will establish both the value of h µ and the necessary convergence. Now, by Thm. 1, there are constants C > 0 and 0 < α < 1 such that Pr(NSYN L ) Cα L, for all L N. Also, note that for all L and we have: π Pr(w) φ(w) Thus, and w L L (M) w SYN L,σ Pr(w) φ(w) Pr(SYN L,σ ). (28) (π Pr(SYN L,σ )) H[X next σ ] 0, (29) (π Pr(SYN L,σ )) H[X next σ ] Also, clearly, and, (π Pr(SYN L,σ )) log A ( log A π ) Pr(SYN L,σ ) log A (1 Pr(SYN L )) log A Pr(NSYN L ) log A Cα L. (30) w NSYN L Pr(w) H[X next w] 0, (31) w NSYN L Pr(w) H[X next w] log A Pr(NSYN L ) log A Cα L. (32)

8 8 Therefore, we have for all L N: h µ(l + 1) π H[X next σ ] Pr(w)H[X next w] π H[X next σ ] w L L (M) Pr(w)H[X next w] w NSYN L + Pr(w)H[X next w] π H[X next σ ] w SY N L Pr(w)H[X next w] w NSYN L + Pr(w)H[X next w] π H[X next σ ] w SYN L,σ Pr(w)H[X next w] w NSYN L + Pr(SYN L,σ )H[X next σ ] π H[X next σ ] Pr(w)H[X next w] w NSYN L (π Pr(SYN L,σ ))H[X next σ ] C log A α L. (33) The last inequality follows from Eqs. (29)-(32), since x y z for all non-negative real numbers x, y, and z with x z and y z. Finally, since, h µ(l + 1) π H[X next σ ] C log A αl, (34) for all L N, there must be some K C log A /α such that: for all L N. h µ(l) π H[X next σ ] KαL, (35) Remar. For any α > α there exists some K > 0 for which Eq. 11 holds. Hence, by the constructive proofs above, we see that the constant α in Props. 1 and 2 can be chosen arbitrarily close to α. IV. CHARACTERIZATION OF EXACT ɛ-machines In this section we provide a set of necessary and sufficient conditions for exactness and an algorithmic test for exactness based upon these conditions. A. Exact Machine Characterization Theorem Definition 11. States σ and σ j are said to be topologically distinct if L(M, σ ) L(M, σ j ). Definition 12. States σ and σ j are said to be path convergent if there exists w L(M, σ ) L(M, σ j ) such that δ(σ, w) δ(σ j, w). If states σ and σ j are topologically distinct (or path convergent) we will also say the pair (σ, σ j ) is topologically distinct (or is path convergent). Theorem 3. An ɛ-machine M is exact if and only if every pair of distinct states (σ, σ j ) satisfies at least one of the following two conditions: (i) The pair (σ, σ j ) is topologically distinct. (ii) The pair (σ, σ j ) is path convergent. Proof. It was noted above that an ɛ-machine M is exact if and only if it has some sync word w of finite length. Therefore, it is enough to show that every pair of distinct states (σ, σ j ) satisfies either (i) or (ii) if and only if M has some sync word w of finite length. We establish the if first: If M has a sync word w, then every pair of distinct states (σ, σ j ) satisfies either (i) or (ii). Let w be a sync word for M. Then w L(M, σ ) for some. Tae words v j, j 1, 2,...N, such that δ(σ j, v j ) σ. Then, the word v j w w j L(M, σ j ) is also a sync word for M for each j. Therefore, for each i j either w j L(M, σ i ) or δ(σ i, w j ) δ(σ j, w j ). This establishes that the pair (σ i, σ j ) is either topologically distinct or path convergent. Since this holds for all j 1, 2,..., N and for all i j, we now every pair of distinct states is either topologically distinct or path convergent. Now, for the only if case: If every pair of distinct states (σ, σ j ) satisfies either (i) or (ii), then M has a sync word w. If each pair of distinct states (σ, σ j ) satisfies either (i) or (ii), then for all and j ( j) there is some word w σ,σ j such that one of the following three conditions is satisfied. 1. w σ,σ j L(M, σ ), but w σ,σ j L(M, σ j ). 2. w σ,σ j L(M, σ j ), but w σ,σ j L(M, σ ). 3. w σ,σ j L(M, σ ) L(M, σ j ) and δ(σ, w σ,σ j ) δ(σ j, w σ,σ j ).

9 9 We construct a sync word w w 1 w 2...w m for M, where each w i w σi,σ ji for some i and j i, as follows. Let S 0 {σ 0 1,..., σ 0 N 0 } S {σ 1,..., σ N }. Tae w 1 w σ 0 1,σ 0 2. Let S 1 {σ 1 1,..., σ 1 N 1 } δ(s 0, w 1 ). Since w 1 w σ 0 1,σ 0 2 satisfies either (1), (2), or (3), we now N 1 < N 0. Tae w 2 w σ 1 1,σ 1 2. Let S 2 {σ 2 1,..., σ 2 N 2 } δ(s 1, w 2 ). Since w 2 w σ 1 1,σ 1 2 satisfies either (1), (2), or (3) we now N 2 < N 1. Tae w 3 w σ 2 1,σ 2 2. Repeat these steps until S m 1 for some m. Note that this will happen after a finite number of steps since N N 0 is finite and N 0 > N 1 > N 2 >. By this construction w w 1 w 2...w m L(M) is a sync word for M. After observing w, an observer nows the machine must be in state σ m 1. B. A Test for Exactness We can now provide an algorithmic test for exactness using the characterization theorem of exact machines. We begin with subalgorithms to test for topological distinctness and path convergence of state pairs. Both are essentially the same algorithm and only a slight modification of the deterministic finite-automata (DFA) tablefilling algorithm that tests for pairs of equivalent states [11]. Algorithm 1. Test States for Topological Distinctness. 1. Initialization: Create a table containing boxes for all pairs of distinct states (σ, σ j ). Initially, all boxes are blan. Then, Loop over distinct pairs (σ, σ j ) Loop over x A If {x L(M, σ ) but x L(M, σ j )} or {x L(M, σ j ) but x L(M, σ )}, then mar box for pair (σ, σ j ). end end 2. Induction: If δ(σ, x) σ and δ(σ j, x) σ j and the box for the pair (σ, σ j ) is mared, then mar the box for pair (σ, σ j ). Repeat until no more inductions are possible. Algorithm 2. Test States for Path Convergence. This algorithm is identical to Algorithm 1 except that the if-statement in the initialization step is replaced with the following: If x L(M, σ ) L(M, σ j ) and δ(σ, x) δ(σ j, x), then mar box for pair (σ, σ j ). With Algorithm 1 all pairs of topologically distinct states end up with mared boxes. With Algorithm 2 all pairs of path convergent states end up with mared boxes. Both of these facts can be proved using induction on the length of the minimal distinguishing (path converging) word w for a given pair of states. The proofs are virtually identical to the proof of the standard DFA table filling algorithm, so the details have been omitted. Note also that both of these are polynomial-time algorithms. Step (1) has run time O( A N 2 ). The inductions in Step (2), if done in a reasonably efficient fashion, can also be completed in run time O( A N 2 ). (See, e.g., the analysis of DFA table filling algorithm in Ref. [11].) Therefore, the total run time of these algorithm is O( A N 2 ). Algorithm 3. Test for Exactness. 1. Use Algorithm 1 to find all pairs of topologically distinct states. 2. Use Algorithm 2 to find all pairs of path convergent states. 3. Loop over all pairs of distinct states (σ, σ j ) to chec if they are either (i) topologically distinct or (ii) path convergent. By Thm. 3, if all distinct pairs of states satisfy (i) or (ii) or both, the machine is exact and otherwise it is not. This, too, is a polynomial-time algorithm. Steps (1) and (2) have run time O( A N 2 ). Step (3) has run time O(N 2 ). Hence, the total run time for this algorithm is O( A N 2 ). V. CONCLUSION We analyzed the process of exact synchronization to finite-state ɛ-machines. In particular, we showed that for exact machines an observer synchronizes exponentially fast and that, as a result, the average uncertainty h µ (L) in an observer s predictions converges exponentially fast to the entropy rate h µ. Additionally, we found an efficient (polynomial-time) algorithm to test ɛ-machines for exactness. In Ref. [8] we similarly analyze asymptotic synchronization to nonexact ɛ-machines. It turns out that qualitatively similar results hold. That is, H(L) and h µ (L) both converge to their respective limits exponentially fast. However, the proof methods in the nonexact case are substantially different. In the future we plan to extend these results to more generalized model classes, such as to ɛ-machines with a countable number of states and to nonunifilar hidden Marov machines.

10 10 Acnowledgments NT was partially supported on a VIGRE fellowship. The wor was partially supported by the Defense Advanced Research Projects Agency (DARPA) Physical Intelligence project. The views, opinions, and/or findings contained in this article are those of the authors and should not be interpreted as representing the official views or policies, either expressed or implied, of the DARPA or the Department of Defense. [1] J. P. Crutchfield and K. Young. Inferring statistical complexity. Phys. Rev. Let., 63: , J. P. Crutchfield, The Calculi of Emergence: Computation, Dynamics, and Induction, Physica D 75: 11 54, C. R. Shalizi and J. P. Crutchfield, Computational Mechanics: Pattern and Prediction, Structure and Simplicity, J. Stat. Phys. 104: , [2] G. D. Forney Jr. The Viterbi algorithm: A personal history. CoRR, abs/cs/ , [3] A. J. Viterbi. Error bounds for convolutional codes and an asymptotically optimum decoding algorithm. IEEE Trans. Info. Th., 13(2): , [4] N. Jonosa. Sofic shifts with synchronizing presentations. Theor. Comput. Sci., 158(1-2):81 115, [5] S. Sandberg. Homing and synchronizing sequences. In M. Broy et al, editor, Lect. Notes Comp. Sci., volume 3472, pages Springer, Berlin, [6] A. Paz. Introduction to Probabilistic Automata. Academic Press, New Yor, [7] J. P. Crutchfield and D. P. Feldman. Regularities unseen, randomness observed: Levels of entropy convergence. CHAOS, 13(1):25 54, [8] N. Travers and J. P. Crutchfield. Asymptotic synchronization for finite-state sources SFI Woring Paper XXX; arxiv.org:10xx.xxxx [XXXX]. [9] T. M. Cover and J. A. Thomas. Elements of Information Theory. Wiley-Interscience, New Yor, second edition, [10] C. E. Shannon. A mathematical theory of communication. Bell Sys. Tech. J., 27: , , [11] J. E. Hopcroft, R. Motwani, and J. D. Ullman. Automata Theory, Languages, and Computation. Addison-Wesley, Reading, [12] M. Reed and B. Simon. Functional Analysis. Academic Press, Appendix A We construct the possibility machine M for the threestate ɛ-machine shown in Fig. 5. The result is shown in Fig. 6. Appendix B We prove Eq. (22) in Sec. III B. Lemma 1. For any exact ɛ-machine M, lim L πb B L 1/L 1 r(b). (B1) p 21 a σ 1 p 12 a p 31 c p 13 c σ 2 σ 3 p 23 b p 11 b p 33 a FIG. 5: A three-state ɛ-machine M with alphabet A {a, b, c}. In what follows A denotes an arbitrary m m matrix and v and w denote row m-vectors. Unless otherwise specified, the entries of matrices and vectors are assumed to be complex. Definition 13. The (left) matrix p-norms (1 p ) are defined as A p max{ v A p : v p 1}. The following facts will be used in our proof. (B2) Fact 1. If A is a matrix with real non-negative entries and v (v 1,..., v m ) is a vector with real non-negative entries, then: v A 1 m (v e )A 1 1 m v e A 1, 1 where e (0,..., 1,..., 0) is the th standard basis vector. Fact 2. Let A be a matrix with real nonnegative entries, let v (v 1,..., v m ) be a vector with complex entries, and let w (w 1,..., w m ) ( v 1,..., v m ). Then: v A 1 w A 1. (B3)

11 11 p 21 a 2, ewns 123 1, ewns 123 3, ewns 123 p 33 a p 23 b p 12 a p 13 c 3, ewns 23 3, ewns 13 1, ewns 13 2, ewns 23 p 31 c p 33 a p 21 a p 12 a p 31 c p 13 c p 11 b 1, ewns 1 p 11 b p 31 c p 11 b p 13 c p 33 a p 31 c p 12 a p 21 a 2, ewns 2 3, ewns 3 p 23 b p 23 b p 33 a FIG. 6: The possibility machine M for the three-state ɛ-machine M of Fig. 5. The state names have been abbreviated for display purposes: e.g., (σ 1, {σ 1, σ 2, σ 3}) (1, 123). Fact 3. For any matrix A {a ij }, the matrix 1-norm is the largest absolute row sum: A 1 max i m a ij. (B4) j1 Fact 4. For any matrix A, L N, and 1 p : A L p A L p. (B5) Fact 5. For any matrix A and 1 p : lim L AL 1/L p r(a), (B6) where r(a) is the (left) spectral radius of A: r(a) max{ λ : λ is a (left) eigenvalue of A}. (B7) (This is, of course, the same as the right spectral radius, but we emphasize the left eigenvalues for the proof of Lemma 1 below.) Fact 1 can be proved by direct computation, and Fact 2 follows from the triangle inequality. Fact 3 is a standard result from linear algebra. Facts 4 and 5 are finitedimensional versions of more general results established in Ref. [12] for bounded linear operators on Banach spaces. Using these facts we now prove Lemma 1. Proof. By Fact 5 we now: lim sup π B B L 1/L 1 r(b). (B8) L Thus, it suffices to show that: lim inf L πb B L 1/L 1 r(b). (B9) Let us define the B-machine to be the restriction of the M machine to it s nonreccurent states. The stateto-state transition matrix for this machine is B. We call the states of this machine B-states and refer to paths in the associated graph as B-paths. Note that the rows of B {b ij } are substochastic: b ij 1, j (B10) for all i, with strict inequality for at least one value of i. By the construction of the B-machine we now that for each of its states q j there exists some initial state q i q i(j) such that q j is accessible from q i(j). Define l j to be the length of the shortest B-path from q i(j) to q j, and l max max j l j. Let c j > 0 be the probability, according to the initial distribution π B, of both starting in state q i(j) at time 0 and ending in state q j at time l j : c j (π i(j) e i(j) B lj ) j. Finally, let C 1 min j c j. Then, for any L > l max and any state q j we have: π B B L 1 π i(j) e i(j) B L 1 (B11) (π i(j) e i(j) B lj )B L lj 1 (B12) c j e j B L lj 1 C 1 e j B L lj 1 C 1 e j B L 1. (B13) (B14) (B15) Equation (B12) follows from Fact 1. The decomposition in Eq. (B13) is possible since L > l max l j. Eq. (B14) follows from Fact 1 and the definition of c j. Equation (B15) follows from the definition of C 1. Finally, Eq. (B16) follows from Fact 3, Fact 4, and Eq. (B11). Now, tae a normalized (left) eigenvector y (y 1,..., y n ) of B whose associated eigenvalue is maximal. That is, y 1 1, y B λ y, and λ r(b). Define z (z1,..., z n ) ( y 1,..., y n ). Then, for any L N: n z e B L 1 z B L 1 1 y B L 1 (B16) (B17) λ L y 1 (B18) λ L y 1 (B19) r(b) L, (B20)

12 12 where Eq. (B17) follows from Fact 1 and Eq. (B18) follows from Fact 2. Therefore, for each L we now there exists some j j(l) in {1,..., n} such that: z j(l) e j(l) B L 1 r(b)l n. (B21) Now, r(b) may be 0, but we can still choose the j(l) s such that z j(l) is never zero. And, in this case, we may divide through by z j(l) on both sides of Eq. (B22) to obtain, for each L: Therefore, for any L > l max we now: π B B L 1 C 1 e j(l) B L 1 C 1 (C 2 r(b) L) (B23) (B24) C 3 r(b) L, (B25) where C 3 C 1 C 2. Equation (B24) follows from Eq. (B16) and Eq. (B25) follows from Eq. (B23). Finally, since this holds for all L > l max, we have: e j(l) B L 1 where C 2 > 0 is defined by r(b)l n z j(l) C 2 r(b) L, 1 C 2 min. z j 0 n z j (B22) lim inf L πb B L 1/L 1 lim inf L ( C3 r(b) L) 1/L r(b). (B26)

Exact Synchronization for Finite-State Sources

Exact Synchronization for Finite-State Sources Santa Fe Institute Woring Paper 10-11-031 arxiv.org:1008.4182 [nlin.cd] Exact Synchronization for Finite-State Sources Nicholas F. Travers 1, 2, 1, 2, 3, 4, and James P. Crutchfield 1 Complexity Sciences

More information

Equivalence of History and Generator ɛ-machines

Equivalence of History and Generator ɛ-machines MCS Codes: 37A50 37B10 60J10 Equivalence of History and Generator ɛ-machines Nicholas F. Travers 1, 2, 1, 2, 3, 4, and James P. Crutchfield 1 Complexity Sciences Center 2 Mathematics Department 3 Physics

More information

Equivalence of History and Generator Epsilon-Machines

Equivalence of History and Generator Epsilon-Machines Equivalence of History and Generator Epsilon-Machines Nicholas F. Travers James P. Crutchfield SFI WORKING PAPER: 2011-11-051 SFI Working Papers contain accounts of scientific work of the author(s) and

More information

Bounds on Convergence of Entropy Rate Approximations in Hidden Markov Processes

Bounds on Convergence of Entropy Rate Approximations in Hidden Markov Processes Bounds on Convergence of Entropy Rate Approximations in Hidden Markov Processes By Nicholas F. Travers B.S. Haverford College 2005 DISSERTATION Submitted in partial satisfaction of the requirements for

More information

Statistical Complexity of Simple 1D Spin Systems

Statistical Complexity of Simple 1D Spin Systems Statistical Complexity of Simple 1D Spin Systems James P. Crutchfield Physics Department, University of California, Berkeley, CA 94720-7300 and Santa Fe Institute, 1399 Hyde Park Road, Santa Fe, NM 87501

More information

Computation in Sofic Quantum Dynamical Systems

Computation in Sofic Quantum Dynamical Systems Santa Fe Institute Working Paper 07-05-007 arxiv:0704.3075 Computation in Sofic Quantum Dynamical Systems Karoline Wiesner 1, and James P. Crutchfield 1, 1 Center for Computational Science & Engineering

More information

Spectral Simplicity of Apparent Complexity, Part I: The Nondiagonalizable Metadynamics of Prediction

Spectral Simplicity of Apparent Complexity, Part I: The Nondiagonalizable Metadynamics of Prediction Santa Fe Institute Woring Paper 207-05-08 arxivorg:70508042 [nlincd] Spectral Simplicity of Apparent Complexity, Part I: The Nondiagonalizable Metadynamics of Prediction Paul M Riechers and James P Crutchfield

More information

Enumerating Finitary Processes

Enumerating Finitary Processes Enumerating Finitary Processes B. D. Johnson J. P. Crutchfield C. J. Ellison C. S. McTague SFI WORKING PAPER: 200--027 SFI Working Papers contain accounts of scientific work of the author(s) and do not

More information

Information Accessibility and Cryptic Processes

Information Accessibility and Cryptic Processes Information Accessibility and Cryptic Processes John R. Mahoney Christopher J. Ellison James P. Crutchfield SFI WORKING PAPER: 2009-05-08 SFI Working Papers contain accounts of scientific work of the author(s)

More information

Computation in Finitary Quantum Processes

Computation in Finitary Quantum Processes Computation in Finitary Quantum Processes Karoline Wiesner James P. Crutchfield SFI WORKING PAPER: 006-09-03 SFI Working Papers contain accounts of scientific work of the author(s) and do not necessarily

More information

The Past and the Future in the Present

The Past and the Future in the Present The Past and the Future in the Present James P. Crutchfield Christopher J. Ellison SFI WORKING PAPER: 2010-12-034 SFI Working Papers contain accounts of scientific work of the author(s) and do not necessarily

More information

Permutation Excess Entropy and Mutual Information between the Past and Future

Permutation Excess Entropy and Mutual Information between the Past and Future Permutation Excess Entropy and Mutual Information between the Past and Future Taichi Haruna 1, Kohei Nakajima 2 1 Department of Earth & Planetary Sciences, Graduate School of Science, Kobe University,

More information

Towards a Theory of Information Flow in the Finitary Process Soup

Towards a Theory of Information Flow in the Finitary Process Soup Towards a Theory of in the Finitary Process Department of Computer Science and Complexity Sciences Center University of California at Davis June 1, 2010 Goals Analyze model of evolutionary self-organization

More information

An Optimal Lower Bound for Nonregular Languages

An Optimal Lower Bound for Nonregular Languages An Optimal Lower Bound for Nonregular Languages Alberto Bertoni Carlo Mereghetti Giovanni Pighizzini Dipartimento di Scienze dell Informazione Università degli Studi di Milano via Comelico, 39 2035 Milano

More information

Informational and Causal Architecture of Discrete-Time Renewal Processes

Informational and Causal Architecture of Discrete-Time Renewal Processes Entropy 2015, 17, 4891-4917; doi:10.3390/e17074891 OPEN ACCESS entropy ISSN 1099-4300 www.mdpi.com/journal/entropy Article Informational and Causal Architecture of Discrete-Time Renewal Processes Sarah

More information

Information Theoretic Properties of Quantum Automata

Information Theoretic Properties of Quantum Automata Information Theoretic Properties of Quantum Automata D. Gier Department of Physics and Astronomy, The University of Kansas, Lawrence, KS 66045 and Complexity Sciences Center and Physics Department, University

More information

arxiv: v3 [cs.fl] 16 Dec 2012

arxiv: v3 [cs.fl] 16 Dec 2012 Santa Fe Institute Working Paper 0--027 arxiv.org:0.0036 [cs.fl] Enumerating Finitary Processes arxiv:0.0036v3 [cs.fl] 6 Dec 202 Benjamin D. Johnson,, 2, James P. Crutchfield,, 2, 3, 4, Christopher J.

More information

WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME)

WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME) WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME ITAI BENJAMINI, GADY KOZMA, LÁSZLÓ LOVÁSZ, DAN ROMIK, AND GÁBOR TARDOS Abstract. We observe returns of a simple random wal on a finite graph to a fixed node,

More information

Lecture 20 : Markov Chains

Lecture 20 : Markov Chains CSCI 3560 Probability and Computing Instructor: Bogdan Chlebus Lecture 0 : Markov Chains We consider stochastic processes. A process represents a system that evolves through incremental changes called

More information

Information Symmetries in Irreversible Processes

Information Symmetries in Irreversible Processes Information Symmetries in Irreversible Processes Christopher J. Ellison John R. Mahoney R. G. James James P. Crutchfield Joerg Reichardt SFI WORKING PAPER: 0-07-08 SFI Working Papers contain accounts of

More information

A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources

A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources Wei Kang Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College

More information

Intrinsic Quantum Computation

Intrinsic Quantum Computation Intrinsic Quantum Computation James P. Crutchfield Karoline Wiesner SFI WORKING PAPER: 2006-11-045 SFI Working Papers contain accounts of scientific work of the author(s and do not necessarily represent

More information

Asymptotic Filtering and Entropy Rate of a Hidden Markov Process in the Rare Transitions Regime

Asymptotic Filtering and Entropy Rate of a Hidden Markov Process in the Rare Transitions Regime Asymptotic Filtering and Entropy Rate of a Hidden Marov Process in the Rare Transitions Regime Chandra Nair Dept. of Elect. Engg. Stanford University Stanford CA 94305, USA mchandra@stanford.edu Eri Ordentlich

More information

On Row-by-Row Coding for 2-D Constraints

On Row-by-Row Coding for 2-D Constraints On Row-by-Row Coding for 2-D Constraints Ido Tal Tuvi Etzion Ron M. Roth Computer Science Department, Technion, Haifa 32000, Israel. Email: {idotal, etzion, ronny}@cs.technion.ac.il Abstract A constant-rate

More information

Lecture notes for Analysis of Algorithms : Markov decision processes

Lecture notes for Analysis of Algorithms : Markov decision processes Lecture notes for Analysis of Algorithms : Markov decision processes Lecturer: Thomas Dueholm Hansen June 6, 013 Abstract We give an introduction to infinite-horizon Markov decision processes (MDPs) with

More information

3515ICT: Theory of Computation. Regular languages

3515ICT: Theory of Computation. Regular languages 3515ICT: Theory of Computation Regular languages Notation and concepts concerning alphabets, strings and languages, and identification of languages with problems (H, 1.5). Regular expressions (H, 3.1,

More information

Prediction, Retrodiction, and the Amount of Information Stored in the Present

Prediction, Retrodiction, and the Amount of Information Stored in the Present Prediction, Retrodiction, and the Amount of Information Stored in the Present Christopher J. Ellison John R. Mahoney James P. Crutchfield SFI WORKING PAPER: 2009-05-017 SFI Working Papers contain accounts

More information

arxiv: v1 [cs.it] 5 Sep 2008

arxiv: v1 [cs.it] 5 Sep 2008 1 arxiv:0809.1043v1 [cs.it] 5 Sep 2008 On Unique Decodability Marco Dalai, Riccardo Leonardi Abstract In this paper we propose a revisitation of the topic of unique decodability and of some fundamental

More information

Minimum Repair Bandwidth for Exact Regeneration in Distributed Storage

Minimum Repair Bandwidth for Exact Regeneration in Distributed Storage 1 Minimum Repair andwidth for Exact Regeneration in Distributed Storage Vivec R Cadambe, Syed A Jafar, Hamed Malei Electrical Engineering and Computer Science University of California Irvine, Irvine, California,

More information

Beyond the typical set: Fluctuations in Intrinsic Computation

Beyond the typical set: Fluctuations in Intrinsic Computation Beyond the typical set: Fluctuations in Intrinsic Computation Cina Aghamohammadi Complexity Sciences Center and Department of Physics, niversity of California at Davis, One Shields Avenue, Davis, CA 9566

More information

Upper Bounds on the Capacity of Binary Intermittent Communication

Upper Bounds on the Capacity of Binary Intermittent Communication Upper Bounds on the Capacity of Binary Intermittent Communication Mostafa Khoshnevisan and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame, Indiana 46556 Email:{mhoshne,

More information

Computational Mechanics of the Two Dimensional BTW Model

Computational Mechanics of the Two Dimensional BTW Model Computational Mechanics of the Two Dimensional BTW Model Rajesh Kommu kommu@physics.ucdavis.edu June 8, 2010 Abstract Some aspects of computational mechanics in two dimensions are investigated in this

More information

arxiv:quant-ph/ v4 28 Apr 2008

arxiv:quant-ph/ v4 28 Apr 2008 Santa Fe Institute Working Paper 06-09-031 arxiv.org/quant-ph/0608206 Computation in Finitary Stochastic and Quantum Processes Karoline Wiesner 1,2, and James P. Crutchfield 1,2, 1 Complexity Sciences

More information

MATRIX REPRESENTATIONS FOR MULTIPLICATIVE NESTED SUMS. 1. Introduction. The harmonic sums, defined by [BK99, eq. 4, p. 1] sign (i 1 ) n 1 (N) :=

MATRIX REPRESENTATIONS FOR MULTIPLICATIVE NESTED SUMS. 1. Introduction. The harmonic sums, defined by [BK99, eq. 4, p. 1] sign (i 1 ) n 1 (N) := MATRIX REPRESENTATIONS FOR MULTIPLICATIVE NESTED SUMS LIN JIU AND DIANE YAHUI SHI* Abstract We study the multiplicative nested sums which are generalizations of the harmonic sums and provide a calculation

More information

On bounded redundancy of universal codes

On bounded redundancy of universal codes On bounded redundancy of universal codes Łukasz Dębowski Institute of omputer Science, Polish Academy of Sciences ul. Jana Kazimierza 5, 01-248 Warszawa, Poland Abstract onsider stationary ergodic measures

More information

Structure or Noise? Susanne Still James P. Crutchfield SFI WORKING PAPER:

Structure or Noise? Susanne Still James P. Crutchfield SFI WORKING PAPER: Structure or Noise? Susanne Still James P. Crutchfield SFI WORKING PAPER: 2007-08-020 SFI Working Papers contain accounts of scientific work of the author(s) and do not necessarily represent the views

More information

Bayesian Structural Inference for Hidden Processes

Bayesian Structural Inference for Hidden Processes Santa Fe Institute Working Paper 13-09-027 arxiv:1309.1392 [stat.ml] Bayesian Structural Inference for Hidden Processes they occur in discrete-time and discrete-state time series. (The approach also applies

More information

Reconstruction Deconstruction:

Reconstruction Deconstruction: Reconstruction Deconstruction: A Brief History of Building Models of Nonlinear Dynamical Systems Jim Crutchfield Center for Computational Science & Engineering Physics Department University of California,

More information

Quantum Finite-State Machines

Quantum Finite-State Machines Quantum Finite-State Machines manda Young Department of Mathematics University of California, Davis amyoung@math.ucdavis.edu June, bstract Stochastic finite-state machines provide information about the

More information

The efficiency of identifying timed automata and the power of clocks

The efficiency of identifying timed automata and the power of clocks The efficiency of identifying timed automata and the power of clocks Sicco Verwer a,b,1,, Mathijs de Weerdt b, Cees Witteveen b a Eindhoven University of Technology, Department of Mathematics and Computer

More information

Computation at the Nanoscale:

Computation at the Nanoscale: Computation at the Nanoscale: thoughts on information processing in novel materials, molecules, & atoms Jim Crutchfield Complexity Sciences Center Physics Department University of California at Davis Telluride

More information

Minimizing response times and queue lengths in systems of parallel queues

Minimizing response times and queue lengths in systems of parallel queues Minimizing response times and queue lengths in systems of parallel queues Ger Koole Department of Mathematics and Computer Science, Vrije Universiteit, De Boelelaan 1081a, 1081 HV Amsterdam, The Netherlands

More information

Computing and Communications 2. Information Theory -Entropy

Computing and Communications 2. Information Theory -Entropy 1896 1920 1987 2006 Computing and Communications 2. Information Theory -Entropy Ying Cui Department of Electronic Engineering Shanghai Jiao Tong University, China 2017, Autumn 1 Outline Entropy Joint entropy

More information

Entropy Rate of Stochastic Processes

Entropy Rate of Stochastic Processes Entropy Rate of Stochastic Processes Timo Mulder tmamulder@gmail.com Jorn Peters jornpeters@gmail.com February 8, 205 The entropy rate of independent and identically distributed events can on average be

More information

On Supervisory Control of Concurrent Discrete-Event Systems

On Supervisory Control of Concurrent Discrete-Event Systems On Supervisory Control of Concurrent Discrete-Event Systems Yosef Willner Michael Heymann March 27, 2002 Abstract When a discrete-event system P consists of several subsystems P 1,..., P n that operate

More information

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1 Kraft s inequality An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if N 2 l i 1 Proof: Suppose that we have a tree code. Let l max = max{l 1,...,

More information

Stochastic Processes

Stochastic Processes qmc082.tex. Version of 30 September 2010. Lecture Notes on Quantum Mechanics No. 8 R. B. Griffiths References: Stochastic Processes CQT = R. B. Griffiths, Consistent Quantum Theory (Cambridge, 2002) DeGroot

More information

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim

More information

Training-Based Schemes are Suboptimal for High Rate Asynchronous Communication

Training-Based Schemes are Suboptimal for High Rate Asynchronous Communication Training-Based Schemes are Suboptimal for High Rate Asynchronous Communication The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation

More information

Stationary Probabilities of Markov Chains with Upper Hessenberg Transition Matrices

Stationary Probabilities of Markov Chains with Upper Hessenberg Transition Matrices Stationary Probabilities of Marov Chains with Upper Hessenberg Transition Matrices Y. Quennel ZHAO Department of Mathematics and Statistics University of Winnipeg Winnipeg, Manitoba CANADA R3B 2E9 Susan

More information

An Extended Fano s Inequality for the Finite Blocklength Coding

An Extended Fano s Inequality for the Finite Blocklength Coding An Extended Fano s Inequality for the Finite Bloclength Coding Yunquan Dong, Pingyi Fan {dongyq8@mails,fpy@mail}.tsinghua.edu.cn Department of Electronic Engineering, Tsinghua University, Beijing, P.R.

More information

Essentials on the Analysis of Randomized Algorithms

Essentials on the Analysis of Randomized Algorithms Essentials on the Analysis of Randomized Algorithms Dimitris Diochnos Feb 0, 2009 Abstract These notes were written with Monte Carlo algorithms primarily in mind. Topics covered are basic (discrete) random

More information

Adaptive State Feedback Nash Strategies for Linear Quadratic Discrete-Time Games

Adaptive State Feedback Nash Strategies for Linear Quadratic Discrete-Time Games Adaptive State Feedbac Nash Strategies for Linear Quadratic Discrete-Time Games Dan Shen and Jose B. Cruz, Jr. Intelligent Automation Inc., Rocville, MD 2858 USA (email: dshen@i-a-i.com). The Ohio State

More information

NORMS ON SPACE OF MATRICES

NORMS ON SPACE OF MATRICES NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system

More information

EECS 144/244: Fundamental Algorithms for System Modeling, Analysis, and Optimization

EECS 144/244: Fundamental Algorithms for System Modeling, Analysis, and Optimization EECS 144/244: Fundamental Algorithms for System Modeling, Analysis, and Optimization Discrete Systems Lecture: Automata, State machines, Circuits Stavros Tripakis University of California, Berkeley Stavros

More information

UTILIZING PRIOR KNOWLEDGE IN ROBUST OPTIMAL EXPERIMENT DESIGN. EE & CS, The University of Newcastle, Australia EE, Technion, Israel.

UTILIZING PRIOR KNOWLEDGE IN ROBUST OPTIMAL EXPERIMENT DESIGN. EE & CS, The University of Newcastle, Australia EE, Technion, Israel. UTILIZING PRIOR KNOWLEDGE IN ROBUST OPTIMAL EXPERIMENT DESIGN Graham C. Goodwin James S. Welsh Arie Feuer Milan Depich EE & CS, The University of Newcastle, Australia 38. EE, Technion, Israel. Abstract:

More information

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING ERIC SHANG Abstract. This paper provides an introduction to Markov chains and their basic classifications and interesting properties. After establishing

More information

Stochastic Shortest Path Problems

Stochastic Shortest Path Problems Chapter 8 Stochastic Shortest Path Problems 1 In this chapter, we study a stochastic version of the shortest path problem of chapter 2, where only probabilities of transitions along different arcs can

More information

EQUIVALENCE OF TOPOLOGIES AND BOREL FIELDS FOR COUNTABLY-HILBERT SPACES

EQUIVALENCE OF TOPOLOGIES AND BOREL FIELDS FOR COUNTABLY-HILBERT SPACES EQUIVALENCE OF TOPOLOGIES AND BOREL FIELDS FOR COUNTABLY-HILBERT SPACES JEREMY J. BECNEL Abstract. We examine the main topologies wea, strong, and inductive placed on the dual of a countably-normed space

More information

Supplemental for Spectral Algorithm For Latent Tree Graphical Models

Supplemental for Spectral Algorithm For Latent Tree Graphical Models Supplemental for Spectral Algorithm For Latent Tree Graphical Models Ankur P. Parikh, Le Song, Eric P. Xing The supplemental contains 3 main things. 1. The first is network plots of the latent variable

More information

Lecture 4: Probability and Discrete Random Variables

Lecture 4: Probability and Discrete Random Variables Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 4: Probability and Discrete Random Variables Wednesday, January 21, 2009 Lecturer: Atri Rudra Scribe: Anonymous 1

More information

Markov Chains and Stochastic Sampling

Markov Chains and Stochastic Sampling Part I Markov Chains and Stochastic Sampling 1 Markov Chains and Random Walks on Graphs 1.1 Structure of Finite Markov Chains We shall only consider Markov chains with a finite, but usually very large,

More information

Lecture 3: Markov chains.

Lecture 3: Markov chains. 1 BIOINFORMATIK II PROBABILITY & STATISTICS Summer semester 2008 The University of Zürich and ETH Zürich Lecture 3: Markov chains. Prof. Andrew Barbour Dr. Nicolas Pétrélis Adapted from a course by Dr.

More information

Kleene Algebras and Algebraic Path Problems

Kleene Algebras and Algebraic Path Problems Kleene Algebras and Algebraic Path Problems Davis Foote May 8, 015 1 Regular Languages 1.1 Deterministic Finite Automata A deterministic finite automaton (DFA) is a model of computation that can simulate

More information

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Chapter 14 SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Today we continue the topic of low-dimensional approximation to datasets and matrices. Last time we saw the singular

More information

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST Information Theory Lecture 5 Entropy rate and Markov sources STEFAN HÖST Universal Source Coding Huffman coding is optimal, what is the problem? In the previous coding schemes (Huffman and Shannon-Fano)it

More information

Markov Processes Hamid R. Rabiee

Markov Processes Hamid R. Rabiee Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete

More information

On the exact bit error probability for Viterbi decoding of convolutional codes

On the exact bit error probability for Viterbi decoding of convolutional codes On the exact bit error probability for Viterbi decoding of convolutional codes Irina E. Bocharova, Florian Hug, Rolf Johannesson, and Boris D. Kudryashov Dept. of Information Systems Dept. of Electrical

More information

Welsh s problem on the number of bases of matroids

Welsh s problem on the number of bases of matroids Welsh s problem on the number of bases of matroids Edward S. T. Fan 1 and Tony W. H. Wong 2 1 Department of Mathematics, California Institute of Technology 2 Department of Mathematics, Kutztown University

More information

http://www.math.uah.edu/stat/markov/.xhtml 1 of 9 7/16/2009 7:20 AM Virtual Laboratories > 16. Markov Chains > 1 2 3 4 5 6 7 8 9 10 11 12 1. A Markov process is a random process in which the future is

More information

Entropy and Ergodic Theory Lecture 3: The meaning of entropy in information theory

Entropy and Ergodic Theory Lecture 3: The meaning of entropy in information theory Entropy and Ergodic Theory Lecture 3: The meaning of entropy in information theory 1 The intuitive meaning of entropy Modern information theory was born in Shannon s 1948 paper A Mathematical Theory of

More information

On the Average Complexity of Brzozowski s Algorithm for Deterministic Automata with a Small Number of Final States

On the Average Complexity of Brzozowski s Algorithm for Deterministic Automata with a Small Number of Final States On the Average Complexity of Brzozowski s Algorithm for Deterministic Automata with a Small Number of Final States Sven De Felice 1 and Cyril Nicaud 2 1 LIAFA, Université Paris Diderot - Paris 7 & CNRS

More information

A Piggybacking Design Framework for Read-and Download-efficient Distributed Storage Codes

A Piggybacking Design Framework for Read-and Download-efficient Distributed Storage Codes A Piggybacing Design Framewor for Read-and Download-efficient Distributed Storage Codes K V Rashmi, Nihar B Shah, Kannan Ramchandran, Fellow, IEEE Department of Electrical Engineering and Computer Sciences

More information

An Introduction to Entropy and Subshifts of. Finite Type

An Introduction to Entropy and Subshifts of. Finite Type An Introduction to Entropy and Subshifts of Finite Type Abby Pekoske Department of Mathematics Oregon State University pekoskea@math.oregonstate.edu August 4, 2015 Abstract This work gives an overview

More information

Stochastic modelling of epidemic spread

Stochastic modelling of epidemic spread Stochastic modelling of epidemic spread Julien Arino Centre for Research on Inner City Health St Michael s Hospital Toronto On leave from Department of Mathematics University of Manitoba Julien Arino@umanitoba.ca

More information

arxiv: v1 [cond-mat.stat-mech] 9 Dec 2014

arxiv: v1 [cond-mat.stat-mech] 9 Dec 2014 Santa Fe Institute Working Paper 4-2-XXX arxiv.org:42.xxxx [cond-mat.stat-mech] Circumventing the Curse of Dimensionality in Prediction: Causal Rate-Distortion for Infinite-Order Markov Processes arxiv:42.2859v

More information

Structured Stochastic Uncertainty

Structured Stochastic Uncertainty Structured Stochastic Uncertainty Bassam Bamieh Department of Mechanical Engineering University of California at Santa Barbara Santa Barbara, CA, 9306 bamieh@engineeringucsbedu Abstract We consider linear

More information

1 The independent set problem

1 The independent set problem ORF 523 Lecture 11 Spring 2016, Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Tuesday, March 29, 2016 When in doubt on the accuracy of these notes, please cross chec with the instructor

More information

Stochastic modelling of epidemic spread

Stochastic modelling of epidemic spread Stochastic modelling of epidemic spread Julien Arino Department of Mathematics University of Manitoba Winnipeg Julien Arino@umanitoba.ca 19 May 2012 1 Introduction 2 Stochastic processes 3 The SIS model

More information

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states. Chapter 8 Finite Markov Chains A discrete system is characterized by a set V of states and transitions between the states. V is referred to as the state space. We think of the transitions as occurring

More information

Pattern Discovery in Time Series, Part I: Theory, Algorithm, Analysis, and Convergence

Pattern Discovery in Time Series, Part I: Theory, Algorithm, Analysis, and Convergence Journal of Machine earning Research? (2002)?? Submitted 0/28; Published?/02 Pattern Discovery in Time Series, Part I: Theory, Algorithm, Analysis, and Convergence Cosma Rohilla Shalizi shalizi@santafe.edu

More information

Markov chains and the number of occurrences of a word in a sequence ( , 11.1,2,4,6)

Markov chains and the number of occurrences of a word in a sequence ( , 11.1,2,4,6) Markov chains and the number of occurrences of a word in a sequence (4.5 4.9,.,2,4,6) Prof. Tesler Math 283 Fall 208 Prof. Tesler Markov Chains Math 283 / Fall 208 / 44 Locating overlapping occurrences

More information

An algebraic characterization of unary two-way transducers

An algebraic characterization of unary two-way transducers An algebraic characterization of unary two-way transducers (Extended Abstract) Christian Choffrut 1 and Bruno Guillon 1 LIAFA, CNRS and Université Paris 7 Denis Diderot, France. Abstract. Two-way transducers

More information

Part III. 10 Topological Space Basics. Topological Spaces

Part III. 10 Topological Space Basics. Topological Spaces Part III 10 Topological Space Basics Topological Spaces Using the metric space results above as motivation we will axiomatize the notion of being an open set to more general settings. Definition 10.1.

More information

Markov Chains, Stochastic Processes, and Matrix Decompositions

Markov Chains, Stochastic Processes, and Matrix Decompositions Markov Chains, Stochastic Processes, and Matrix Decompositions 5 May 2014 Outline 1 Markov Chains Outline 1 Markov Chains 2 Introduction Perron-Frobenius Matrix Decompositions and Markov Chains Spectral

More information

The 2-valued case of makespan minimization with assignment constraints

The 2-valued case of makespan minimization with assignment constraints The 2-valued case of maespan minimization with assignment constraints Stavros G. Kolliopoulos Yannis Moysoglou Abstract We consider the following special case of minimizing maespan. A set of jobs J and

More information

Modeling and Stability Analysis of a Communication Network System

Modeling and Stability Analysis of a Communication Network System Modeling and Stability Analysis of a Communication Network System Zvi Retchkiman Königsberg Instituto Politecnico Nacional e-mail: mzvi@cic.ipn.mx Abstract In this work, the modeling and stability problem

More information

MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing

MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing Afonso S. Bandeira April 9, 2015 1 The Johnson-Lindenstrauss Lemma Suppose one has n points, X = {x 1,..., x n }, in R d with d very

More information

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16 EE539R: Problem Set 4 Assigned: 3/08/6, Due: 07/09/6. Cover and Thomas: Problem 3.5 Sets defined by probabilities: Define the set C n (t = {x n : P X n(x n 2 nt } (a We have = P X n(x n P X n(x n 2 nt

More information

The Perron Frobenius theorem and the Hilbert metric

The Perron Frobenius theorem and the Hilbert metric The Perron Frobenius theorem and the Hilbert metric Vaughn Climenhaga April 7, 03 In the last post, we introduced basic properties of convex cones and the Hilbert metric. In this post, we loo at how these

More information

Reconstruction. Reading for this lecture: Lecture Notes.

Reconstruction. Reading for this lecture: Lecture Notes. ɛm Reconstruction Reading for this lecture: Lecture Notes. The Learning Channel... ɛ -Machine of a Process: Intrinsic representation! Predictive (or causal) equivalence relation: s s Pr( S S= s ) = Pr(

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours

More information

1 Stochastic Dynamic Programming

1 Stochastic Dynamic Programming 1 Stochastic Dynamic Programming Formally, a stochastic dynamic program has the same components as a deterministic one; the only modification is to the state transition equation. When events in the future

More information

Maximizing the numerical radii of matrices by permuting their entries

Maximizing the numerical radii of matrices by permuting their entries Maximizing the numerical radii of matrices by permuting their entries Wai-Shun Cheung and Chi-Kwong Li Dedicated to Professor Pei Yuan Wu. Abstract Let A be an n n complex matrix such that every row and

More information

Quantum versus Classical Measures of Complexity on Classical Information

Quantum versus Classical Measures of Complexity on Classical Information Quantum versus Classical Measures of Complexity on Classical Information Lock Yue Chew, PhD Division of Physics and Applied Physics School of Physical & Mathematical Sciences & Complexity Institute Nanyang

More information

CONCATENATION AND KLEENE STAR ON DETERMINISTIC FINITE AUTOMATA

CONCATENATION AND KLEENE STAR ON DETERMINISTIC FINITE AUTOMATA 1 CONCATENATION AND KLEENE STAR ON DETERMINISTIC FINITE AUTOMATA GUO-QIANG ZHANG, XIANGNAN ZHOU, ROBERT FRASER, LICONG CUI Department of Electrical Engineering and Computer Science, Case Western Reserve

More information

Combinatorial Interpretations of a Generalization of the Genocchi Numbers

Combinatorial Interpretations of a Generalization of the Genocchi Numbers 1 2 3 47 6 23 11 Journal of Integer Sequences, Vol. 7 (2004), Article 04.3.6 Combinatorial Interpretations of a Generalization of the Genocchi Numbers Michael Domaratzki Jodrey School of Computer Science

More information

Unsupervised Inductive Learning In Symbolic Sequences via Recursive Identification of Self-Similar Semantics

Unsupervised Inductive Learning In Symbolic Sequences via Recursive Identification of Self-Similar Semantics 2011 American Control Conference on O'Farrell Street, San Francisco, CA, USA June 29 - July 01, 2011 Unsupervised Inductive Learning In Symbolic Sequences via Recursive Identification of Self-Similar Semantics

More information

No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1.

No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1. Stationary Distributions Monday, September 28, 2015 2:02 PM No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1. Homework 1 due Friday, October 2 at 5 PM strongly

More information

cse303 ELEMENTS OF THE THEORY OF COMPUTATION Professor Anita Wasilewska

cse303 ELEMENTS OF THE THEORY OF COMPUTATION Professor Anita Wasilewska cse303 ELEMENTS OF THE THEORY OF COMPUTATION Professor Anita Wasilewska LECTURE 6 CHAPTER 2 FINITE AUTOMATA 2. Nondeterministic Finite Automata NFA 3. Finite Automata and Regular Expressions 4. Languages

More information