Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 1. Random Compositions. Evangelos Kranakis School of Computer Science Carleton University

Size: px
Start display at page:

Download "Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 1. Random Compositions. Evangelos Kranakis School of Computer Science Carleton University"

Transcription

1 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 1 Random Compositions Evangelos Kranakis School of Computer Science Carleton University

2 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 2 Outline 1. Random functions and iterations. 2. Iterated random compositions. 3. Markov chain for random compositions. 4. Waiting time until absorption: (a) Lower bound. (b) Upper bound. 5. Open Problems.

3 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 3 Random Functions For n integer, [n] = {1, 2,..., n}. F(m, n) is the space of functions f : [m] [n]. F(n) := F(n, n). A function f is chosen from F(n) randomly with the uniform distribution. Range of f is defined by Range(f) := {y : x [n]f(x) = y}. Question: What is the expected size of the range of a random function?

4 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 4 Values of Random Functions For random function f : [m] [n] and given x [m], y [n], Pr[f(x) = y] = nm 1 n m = 1 n. For random function f : [m] [n] and given y [n], the size of the inverse image f 1 ({y}) satisfies, Pr[ f 1 ({y}) = k] = ( ) m (n 1) m k k n m, for k = 0, 1,..., n.

5 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 5 Graph of a Random Function Given f F(n) the graph G(f) has [n] as its set of vertices and directed edges (x, y) E f(x) = y. Define the k-the iterate f k of f by f 0 = identity function and f k := f f k 1. The orbit (or cycle) of x is defined by O f (x) := {x, f(x), f(f(x)),..., f n 1 (x)}. If L n is the r.v. that counts the length of an orbit it can be shown (Sachkov 1997) Pr[L n = j] = E[L n ] = n k=j n j j=1 (n) k, j = 1,..., n nk+1 n k=j (n) k n k+1

6 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 6 Connected Components of a Random Function The connected components of the graph G(f) consist of the orbits and trees attached to them. f(f(f(x))) f(f(x)) x f(x) The elements of a cycle are called cyclic elements of G(f). The distance of an element from its cycle is called its height.

7 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 7 # of Connected Components of a Random Function Let K n be the r.v. that counts the number of connected components of a random function in F(n). A result of Stepanov (1966) states that Pr[K n = j] = n k=j ( ) n 1 S(k, j), j = 1, 2,..., n k 1 n k E[K n ] = 1 2 V ar[k n ] = 1 2 log n(1 + o(1)) log n(1 + o(1)), where S(k, j) is the # of ways to partition a k element set into j disjoint subsets, a Stirling number of the 2nd kind.

8 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 8 # of Cyclic Elements of a Random Function Let Z n be the r.v. that counts the number of distinct cyclic elements of a random function in F(n). A result of Harris (1973) states that Pr[Z n = k] = k(n) k, k = 0, 1,..., n nk+1 πn E[Z n ] = (1 + o(1)) 2 ( V ar[z n ] = 2 π ) n(1 + o(1)), 2 where (n) k = n(n 1) (n k + 1).

9 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 9 Expected Size of Range of a Random Function f : [m] [n] Consider the indicator function I i : I i = 1 if i Range(f), and I i = 0, otherwise. Then we have [ n ] E[ Range(f) ] = E I i = i=1 n E[I i ] i=1 = ne[i 1 ] = n Pr[I 1 = 1] = n(1 Pr[1 Range(f)]) ( ( = n ) m ). n

10 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 10 Random Functions Cause Shrinkage! Given a random function f F(n), ( E[ Range(f) ] = n n 1 n) 1 n n(1 1/e). [n] [n] f Range(f) n n(e 1)/e

11 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 11 What Causes Shrinkage? Consider a random function f : [m] [n] (with m n): shrinkage is caused by collisions among the elements f(1), f(2),..., f(m), i.e., f(x) = f(y), for some x y. Pr[collision] = Pr[ x y(f(x) = f(y))] = 1 Pr[ x y(f(x) f(y))] = 1 = 1 m 1 i=0 m 1 i=0 n i n ( 1 i ) n (Birthday paradox) Hence, the bigger the m the higher the probability of a collision! We see later that this causes the Markov Chain to skip large states!

12 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 12 Variance of Range(f) for a Random Function f : [m] [n] Let X be the r.v. that counts the size of Range(f) and U = n X. Consider the indicator function I i : I i = 1 if i Range(f), and = 1, otherwise. Observe that I i V ar(x) = V ar(u) = E[U 2 ] E[U] 2 = i j E[I ii j] + i E[I i] E[U] 2 = n(n 1) = n 2 ( ( 1 2 n) m +n ( 1 n) 2 m ( + n 1 1 ) m ( n ) 2m n n ( 1 1 ) ) 2m n (( 1 1 n) m ( 1 2 ) m ). n

13 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 13 Compositions of Random Functions k functions f 1, f 2,..., f k are chosen from F(n) randomly and independently with the uniform distribution. Let f (k) := f k f k 1 f 1. Convention: f (0) = identity function on [n]. f (k) is called a random composition. Question: Since the expected size of the range of a random function is a constant fraction of the size of its domain, how long does it take for a random composition to become constant?

14 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 14 Iterations of Random Functions Model proposed by (Diaconis & Freedman 1999) There is a state space S. There is a family of functions F such that each F F maps the state space into itself F : S S. There is a probability distribution µ on F. If the chain is at state s S it moves to state F (s) by choosing F F at random. The process starts with F 0 and inductively defines X t+1 = F (X t ) where F is a random function F F.

15 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 15 Example I: Linear Affine Functions The state space S is the real line. F = {F +, F } has just two functions defined as follows where 0 < a < 1. F + : R R : x F + (x) = ax + 1 F : R R : x F (x) = ax 1, The probability distribution µ on F is µ(f + ) = µ(f ) = 1/2. Let ξ i = ±1 with probability 1/2, respectively. The process starts with ξ 0 and inductively defines X t+1 = F (X t ) where F is a random function F F. Clearly, X t+1 = ax t + ξ t and the stationary distribution X = ξ 1 + aξ 2 + a 2 ξ 3 + converges since a < 1.

16 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 16 Example II: d-dimensional Affine Functions The state space S is the d-dimensional space R d. F contains a set of functions of the form F : R d R d : x F (x) = Ax + B, A is an d d matrix and B is a d 1 vector. F can be identified with a set of pairs (A, B) of matrices and we have a probability distribution µ on F. The basic chain is X t+1 = A t X t + B t, where A t is an d d matrix and B t a d 1 vector, and (A t, B t ) are idependent and identically distributed. This has applications in fractal geometry.

17 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 17 Example III: Random Compositions States specify the size of the range of a random composition and these states form the state space S. A family of functions { f : f F(n)} maps the state space into itself as follows: Given a function g F(n) already in state s, s f (s) := size of range of f g. The probability distribution on { f : f F(n)} is uniform. If the chain is at state s S it moves to state f (s) by choosing f at random. The process starts with f 0 (identity function) and inductively defines X t+1 = f (X t ) where f is a random function f F(n).

18 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 18 Example IV: Random Compositions of Hashes States specify the size of the range of a random composition and these states form the state space S. H(n) is the set of functions f : [n] [n/2]. A family of functions { h : h H(n)} maps the state space into itself as follows: Given a function g H(n) already in state s, s h (s) := size of range of h g. The probability distribution on { h : h H(n)} is uniform. If the chain is at state s S it moves to state f (s) by choosing f at random. The process starts with a given function h 0 and inductively defines X t+1 = h (X t ) where h is a random function h H(n).

19 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 19 Waiting Time until Absorption For t > 0, we are in state s r iff Range(f (t) ) = r. τ r = {t : Range(f (t) ) = r} is the amount of time in state s r. State s r is visited if τ r > 0. and T is the set of states that are actually visited. Let T be the smallest t for which f (t) is constant, i.e., T = n r=1 τ r s s s n s n 1 s n 2 i 1

20 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 20 How is T Computed The Markov chain starts with the identity function f (0) at time 0 in state s n. By the nature of the problem, states are visited in non-increasing order. It is possible that states may be skipped. Eventually it reaches s 1, the absorbing state. T is really the time it takes to reach the absorbing state s 1. The main result is the following Theorem: E[T ] = 2n(1 + o(1)), as n.

21 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 21 Transition Probabilities For j i, what is the probability f (t) s j given that f (t 1) s i? Given that f (t 1) s i, how many functions f are there such that f f (t 1) has j elements in its range? The are ( n j) ways to choose the range of f f (t 1), S(i, j)j! ways to map the i-element range of f (t 1) onto a given j element set, where S(i, j) is the # of ways to partition a i element set into j disjoint subsets, a Stirling number of the 2nd kind, and n n i ways to map them into [n]. It follows that p(i, j) := Pr[f (t) s j f (t 1) s i ] = ( ) n S(i, j)j!, j n i

22 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 22 Upper Bound on Stirling Numbers S(i, j) S(i, j) = # of ways to partition a i element set into j subsets. Prove by induction S(i, j) (2j) i. For i = 1: S(1, j) 2j We have that S(i, j) = S(i 1, j 1) + js(i 1, j) (Identity) (2(j 1)) i 1 + j(2j) i 1 (Induction) ( (j 1) = (2j) i i 1 2j i + 1 ) 2 (2j) i.

23 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 23 Eigenvalues of the Transition Matrix The transition matrix P := (p(i, j) i,j is lower diagonal. Eigenvalues are the diagonal elements of the matrix, i.e., λ r = p(r, r) ( ) n S(r, r)r! = r n r = r 1 i=1 = 1 ( 1 i n ) ( r ( ) 2) r 4 n + O. n 2 Note: 1 > 1 1 n = λ 2 λ n > 0 and 1 1 λ r+d n 1

24 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 24 Transition Probabilities for Affine Matrices Consider d d matrices over, say, the finite field Z p and let A (0) := I be the identity matrix. A (t) = A t A (t 1), where A t is a random matrix. For t > 0 we are in state r iff rank(a (t) ) = r. τ r = {t : rank(a (t) ) = r} is the amount of time in state s r. Open Question: For j i, compute the transition probabilities p(i, j) := Pr[rank(A (t) ) = j rank(a (t 1) ) = i]. This is equivalent to computing the number of matrices B such that rank(ba) = j, given that rank(a) = i.

25 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 25 Lower Bound on E[T ] We have the identity [ n ] E[T ] = E τ r = = It remains r=2 n E [τ r ] r=2 n E [τ r τ r > 0] Pr[τ r > 0]. r=2 to compute E [τ r τ r > 0], and give a lower bound on Pr[τ r > 0].

26 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 26 Computing E [τ r τ r > 0] This is the expected amount of time you stay in state τ r, given that you visit it? Given τ r > 0, τ r follows the geometric distribution, with probability of success p(r, r) = λ r. E [τ r τ r > 0] = = t=1 1 1 λ r = n ( r 2) tλ t 1 r (1 λ r ) ( 1 + O ( )) r 2 n

27 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 27 Estimating Pr[τ r = 0] We give an upper bound on Pr[τ r = 0]. If τ r = 0 then the state s r is never visited. Therefore there must exist a transition for some positive integers d, j. s r+d s r j, In fact, before moving to state s r j it may stay in state s r+d a number of times t = 0, 1, 2, 3,.... Therefore we must take into account how long we stay in state s r+d given that this state is visited.

28 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 28 We can show that Pr[τ r = 0] = = n r d=1 n r d=1 n r d=1 Upper Bound on Pr[τ r = 0] r 1 j=1 r 1 j=1 r 1 j=1 (n 1) Pr[τ r+d > 0] Pr[τ r+d > 0] p(r + d, r j)p(r + d, r + d) t t=0 p(r + d, r j) 1 λ r+d ( n ) S(r + d, r j)(r j)! r j n r+d (1 λ r+d ) n r d=1 1 n d r 1 j=1 S(r + d, r j) n j Recall: 1) Pr[τ r+d > 0] 1, 2) 1 > 1 1 n = λ 2 λ n > 0 and 1 1 λ r+d n 1

29 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 29 Lower Bound on Pr[τ r > 0] Hence we obtain for r log log n Pr[τ r > 0] 1 (n 1) 1 (n 1) n r d=1 n r d=1 n r 1 n d 1 n d r 1 j=1 r 1 j=1 1 r(2r) r+d 1 (n 1) n d n d=1 ( ) (2l) l+2 1 O n = 1 o(1). S(r + d, r j) n j (2(r j)) r+d n j

30 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 30 Proving the Lower Bound: E[T ] 2n(1 + o(1)) We have the idequality E[T ] l E [τ r τ r > 0] Pr[τ r > 0] r=1 l r=2 = 2n n ( r 2) l r=2 1 r(r 1) l ( 1 = 2n r 1 1 ) r r=2 ( = 2n 1 1 ) l This completes the proof of the lower bound.

31 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 31 Idea of Proof of Upper Bound Recall that T = n m=2 τ m. We split E[T ] into three sums n E[T ] = E [τ m τ m > 0] Pr[τ m > 0] whereby = m=2 ξ 1 m=2 + ξ 1 = ξ 2 = ξ 2 m=ξ n log n n log 2 n n m=ξ 2 +1 and make upper bound estimates on each of them.

32 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 32 Observe that 1st Sum: = ξ 1 m=2 E [τ m τ m > 0] Pr[τ m > 0] 1st Sum = = ξ 1 m=2 ξ 1 m=2 ξ 1 m=2 2n 1 1 λ m 1 ( m 2 ) n + O(m4 /n 2 ) ( m 2 ξ 1 m=2 n ) + O(m4 /n) 1 m(m 1) = 2n(1 1/ξ 1 ) = 2n(1 + o(1))

33 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 33 2nd Sum: = ξ 2 m=ξ 1 +1 E [τ m τ m > 0] Pr[τ m > 0] Observe that λ ξ1 = log n + O(1/ log2 n). 2nd Sum = ξ 2 m=ξ λ ξ1 1 1 λ m = O(ξ 2 log n) = O(n/ log n) = o(1). ξ 2 m=ξ

34 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 34 3rd Sum: = n m=ξ 2 +1 E [τ m τ m > 0] Pr[τ m > 0] Observe that max m>ξ2 1 1 λ m = 1 1 λ ξ2 +1. Hence, 3rd Sum n 1 Pr[τ m > 0] 1 λ m m=ξ n Pr[τ m > 0] 1 λ ξ2 +1 m=ξ n 1 exp ( ( ξ 22 ) ) Pr[τ m > 0] /n n 2 Pr[τ m > 0], m=ξ 2 +1 m=ξ 2 +1 for n large enough. It remains to bound the RHS above.

35 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 35 On Skipping Large States To deal with n m=ξ 2 +1 Pr[τ m > 0] we will show that every hit (i.e., τ m > 0) is followed by many (i.e., β) misses (i.e., τ m δ = 0, forall δ such that 1 δ β) Let us define β := β(n) = 1 2 ( ξ 2 n + n ( 1 1 ) ) ξ2 n Observe that β(n) >> n log 4 n. Suppose we are in state s m at time t 1, i.e., Range(f (t 1) ) = m, and select the next function f t at random.

36 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 36 Claim 1: If B > β then τ m δ = 0, for 1 δ β. Let h be the restriction of f t to the range of f (t 1). Let R be the cardinality of the range of h, and B = m R. To prove the claim notice that B > β m R > β R < m β and Range(f (t) ) = f (t) ([n]) = f t (f (t 1) ([n])) = h(range(f (t 1) )) = R.

37 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 37 Observe that Claim 2: E[B] ξ 2 n + n(1 1 n )ξ 2 >> n log 4 n. E[B] = E[m R] = m E[R] = m n + n ( 1 1 ) m n > ξ 2 n + n(1 1 n )ξ 2 = 2β n >> log 4 n. Also ( recall, V ar(b) = (1 ) n 2 2 m ( ) ) n 1 1 2m n + n (( 1 1 n) m ( 1 2 n) m ) = O(m).

38 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 38 Claim 4: Pr[ 1 δ β(τ m δ = 0) τ m > 0] By Chebeshev s Inequality and since m > ξ 2 we have Pr[ B β] Pr [ B 12 ] E[B] 4V ar(b) (E[B]) 2 ( m log 8 n = O = o(1). Hence, Pr[ B > β] = 1 o(1). It follows, by Claim 1, n 2 Pr[ 1 δ β(τ m δ = 0) τ m > 0] = 1 o(1). )

39 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 39 Back to the 3rd Sum: = n m=ξ 2 +1 E [τ m τ m > 0] Pr[τ m > 0] Define χ m = 1 if τ m > 0 and χ m = 0, otherwise. Recall that n 3rd Sum 2 Pr[τ m > 0] = 2 = 2E m=ξ 2 +1 n m=ξ 2 +1 = 2E[V ], n E[χ m ] χ m m=ξ 2 +1 where V := n m=ξ 2 +1 χ m. Define W := n m=ξ 2 +1 (1 χ m) and observe that V + W = n ξ 2.

40 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 40 Back to the 3rd Sum: = n m=ξ 2 +1 E [τ m τ m > 0] Pr[τ m > 0] Note that if τ m > 0 and 1 δ β(τ m δ = 0) then these β missed states contribute exactly β to W. Hence, if we define J m = χ m β δ=1 (1 χ m δ) then W β m>ξ 2 J m. Hence, E[W ] β m>ξ 2 E[J m ] = β m>ξ 2 Pr[J m = 1] = β m>ξ 2 Pr[τ m > 0] Pr[ 1 δ β(τ m δ = 0) τ m > 0] β(1 + o(1)) m>ξ 2 Pr[τ m > 0] = β(1 + o(1))e[v ].

41 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 41 Back to the 3rd Sum: = n m=ξ 2 +1 E [τ m τ m > 0] Pr[τ m > 0] It follows that Hence, E[V ] = n ξ 2 E[W ]) n ξ 2 β(1 + o(1))e[v ]) 3rd Sum 2E[V ] 2(n ξ 2 ) 1 + β(1 + o(1)) = O(log 4 n) = o(n). This completes the proof of the theorem.

42 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 42 Another Idea: Hitting Times For r = 1, 2,..., n, let t r be the (hitting) time that the chain spends in transient state r until absorbtion by state 1. Let the (n 1) (n 1) matrix Q be obtained from the transition matrix P by removing the first row and column. It is easy to see that t 1 = 1 and t r = 1 r=1 + r r =2 p(r, r )t r If I is the unit matrix and t is the vector of hitting times then t = I + Qt, which is equivalent to t(i Q) = I. So to compute the hitting times it is enough to compute the inverse of the matrix I Q. Easy to prove: if lim k Q k = 0 then (I Q) 1 = k=0 Qk. Speed of convergence depends on 2nd largest eigenvalue!

43 Dalal-Schmutz (2002) and Diaconis-Freedman (1999) 43 Open Questions Consider the space H(n) := F(n, n/2) of hash functions. Consider different probability distributions on F(n). The reason is that in practice one has preference over certain types of random functions. Consider a space {(A, B)} of pairs of d d (random) matrices and the functions x Ax + B. States are determined by the rank of a random matrix. This problem is equivalent to estimating the time T until the product of T random matrices is 0.

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time

More information

Stat 516, Homework 1

Stat 516, Homework 1 Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball

More information

Solution Set for Homework #1

Solution Set for Homework #1 CS 683 Spring 07 Learning, Games, and Electronic Markets Solution Set for Homework #1 1. Suppose x and y are real numbers and x > y. Prove that e x > ex e y x y > e y. Solution: Let f(s = e s. By the mean

More information

18.175: Lecture 30 Markov chains

18.175: Lecture 30 Markov chains 18.175: Lecture 30 Markov chains Scott Sheffield MIT Outline Review what you know about finite state Markov chains Finite state ergodicity and stationarity More general setup Outline Review what you know

More information

STOCHASTIC PROCESSES Basic notions

STOCHASTIC PROCESSES Basic notions J. Virtamo 38.3143 Queueing Theory / Stochastic processes 1 STOCHASTIC PROCESSES Basic notions Often the systems we consider evolve in time and we are interested in their dynamic behaviour, usually involving

More information

Markov Chains Handout for Stat 110

Markov Chains Handout for Stat 110 Markov Chains Handout for Stat 0 Prof. Joe Blitzstein (Harvard Statistics Department) Introduction Markov chains were first introduced in 906 by Andrey Markov, with the goal of showing that the Law of

More information

Analysis of Algorithms I: Perfect Hashing

Analysis of Algorithms I: Perfect Hashing Analysis of Algorithms I: Perfect Hashing Xi Chen Columbia University Goal: Let U = {0, 1,..., p 1} be a huge universe set. Given a static subset V U of n keys (here static means we will never change the

More information

2. Transience and Recurrence

2. Transience and Recurrence Virtual Laboratories > 15. Markov Chains > 1 2 3 4 5 6 7 8 9 10 11 12 2. Transience and Recurrence The study of Markov chains, particularly the limiting behavior, depends critically on the random times

More information

Lecture 20 : Markov Chains

Lecture 20 : Markov Chains CSCI 3560 Probability and Computing Instructor: Bogdan Chlebus Lecture 0 : Markov Chains We consider stochastic processes. A process represents a system that evolves through incremental changes called

More information

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321 Lecture 11: Introduction to Markov Chains Copyright G. Caire (Sample Lectures) 321 Discrete-time random processes A sequence of RVs indexed by a variable n 2 {0, 1, 2,...} forms a discretetime random process

More information

Lecture 9 Classification of States

Lecture 9 Classification of States Lecture 9: Classification of States of 27 Course: M32K Intro to Stochastic Processes Term: Fall 204 Instructor: Gordan Zitkovic Lecture 9 Classification of States There will be a lot of definitions and

More information

Stochastic modelling of epidemic spread

Stochastic modelling of epidemic spread Stochastic modelling of epidemic spread Julien Arino Centre for Research on Inner City Health St Michael s Hospital Toronto On leave from Department of Mathematics University of Manitoba Julien Arino@umanitoba.ca

More information

2 Problem Set 2 Graphical Analysis

2 Problem Set 2 Graphical Analysis 2 PROBLEM SET 2 GRAPHICAL ANALYSIS 2 Problem Set 2 Graphical Analysis 1. Use graphical analysis to describe all orbits of the functions below. Also draw their phase portraits. (a) F(x) = 2x There is only

More information

Solutions to Homework Assignment 2

Solutions to Homework Assignment 2 Solutions to Homework Assignment Real Analysis I February, 03 Notes: (a) Be aware that there maybe some typos in the solutions. If you find any, please let me know. (b) As is usual in proofs, most problems

More information

Stat 150 Practice Final Spring 2015

Stat 150 Practice Final Spring 2015 Stat 50 Practice Final Spring 205 Instructor: Allan Sly Name: SID: There are 8 questions. Attempt all questions and show your working - solutions without explanation will not receive full credit. Answer

More information

Solution: (Course X071570: Stochastic Processes)

Solution: (Course X071570: Stochastic Processes) Solution I (Course X071570: Stochastic Processes) October 24, 2013 Exercise 1.1: Find all functions f from the integers to the real numbers satisfying f(n) = 1 2 f(n + 1) + 1 f(n 1) 1. 2 A secial solution

More information

MATH 56A: STOCHASTIC PROCESSES CHAPTER 2

MATH 56A: STOCHASTIC PROCESSES CHAPTER 2 MATH 56A: STOCHASTIC PROCESSES CHAPTER 2 2. Countable Markov Chains I started Chapter 2 which talks about Markov chains with a countably infinite number of states. I did my favorite example which is on

More information

Markov Chains and Stochastic Sampling

Markov Chains and Stochastic Sampling Part I Markov Chains and Stochastic Sampling 1 Markov Chains and Random Walks on Graphs 1.1 Structure of Finite Markov Chains We shall only consider Markov chains with a finite, but usually very large,

More information

18.440: Lecture 33 Markov Chains

18.440: Lecture 33 Markov Chains 18.440: Lecture 33 Markov Chains Scott Sheffield MIT 1 Outline Markov chains Examples Ergodicity and stationarity 2 Outline Markov chains Examples Ergodicity and stationarity 3 Markov chains Consider a

More information

18.600: Lecture 32 Markov Chains

18.600: Lecture 32 Markov Chains 18.600: Lecture 32 Markov Chains Scott Sheffield MIT Outline Markov chains Examples Ergodicity and stationarity Outline Markov chains Examples Ergodicity and stationarity Markov chains Consider a sequence

More information

MARKOV CHAINS AND HIDDEN MARKOV MODELS

MARKOV CHAINS AND HIDDEN MARKOV MODELS MARKOV CHAINS AND HIDDEN MARKOV MODELS MERYL SEAH Abstract. This is an expository paper outlining the basics of Markov chains. We start the paper by explaining what a finite Markov chain is. Then we describe

More information

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1 MATH 56A: STOCHASTIC PROCESSES CHAPTER. Finite Markov chains For the sake of completeness of these notes I decided to write a summary of the basic concepts of finite Markov chains. The topics in this chapter

More information

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006.

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006. Markov Chains As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006 1 Introduction A (finite) Markov chain is a process with a finite number of states (or outcomes, or

More information

Randomized Load Balancing:The Power of 2 Choices

Randomized Load Balancing:The Power of 2 Choices Randomized Load Balancing: The Power of 2 Choices June 3, 2010 Balls and Bins Problem We have m balls that are thrown into n bins, the location of each ball chosen independently and uniformly at random

More information

P(X 0 = j 0,... X nk = j k )

P(X 0 = j 0,... X nk = j k ) Introduction to Probability Example Sheet 3 - Michaelmas 2006 Michael Tehranchi Problem. Let (X n ) n 0 be a homogeneous Markov chain on S with transition matrix P. Given a k N, let Z n = X kn. Prove that

More information

6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities

6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities 6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities 1 Outline Outline Dynamical systems. Linear and Non-linear. Convergence. Linear algebra and Lyapunov functions. Markov

More information

A Lower Bound for the Size of Syntactically Multilinear Arithmetic Circuits

A Lower Bound for the Size of Syntactically Multilinear Arithmetic Circuits A Lower Bound for the Size of Syntactically Multilinear Arithmetic Circuits Ran Raz Amir Shpilka Amir Yehudayoff Abstract We construct an explicit polynomial f(x 1,..., x n ), with coefficients in {0,

More information

6.842 Randomness and Computation March 3, Lecture 8

6.842 Randomness and Computation March 3, Lecture 8 6.84 Randomness and Computation March 3, 04 Lecture 8 Lecturer: Ronitt Rubinfeld Scribe: Daniel Grier Useful Linear Algebra Let v = (v, v,..., v n ) be a non-zero n-dimensional row vector and P an n n

More information

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains 8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains 8.1 Review 8.2 Statistical Equilibrium 8.3 Two-State Markov Chain 8.4 Existence of P ( ) 8.5 Classification of States

More information

The Inclusion Exclusion Principle

The Inclusion Exclusion Principle The Inclusion Exclusion Principle 1 / 29 Outline Basic Instances of The Inclusion Exclusion Principle The General Inclusion Exclusion Principle Counting Derangements Counting Functions Stirling Numbers

More information

Lecture 4: Two-point Sampling, Coupon Collector s problem

Lecture 4: Two-point Sampling, Coupon Collector s problem Randomized Algorithms Lecture 4: Two-point Sampling, Coupon Collector s problem Sotiris Nikoletseas Associate Professor CEID - ETY Course 2013-2014 Sotiris Nikoletseas, Associate Professor Randomized Algorithms

More information

Convergence of Random Walks and Conductance - Draft

Convergence of Random Walks and Conductance - Draft Graphs and Networks Lecture 0 Convergence of Random Walks and Conductance - Draft Daniel A. Spielman October, 03 0. Overview I present a bound on the rate of convergence of random walks in graphs that

More information

Boolean Inner-Product Spaces and Boolean Matrices

Boolean Inner-Product Spaces and Boolean Matrices Boolean Inner-Product Spaces and Boolean Matrices Stan Gudder Department of Mathematics, University of Denver, Denver CO 80208 Frédéric Latrémolière Department of Mathematics, University of Denver, Denver

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

Lecture 5: Random Walks and Markov Chain

Lecture 5: Random Walks and Markov Chain Spectral Graph Theory and Applications WS 20/202 Lecture 5: Random Walks and Markov Chain Lecturer: Thomas Sauerwald & He Sun Introduction to Markov Chains Definition 5.. A sequence of random variables

More information

Lecture Notes 6: Dynamic Equations Part A: First-Order Difference Equations in One Variable

Lecture Notes 6: Dynamic Equations Part A: First-Order Difference Equations in One Variable University of Warwick, EC9A0 Maths for Economists Peter J. Hammond 1 of 54 Lecture Notes 6: Dynamic Equations Part A: First-Order Difference Equations in One Variable Peter J. Hammond latest revision 2017

More information

Markov Chains Absorption (cont d) Hamid R. Rabiee

Markov Chains Absorption (cont d) Hamid R. Rabiee Markov Chains Absorption (cont d) Hamid R. Rabiee 1 Absorbing Markov Chain An absorbing state is one in which the probability that the process remains in that state once it enters the state is 1 (i.e.,

More information

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities

More information

P (A G) dp G P (A G)

P (A G) dp G P (A G) First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume

More information

2.1 Laplacian Variants

2.1 Laplacian Variants -3 MS&E 337: Spectral Graph heory and Algorithmic Applications Spring 2015 Lecturer: Prof. Amin Saberi Lecture 2-3: 4/7/2015 Scribe: Simon Anastasiadis and Nolan Skochdopole Disclaimer: hese notes have

More information

Countable and uncountable sets. Matrices.

Countable and uncountable sets. Matrices. Lecture 11 Countable and uncountable sets. Matrices. Instructor: Kangil Kim (CSE) E-mail: kikim01@konkuk.ac.kr Tel. : 02-450-3493 Room : New Milenium Bldg. 1103 Lab : New Engineering Bldg. 1202 Next topic:

More information

ASSIGNMENT 1 SOLUTIONS

ASSIGNMENT 1 SOLUTIONS MATH 271 ASSIGNMENT 1 SOLUTIONS 1. (a) Let S be the statement For all integers n, if n is even then 3n 11 is odd. Is S true? Give a proof or counterexample. (b) Write out the contrapositive of statement

More information

6 Markov Chain Monte Carlo (MCMC)

6 Markov Chain Monte Carlo (MCMC) 6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution

More information

The Degree of the Splitting Field of a Random Polynomial over a Finite Field

The Degree of the Splitting Field of a Random Polynomial over a Finite Field The Degree of the Splitting Field of a Random Polynomial over a Finite Field John D. Dixon and Daniel Panario School of Mathematics and Statistics Carleton University, Ottawa, Canada fjdixon,danielg@math.carleton.ca

More information

MAA704, Perron-Frobenius theory and Markov chains.

MAA704, Perron-Frobenius theory and Markov chains. November 19, 2013 Lecture overview Today we will look at: Permutation and graphs. Perron frobenius for non-negative. Stochastic, and their relation to theory. Hitting and hitting probabilities of chain.

More information

A Unified Theorem on SDP Rank Reduction. yyye

A Unified Theorem on SDP Rank Reduction.   yyye SDP Rank Reduction Yinyu Ye, EURO XXII 1 A Unified Theorem on SDP Rank Reduction Yinyu Ye Department of Management Science and Engineering and Institute of Computational and Mathematical Engineering Stanford

More information

MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS

MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS There will be eight problems on the final. The following are sample problems. Problem 1. Let F be the vector space of all real valued functions on

More information

Groups and Symmetries

Groups and Symmetries Groups and Symmetries Definition: Symmetry A symmetry of a shape is a rigid motion that takes vertices to vertices, edges to edges. Note: A rigid motion preserves angles and distances. Definition: Group

More information

Stochastic Processes

Stochastic Processes Stochastic Processes 8.445 MIT, fall 20 Mid Term Exam Solutions October 27, 20 Your Name: Alberto De Sole Exercise Max Grade Grade 5 5 2 5 5 3 5 5 4 5 5 5 5 5 6 5 5 Total 30 30 Problem :. True / False

More information

One-to-one functions and onto functions

One-to-one functions and onto functions MA 3362 Lecture 7 - One-to-one and Onto Wednesday, October 22, 2008. Objectives: Formalize definitions of one-to-one and onto One-to-one functions and onto functions At the level of set theory, there are

More information

Math Homework 5 Solutions

Math Homework 5 Solutions Math 45 - Homework 5 Solutions. Exercise.3., textbook. The stochastic matrix for the gambler problem has the following form, where the states are ordered as (,, 4, 6, 8, ): P = The corresponding diagram

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME)

WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME) WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME ITAI BENJAMINI, GADY KOZMA, LÁSZLÓ LOVÁSZ, DAN ROMIK, AND GÁBOR TARDOS Abstract. We observe returns of a simple random wal on a finite graph to a fixed node,

More information

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1 Properties of Markov Chains and Evaluation of Steady State Transition Matrix P ss V. Krishnan - 3/9/2 Property 1 Let X be a Markov Chain (MC) where X {X n : n, 1, }. The state space is E {i, j, k, }. The

More information

Determinants of Partition Matrices

Determinants of Partition Matrices journal of number theory 56, 283297 (1996) article no. 0018 Determinants of Partition Matrices Georg Martin Reinhart Wellesley College Communicated by A. Hildebrand Received February 14, 1994; revised

More information

CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions

CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions Instructor: Erik Sudderth Brown University Computer Science April 14, 215 Review: Discrete Markov Chains Some

More information

14.1 Finding frequent elements in stream

14.1 Finding frequent elements in stream Chapter 14 Streaming Data Model 14.1 Finding frequent elements in stream A very useful statistics for many applications is to keep track of elements that occur more frequently. It can come in many flavours

More information

Lecture 7. µ(x)f(x). When µ is a probability measure, we say µ is a stationary distribution.

Lecture 7. µ(x)f(x). When µ is a probability measure, we say µ is a stationary distribution. Lecture 7 1 Stationary measures of a Markov chain We now study the long time behavior of a Markov Chain: in particular, the existence and uniqueness of stationary measures, and the convergence of the distribution

More information

T. Liggett Mathematics 171 Final Exam June 8, 2011

T. Liggett Mathematics 171 Final Exam June 8, 2011 T. Liggett Mathematics 171 Final Exam June 8, 2011 1. The continuous time renewal chain X t has state space S = {0, 1, 2,...} and transition rates (i.e., Q matrix) given by q(n, n 1) = δ n and q(0, n)

More information

7.1 Coupling from the Past

7.1 Coupling from the Past Georgia Tech Fall 2006 Markov Chain Monte Carlo Methods Lecture 7: September 12, 2006 Coupling from the Past Eric Vigoda 7.1 Coupling from the Past 7.1.1 Introduction We saw in the last lecture how Markov

More information

Dynamical Systems 2, MA 761

Dynamical Systems 2, MA 761 Dynamical Systems 2, MA 761 Topological Dynamics This material is based upon work supported by the National Science Foundation under Grant No. 9970363 1 Periodic Points 1 The main objects studied in the

More information

ACO Comprehensive Exam October 14 and 15, 2013

ACO Comprehensive Exam October 14 and 15, 2013 1. Computability, Complexity and Algorithms (a) Let G be the complete graph on n vertices, and let c : V (G) V (G) [0, ) be a symmetric cost function. Consider the following closest point heuristic for

More information

The integers. Chapter 3

The integers. Chapter 3 Chapter 3 The integers Recall that an abelian group is a set A with a special element 0, and operation + such that x +0=x x + y = y + x x +y + z) =x + y)+z every element x has an inverse x + y =0 We also

More information

The collection of all subsets of cardinality k of a set with v elements (k < v) has the

The collection of all subsets of cardinality k of a set with v elements (k < v) has the Chapter 7 Block Designs One simile that solitary shines In the dry desert of a thousand lines. Epilogue to the Satires, ALEXANDER POPE The collection of all subsets of cardinality k of a set with v elements

More information

Two Simple Nonlinear Edge Detectors

Two Simple Nonlinear Edge Detectors Two Simple Nonlinear Edge Detectors Mladen Victor Wickerhauser Washington University in St. Louis, Missouri victor@math.wustl.edu http://www.math.wustl.edu/~victor International Conference on Wavelet Analysis

More information

Probability Background

Probability Background Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second

More information

What you learned in Math 28. Rosa C. Orellana

What you learned in Math 28. Rosa C. Orellana What you learned in Math 28 Rosa C. Orellana Chapter 1 - Basic Counting Techniques Sum Principle If we have a partition of a finite set S, then the size of S is the sum of the sizes of the blocks of the

More information

Symmetric Matrices and Eigendecomposition

Symmetric Matrices and Eigendecomposition Symmetric Matrices and Eigendecomposition Robert M. Freund January, 2014 c 2014 Massachusetts Institute of Technology. All rights reserved. 1 2 1 Symmetric Matrices and Convexity of Quadratic Functions

More information

Homework 6: Solutions Sid Banerjee Problem 1: (The Flajolet-Martin Counter) ORIE 4520: Stochastics at Scale Fall 2015

Homework 6: Solutions Sid Banerjee Problem 1: (The Flajolet-Martin Counter) ORIE 4520: Stochastics at Scale Fall 2015 Problem 1: (The Flajolet-Martin Counter) In class (and in the prelim!), we looked at an idealized algorithm for finding the number of distinct elements in a stream, where we sampled uniform random variables

More information

arxiv: v1 [math.co] 3 Nov 2014

arxiv: v1 [math.co] 3 Nov 2014 SPARSE MATRICES DESCRIBING ITERATIONS OF INTEGER-VALUED FUNCTIONS BERND C. KELLNER arxiv:1411.0590v1 [math.co] 3 Nov 014 Abstract. We consider iterations of integer-valued functions φ, which have no fixed

More information

h-vectors OF SMALL MATROID COMPLEXES

h-vectors OF SMALL MATROID COMPLEXES h-vectors OF SMALL MATROID COMPLEXES JESÚS A. DE LOERA, YVONNE KEMPER, STEVEN KLEE Abstract. Stanley conjectured in 1977 that the h-vector of a matroid simplicial complex is a pure O-sequence. We give

More information

Compositions of Random Functions on a Finite Set

Compositions of Random Functions on a Finite Set Compositios of Radom Fuctios o a Fiite Set Aviash Dalal MCS Departmet, Drexel Uiversity Philadelphia, Pa. 904 ADalal@drexel.edu Eric Schmutz Drexel Uiversity ad Swarthmore College Philadelphia, Pa., 904

More information

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC)

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC) Markov Chains (2) Outlines Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC) 2 pj ( n) denotes the pmf of the random variable p ( n) P( X j) j We will only be concerned with homogenous

More information

Lecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems

Lecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems University of Warwick, EC9A0 Maths for Economists Peter J. Hammond 1 of 45 Lecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems Peter J. Hammond latest revision 2017 September

More information

SPRING 2006 PRELIMINARY EXAMINATION SOLUTIONS

SPRING 2006 PRELIMINARY EXAMINATION SOLUTIONS SPRING 006 PRELIMINARY EXAMINATION SOLUTIONS 1A. Let G be the subgroup of the free abelian group Z 4 consisting of all integer vectors (x, y, z, w) such that x + 3y + 5z + 7w = 0. (a) Determine a linearly

More information

process on the hierarchical group

process on the hierarchical group Intertwining of Markov processes and the contact process on the hierarchical group April 27, 2010 Outline Intertwining of Markov processes Outline Intertwining of Markov processes First passage times of

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional

More information

The Degree of the Splitting Field of a Random Polynomial over a Finite Field

The Degree of the Splitting Field of a Random Polynomial over a Finite Field The Degree of the Splitting Field of a Random Polynomial over a Finite Field John D. Dixon and Daniel Panario School of Mathematics and Statistics Carleton University, Ottawa, Canada {jdixon,daniel}@math.carleton.ca

More information

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i := 2.7. Recurrence and transience Consider a Markov chain {X n : n N 0 } on state space E with transition matrix P. Definition 2.7.1. A state i E is called recurrent if P i [X n = i for infinitely many n]

More information

Matrix analytic methods. Lecture 1: Structured Markov chains and their stationary distribution

Matrix analytic methods. Lecture 1: Structured Markov chains and their stationary distribution 1/29 Matrix analytic methods Lecture 1: Structured Markov chains and their stationary distribution Sophie Hautphenne and David Stanford (with thanks to Guy Latouche, U. Brussels and Peter Taylor, U. Melbourne

More information

12 Sequences and Recurrences

12 Sequences and Recurrences 12 Sequences and Recurrences A sequence is just what you think it is. It is often given by a formula known as a recurrence equation. 12.1 Arithmetic and Geometric Progressions An arithmetic progression

More information

Economics 204 Fall 2011 Problem Set 1 Suggested Solutions

Economics 204 Fall 2011 Problem Set 1 Suggested Solutions Economics 204 Fall 2011 Problem Set 1 Suggested Solutions 1. Suppose k is a positive integer. Use induction to prove the following two statements. (a) For all n N 0, the inequality (k 2 + n)! k 2n holds.

More information

International Competition in Mathematics for Universtiy Students in Plovdiv, Bulgaria 1994

International Competition in Mathematics for Universtiy Students in Plovdiv, Bulgaria 1994 International Competition in Mathematics for Universtiy Students in Plovdiv, Bulgaria 1994 1 PROBLEMS AND SOLUTIONS First day July 29, 1994 Problem 1. 13 points a Let A be a n n, n 2, symmetric, invertible

More information

Countable and uncountable sets. Matrices.

Countable and uncountable sets. Matrices. CS 441 Discrete Mathematics for CS Lecture 11 Countable and uncountable sets. Matrices. Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Arithmetic series Definition: The sum of the terms of the

More information

ECE534, Spring 2018: Solutions for Problem Set #4 Due Friday April 6, 2018

ECE534, Spring 2018: Solutions for Problem Set #4 Due Friday April 6, 2018 ECE534, Spring 2018: s for Problem Set #4 Due Friday April 6, 2018 1. MMSE Estimation, Data Processing and Innovations The random variables X, Y, Z on a common probability space (Ω, F, P ) are said to

More information

Tactical Decompositions of Steiner Systems and Orbits of Projective Groups

Tactical Decompositions of Steiner Systems and Orbits of Projective Groups Journal of Algebraic Combinatorics 12 (2000), 123 130 c 2000 Kluwer Academic Publishers. Manufactured in The Netherlands. Tactical Decompositions of Steiner Systems and Orbits of Projective Groups KELDON

More information

1 Complex Networks - A Brief Overview

1 Complex Networks - A Brief Overview Power-law Degree Distributions 1 Complex Networks - A Brief Overview Complex networks occur in many social, technological and scientific settings. Examples of complex networks include World Wide Web, Internet,

More information

Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank

Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank David Glickenstein November 3, 4 Representing graphs as matrices It will sometimes be useful to represent graphs

More information

40th Canadian Mathematical Olympiad

40th Canadian Mathematical Olympiad 40th Canadian Mathematical Olympiad Wednesday, March 26, 2008 Solutions - CMO 2008 1. ABCD is a convex quadrilateral in which AB is the longest side. Points M and N are located on sides AB and BC respectively,

More information

Vertex colorings of graphs without short odd cycles

Vertex colorings of graphs without short odd cycles Vertex colorings of graphs without short odd cycles Andrzej Dudek and Reshma Ramadurai Department of Mathematical Sciences Carnegie Mellon University Pittsburgh, PA 1513, USA {adudek,rramadur}@andrew.cmu.edu

More information

Solved Examples. Given are two sets A {1, 2, -2, 3} and B = {1, 2, 3, 5}. Is the function f(x) = 2x - 1 defined from A to B?

Solved Examples. Given are two sets A {1, 2, -2, 3} and B = {1, 2, 3, 5}. Is the function f(x) = 2x - 1 defined from A to B? Solved Examples Example 1: Given are two sets A {1, 2, -2, 3} and B = {1, 2, 3, 5}. Is the function f(x) = 2x - 1 defined from A to B? Solution : Out of all the ordered pairs, the ordered pairs which are

More information

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009 LMI MODELLING 4. CONVEX LMI MODELLING Didier HENRION LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ Universidad de Valladolid, SP March 2009 Minors A minor of a matrix F is the determinant of a submatrix

More information

TMA 4265 Stochastic Processes Semester project, fall 2014 Student number and

TMA 4265 Stochastic Processes Semester project, fall 2014 Student number and TMA 4265 Stochastic Processes Semester project, fall 2014 Student number 730631 and 732038 Exercise 1 We shall study a discrete Markov chain (MC) {X n } n=0 with state space S = {0, 1, 2, 3, 4, 5, 6}.

More information

Discrete dynamics on the real line

Discrete dynamics on the real line Chapter 2 Discrete dynamics on the real line We consider the discrete time dynamical system x n+1 = f(x n ) for a continuous map f : R R. Definitions The forward orbit of x 0 is: O + (x 0 ) = {x 0, f(x

More information

Ma/CS 6b Class 23: Eigenvalues in Regular Graphs

Ma/CS 6b Class 23: Eigenvalues in Regular Graphs Ma/CS 6b Class 3: Eigenvalues in Regular Graphs By Adam Sheffer Recall: The Spectrum of a Graph Consider a graph G = V, E and let A be the adjacency matrix of G. The eigenvalues of G are the eigenvalues

More information

The SIS and SIR stochastic epidemic models revisited

The SIS and SIR stochastic epidemic models revisited The SIS and SIR stochastic epidemic models revisited Jesús Artalejo Faculty of Mathematics, University Complutense of Madrid Madrid, Spain jesus_artalejomat.ucm.es BCAM Talk, June 16, 2011 Talk Schedule

More information

5 Flows and cuts in digraphs

5 Flows and cuts in digraphs 5 Flows and cuts in digraphs Recall that a digraph or network is a pair G = (V, E) where V is a set and E is a multiset of ordered pairs of elements of V, which we refer to as arcs. Note that two vertices

More information

Component structure of the vacant set induced by a random walk on a random graph. Colin Cooper (KCL) Alan Frieze (CMU)

Component structure of the vacant set induced by a random walk on a random graph. Colin Cooper (KCL) Alan Frieze (CMU) Component structure of the vacant set induced by a random walk on a random graph Colin Cooper (KCL) Alan Frieze (CMU) Vacant set definition and results Introduction: random walks and cover time G n,p vacant

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours

More information

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks Recap Probability, stochastic processes, Markov chains ELEC-C7210 Modeling and analysis of communication networks 1 Recap: Probability theory important distributions Discrete distributions Geometric distribution

More information