Unpredictable Walks on the Integers

Similar documents
The Boundary Problem: Markov Chain Solution

Randomized Algorithms

6.842 Randomness and Computation March 3, Lecture 8

20.1 2SAT. CS125 Lecture 20 Fall 2016

Chapter 3: Polynomial and Rational Functions

Quantum Entanglement and Cryptography. Deepthi Gopal, Caltech

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Coupling. 2/3/2010 and 2/5/2010

Nearest-neighbor walks with low predictability profile and percolation in 2 + ε dimensions

Final Exam December 12, 2017

Unit 4 Systems of Equations Systems of Two Linear Equations in Two Variables

Unit 1: Polynomial Functions SuggestedTime:14 hours

MATH Mathematics for Agriculture II

Final Exam December 12, 2017

(x 1 +x 2 )(x 1 x 2 )+(x 2 +x 3 )(x 2 x 3 )+(x 3 +x 1 )(x 3 x 1 ).

2.3 Some Properties of Continuous Functions

Probability (continued)

Biased random walk on percolation clusters. Noam Berger, Nina Gantert and Yuval Peres

The Minesweeper game: Percolation and Complexity

Lecture 9 Classification of States

STAT/MA 416 Midterm Exam 3 Monday, November 19, Circle the section you are enrolled in:

Outline Goals and Assumptions Real Numbers Rational and Irrational. L11: Numbers. Alice E. Fischer

and problem sheet 1

Propp-Wilson Algorithm (and sampling the Ising model)

CS 6820 Fall 2014 Lectures, October 3-20, 2014

YALE UNIVERSITY DEPARTMENT OF COMPUTER SCIENCE

For a list of topics, look over the previous review sheets. Since the last quiz we have... Benford s Law. When does it appear? How do people use it?

Section 11.1: Sequences

NORTHWESTERN UNIVERSITY Thrusday, Oct 6th, 2011 ANSWERS FALL 2011 NU PUTNAM SELECTION TEST

USA Mathematical Talent Search Round 1 Solutions Year 22 Academic Year

EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015.

Math 016 Lessons Wimayra LUY

May 16, Aim: To review for Quadratic Function Exam #2 Homework: Study Review Materials. Warm Up - Solve using factoring: 5x 2 + 7x + 2 = 0

EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2017 Kannan Ramchandran March 21, 2017.

3.3 Real Zeros of Polynomial Functions

QUADRATIC FUNCTIONS. ( x 7)(5x 6) = 2. Exercises: 1 3x 5 Sum: 8. We ll expand it by using the distributive property; 9. Let s use the FOIL method;

Pigeonhole Principle and Ramsey Theory

General Least Squares Fitting

A FRAMEWORK FOR UNCONDITIONALLY SECURE PUBLIC-KEY ENCRYPTION (WITH POSSIBLE DECRYPTION ERRORS)

A New Interpretation of Information Rate

Inference for Stochastic Processes

MATH 151, FINAL EXAM Winter Quarter, 21 March, 2014

PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY

CPSC 467b: Cryptography and Computer Security

Clairvoyant scheduling of random walks

It only took one drop of strong base to dramatically raise the ph of solution A. The ph of the other solution hasn't

Theoretical Cryptography, Lecture 10

CALC 2 CONCEPT PACKET Complete

Introducing Proof 1. hsn.uk.net. Contents

Name: How Long Does It Take? (Part 1)

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

A General Overview of Parametric Estimation and Inference Techniques.

Continuous-time Markov Chains

Saturday Morning Math Group Austin Math Circle Austin Area Problem Solving Challenge 2009

Math 320-2: Midterm 2 Practice Solutions Northwestern University, Winter 2015

Polynomial Functions and Their Graphs. Definition of a Polynomial Function: numbers, with a n 0. The function defined by

Public Key Encryption

1. Let X and Y be independent exponential random variables with rate α. Find the densities of the random variables X 3, X Y, min(x, Y 3 )

MATH 2200 Final Review

Checkpoint Questions Due Monday, October 1 at 2:15 PM Remaining Questions Due Friday, October 5 at 2:15 PM

1. Introductory Examples

4.1 Real-valued functions of a real variable

Chapter 4E - Combinations of Functions

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10

Section 4.1: Sequences and Series

Equations and Inequalities

{ }. The dots mean they continue in that pattern to both

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science ALGORITHMS FOR INFERENCE Fall 2014

CS173 Strong Induction and Functions. Tandy Warnow

arxiv:math/ v1 [math.pr] 29 Nov 2002

Recursive Estimation

1 Basic Game Modelling

General Boundary Problem: System of Equations Solution

SO x is a cubed root of t

Chapter 1 Review of Equations and Inequalities

Intersecting Two Lines, Part Two

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 3

SHORT ANSWER. Write the word or phrase that best completes each statement or answers the question.

Public Key Cryptography. All secret key algorithms & hash algorithms do the same thing but public key algorithms look very different from each other.

Secondary Math 2H Unit 3 Notes: Factoring and Solving Quadratics

36-309/749 Math Review 2014

Sequential Decision Problems

1 Review of The Learning Setting

Interlude: Practice Final

THE N-VALUE GAME OVER Z AND R

Metric Spaces Lecture 17

Final. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes.

Asymmetric Communication Complexity and Data Structure Lower Bounds

A Brief Overview of the Clairvoyant Demon

Poisson Processes for Neuroscientists

. Find E(V ) and var(v ).

Notes for Lecture 11

Solving Linear and Rational Inequalities Algebraically. Definition 22.1 Two inequalities are equivalent if they have the same solution set.

1.) Suppose the graph of f(x) looks like this (each tick mark denotes 1 unit). x y

A Quick Algebra Review

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Error Correcting Codes: Combinatorics, Algorithms and Applications Spring Homework Due Monday March 23, 2009 in class

1 Normal Distribution.

Password Cracking: The Effect of Bias on the Average Guesswork of Hash Functions

Network Security Technology Spring, 2018 Tutorial 3, Week 4 (March 23) Due Date: March 30

Transcription:

Unpredictable Walks on the Integers EJ Infeld March 5, 213 Abstract In 1997 Benjamini, Premantle and Peres introduced a construction of a walk on Z that has a faster decaying predictability profile than that of the simple random walk [1], at the rate of ck α for some α > 1/2. A year later, Häggstrom and Mossel improved on this result by constructing two walks with predictability profiles that decay at close to the rate corresponding to the uniform probability distribution [2]. The same year, Hoffman showed in [3] that this new decay rate is in fact optimal. After introducing the necessary concepts, we will have a look at a few familiar examples and consider their predictability. Then we will follow a construction of a walk introduced in [2] as well as the proof that its predictability profile decays at the postulated rate. 1 Introduction We have seen many incarnations of a walk over the course of the term. Since we shall limit this discussion to walks on the integers and the nature of the problem is algorithmic, we will define it as follows: Definition A (random) walk on Z is a sequence of variables {S n } n N such that for n >, S n is determined by a (randomised) algorithm A that takes {S,..., S n 1 } (and a random element) as input and returns S n = S n 1 1 or S n = S n 1 + 1. Throughout this discussion we will take S =. As an example, consider a simple random walk on Z. At each step n, the input is a history {S,..., S n 1 }, and a uniform random variable U n [, 1], upon which A returns S n 1 + 1 if Department of Mathematics, Dartmouth College, Hanover NH. 1

U n < 1/2 and S n 1 1 otherwise. Note that in fact for any walk that is a Markov Chain the output of the algorithm A at step n is going to be independent of {S i } i<n 1. So what does it mean for a walk to be unpredictable? Suppose that Alice and Bob are on opposite sides of the fence - Alice wants to construct an algorithm that will generate an unpredictable walk, and Bob will make his best bet on where the walk will go. Bob is going to know Alice s algorithm and can watch the walk for as long as he wants, say n steps. Then he can, based on his knowledge of the algorithm and the observed history, make a bet on where the walk will be at step n + k, so k steps into the future. How good is his best bet is the measure of predictability of A. More formally, suppose we have an algorithm A that generates a walk on Z. Then the measure of predictability of A is the following: Definition [1] The predictability profile of A is: P RE A (k) = sup {P[S n+k = i : A, {S j } n j=]}, i Z, n N, {S j } n j= i.e. the highest probability over all possible values i Z, given A and any recorded history, of the walk having that particular value i k steps after the history is no longer recorded. Notice that if the algorithm A is deterministic, P RE A (k) = 1 for all k always. Also, P RE A (k) 1/(k + 1), since k + 1 is how many integers at most we can possibly reach on the kth step (remember that the path is bipartite). So the property that we are truly interested in is how fast P RE A (k) decays in k and in particular, whether its decay is sufficiently close to that of 1/(k + 1). It s not hard to guess that we would like the corresponding probability distribution to be close to uniform. This is very much a programmer s problem. Suppose Alice wants to send some information through a network of proxies, and Bob might watch its walk for a while before he decides to try to intercept it, and that he might take time equivalent to k steps to establish the interception attempt. The algorithm is set, the code is open and all Alice can do is make sure Bob, who knows as much about the walk as she does still has little chance of successful interception. This chance is the predictability profile of the walk. 2

2 Some walks and their predictability profiles 2.1 Simple random walk Consider a simple random walk. This is a Markov Chain, so the probability of being at any given point is going to be independent of history except for its final entry, and in fact the predictability profile itself will be independent of history entirely - the probability distribution is symmetric regardless of where on Z we start the walk, and so its maximum value will stay the same. So suppose without loss of generality that the history is {S = }. We are interested in P RE SRW (k) = sup{s k = m : S = }. m N Note that the probability of any individual sequence of moves left or right is the same, and equal to 2 k, so the question comes down to: how many walks terminate at m? The probability of terminating at m is that number times 2 k. Firstly, note that the path is bipartite. the only values of m Z with nonzero probability P[S k = m] are k m k of the same parity as k. For m = k + 2j, some natural number j k, there are ( k j) possible sequences of left or right moves that terminate at m. Then the highest probability is cumulated around, and in fact for k even we expect to be at after k steps with probability about 1/ k by the Central Limit Theorem. This is not all that close to 1/k. Can Alice do better? 2.2 Biased random walk with a random bias Suppose that at the beginning of the walk we pick p [, 1] uniformly at random and then perform a random walk in which at every stage i, we go right with probability p and left with probability 1 p. The probability that we will end up anywhere between k and k, provided the parity is correct, is now uniformly distributed. We can check this as follows. The probability that all k values σ i are equal to 1 is 1 p k dp = 1/(k + 1). Similarly, the probability that they are all equal to -1 is also 1/(k + 1). Now suppose that the probability that exactly j values σ i are -1 is ( ) k 1 p k j (1 p) j dp = 1/(k + 1). Then as long as j < k, by j integration by parts: 3

( ) 1 ( ) [ ] k k p p k j 1 (1 p) j+1 k j 1 ( ) 1 k = (1 p)j+1 + k j = + So what is wrong with this model? ( ) 1 k p k j (1 p) j dp = 1 j k + 1 k j pk j (1 p) j dp From the history S,..., S n the probability p can be approximated, by simply waiting long enough for {i n : σ i = 1} /n to be likely close to p. So the longer Bob watches the walk, the less optimal the probability profile becomes. And in fact if we assume Bob knows p, since he can watch the walk for as long as he wants, this model turns out to be at best as unpredictable as the simple random walk. This might tell us that an ideal walk would keep this random element in the bias, but be constructed in such a way as to hide this bias from the history. Intuition suggests that we then want the bias to be shifting slightly, enough to conceal itself from the history but not enough to become equivalent to the simple random walk. 3 Häggstrom and Mossel construction [2] The following construction attempts to balance the advantages of a random walk with a random bias with concealing the bias from the history. Let Alice pick an integer b > 1 and a sequence {a j } j=1 such that a j < 1/2. For any j, define a random process {p (j) i } i=1 by choosing a value uniformly at random from [ a j, a j ] and assigning it to b j consecutive terms, then choosing another value and assigning it to the next b j terms and so on. For example if b = 2, j = 1 then {p (j) i } i=1 = α 1, α 1, α 2, α 2,... where the α i s, i odd, are chosen independently, uniformly at random from [ a j, a j ]. Once we have determined {p (j) i } i=1 for each j, define p i = 1/2 + p (1) i + p (2) i +.... This is the bias for the walk at time i. Then the process is realized by a series of independent uniform j=1 4

random variables {U i } i=1, with the step of the walk at time i being σ i satisfying: { +1 if U i < p i σ i = 1 otherwise. So the walk is S n = σ i, where {σ i } i=1 are { 1, 1}-valued independent random variables i=1 conditioned on the environment {p i } i=1. At each time i, σ i takes value 1 with probability p i and -1 with probability 1 p i. We are dealing with a random walk that has varying bias. Theorem 3.1. Given b and {a j } j=1, there exists a constant c < such that P RE S (k) c ka logb (k/2) for all k. Proof. Assume without loss of generality that k 2b, since for any finite set of values of P RE S (k) we can choose an appropriate c, and denote m k = log b (k/2). Suppose we know S,..., S n and want to predict S n+k. Consider p (m k) i, constant on time intervals of length b m k = b log b (k/2) b log b(k/2) 1 = k/2b. So the time interval from n + 1 to n + k necessarily contains a subinterval I of length k/2b on which p (m k) i is constant. Write p (m k) i = p i p (m k) i = 1/2 + p (1) i + p (2) i + + p (m 1) i + p (m+1) i +..., and note that {p (m k) i } i=1 is independent of { p (m k) i } i=1. Define X i = U i p (m k) i and let: 1 if X i < a mk Y i = if X i [ a mk, a mk ] 1 if X i > a mk. Note that the relation between these variables and the original variables U i is as follows: X i < p (m k) i U i < p i, X i < a mk U i < p i, X i > a mk U i > p i, and so: Y i = 1 σ i = 1 5

Y i = 1 σ i = 1. Further, each of the Y i takes value with probability 2a mk, so the number of times this happens has a binomial (b m k, 2amk ) distribution. Given that Y i =, what is the probability that σ i = 1? We arrive at U i [ a mk + p (m k) i, a mk + p (m k) i ], less than p i with probability: a mk + p (m k) i 2a mk = 1 2 + p(mk) i. 2a mk Note that if Bob is given any extra information, his best odds can only improve, since he has the option to ignore that information. We will show that the bound we are attempting to prove can be established for a situation when Bob in fact knows more than we have previously assumed. Suppose that in addition to {σ i } n i=1, Bob knows: the values of {σ i I n+k \I} i.e. the movements outside of the subinterval of constant p(m k) i n+1 {Y i } i I. Then S n+k is necessarily in the interval L = S n + σ i + Y i 1 {Yi =}, S n + σ i + Y i + i I n+k n+1 \I i I i I i I n+k n+1 \I i I i I 1 {Yi =} And in fact, since the graph is bipartite, S n+k is necessarily an even number of steps from either end of this interval. We then note that on this subset of L of admissible elements, the probability distribution given the extra information is uniform, since the bias p (m k) I, and X i are independent, with uniform distribution on [ a mk, a mk ].. Then: b m k P[ {i Y : Y i = } = j] P[S n+k = x S,..., S n ] j= P [ {i I : Y i = } b m k a mk ]+ 1 b m k amk The term P [ {i I : Y i = } b m k amk ] decays exponentially in b m k amk, so it will not affect the overall rate of decay we are investigating. Therefore: and so for c = 2bC, have P RE S (k) P[S n+k = x S,..., S n ] c. ka logb (k/2) 6 C b m k amk = 2bC ka mk,

To sum up, how did we get around the fact that Bob might deduce partial information about the bias from the history? By giving him extra data, we have effectively cut off the part of Alice s algorithm that relied on information he might infer. We distilled the non-deterministic part of the walk to steps that draw their bias afresh after Bob is already reacting. It was shown in [3], that the decay rate achieved above is optimal. References [1] I. Benjamini, R. Premantle and Y. Peres, Unpredictable Paths and Percolation (1997) [2] O. Haggstrom and E. Mossel, Nearest-neighbor walks with low predictability profile and percolation in 2+ɛ dimensions (1998) [3] C. Hoffman, Unpredictable Nearest Neighbor Processes (1998) 7