Flip dynamics on canonical cut and project tilings

Similar documents
The coupling method - Simons Counting Complexity Bootcamp, 2016

25.1 Markov Chain Monte Carlo (MCMC)

Coupling AMS Short Course

Coupling. 2/3/2010 and 2/5/2010

Convergence Rate of Markov Chains

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

Convex Optimization CMU-10725

215 Problem 1. (a) Define the total variation distance µ ν tv for probability distributions µ, ν on a finite set S. Show that

The Markov Chain Monte Carlo Method

Introduction to Machine Learning CMU-10701

Stochastic optimization Markov Chain Monte Carlo

CONVERGENCE THEOREM FOR FINITE MARKOV CHAINS. Contents

Discrete quantum random walks

Markov Chains, Random Walks on Graphs, and the Laplacian

Characterization of cutoff for reversible Markov chains

Modern Discrete Probability Spectral Techniques

MARKOV CHAINS AND MIXING TIMES

6 Markov Chain Monte Carlo (MCMC)

The Theory behind PageRank

Some Definition and Example of Markov Chain

3. The Voter Model. David Aldous. June 20, 2012

7.1 Coupling from the Past

Lecture 3: September 10

Lecture 2: September 8

Discrete solid-on-solid models

On Markov Chain Monte Carlo

MAA704, Perron-Frobenius theory and Markov chains.

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

MARKOV CHAINS AND HIDDEN MARKOV MODELS

Edge Flip Chain for Unbiased Dyadic Tilings

Lecture 28: April 26

Propp-Wilson Algorithm (and sampling the Ising model)

INTRODUCTION TO MARKOV CHAIN MONTE CARLO

Lecture 5: Random Walks and Markov Chain

Markov Chains and Stochastic Sampling

DOMINO TILINGS INVARIANT GIBBS MEASURES

Markov chains. Randomness and Computation. Markov chains. Markov processes

1.3 Convergence of Regular Markov Chains

Lab 8: Measuring Graph Centrality - PageRank. Monday, November 5 CompSci 531, Fall 2018

TOTAL VARIATION CUTOFF IN BIRTH-AND-DEATH CHAINS

Markov and Gibbs Random Fields

Introduction to Markov Chains and Riffle Shuffling

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm

EECS 495: Randomized Algorithms Lecture 14 Random Walks. j p ij = 1. Pr[X t+1 = j X 0 = i 0,..., X t 1 = i t 1, X t = i] = Pr[X t+

Mixing times via super-fast coupling

MARKOV PROCESSES. Valerio Di Valerio

Advanced Sampling Algorithms

Lect4: Exact Sampling Techniques and MCMC Convergence Analysis

STOCHASTIC PROCESSES Basic notions

University of Chicago Autumn 2003 CS Markov Chain Monte Carlo Methods

Stochastic Processes

MIXING TIMES OF LOZENGE TILING AND CARD SHUFFLING MARKOV CHAINS 1. BY DAVID BRUCE WILSON Microsoft Research

Powerful tool for sampling from complicated distributions. Many use Markov chains to model events that arise in nature.

Quantum algorithms based on quantum walks. Jérémie Roland (ULB - QuIC) Quantum walks Grenoble, Novembre / 39

Probability & Computing

A note on adiabatic theorem for Markov chains and adiabatic quantum computation. Yevgeniy Kovchegov Oregon State University

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

An Introduction to Entropy and Subshifts of. Finite Type

INTRODUCTION TO MCMC AND PAGERANK. Eric Vigoda Georgia Tech. Lecture for CS 6505

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

Lecture 14: October 22

Introduction to Machine Learning CMU-10701

Math 456: Mathematical Modeling. Tuesday, March 6th, 2018

Markov Chains, Stochastic Processes, and Matrix Decompositions

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Can extra updates delay mixing?

Markov Processes Hamid R. Rabiee

Bayesian Methods with Monte Carlo Markov Chains II

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

Ergodic Properties of Markov Processes

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes

Markov Chains on Countable State Space

Lecture 7. µ(x)f(x). When µ is a probability measure, we say µ is a stationary distribution.

Random Times and Their Properties

Interlude: Practice Final

INTRODUCTION TO MCMC AND PAGERANK. Eric Vigoda Georgia Tech. Lecture for CS 6505

Random walks, Markov chains, and how to analyse them

1 Probabilities. 1.1 Basics 1 PROBABILITIES

Latent voter model on random regular graphs

Homework set 3 - Solutions

Necessary and sufficient conditions for strong R-positivity

Chapter 7. Markov chain background. 7.1 Finite state space

If we want to analyze experimental or simulated data we might encounter the following tasks:

Stat 516, Homework 1

Markov Chains and Spectral Clustering

Modelling data networks stochastic processes and Markov chains

Markov Chains (Part 3)

Homework 2 will be posted by tomorrow morning, due Friday, October 16 at 5 PM.

Markov Chains and MCMC

Monte Carlo Methods. Leon Gu CSD, CMU

Math 456: Mathematical Modeling. Tuesday, April 9th, 2018

Markov Chains. Andreas Klappenecker by Andreas Klappenecker. All rights reserved. Texas A&M University

Exponents. Reteach. Write each expression in exponential form (0.4)

LTCC. Exercises. (1) Two possible weather conditions on any day: {rainy, sunny} (2) Tomorrow s weather depends only on today s weather

The dimer model: universality and conformal invariance. Nathanaël Berestycki University of Cambridge. Colloque des sciences mathématiques du Québec

Reinforcement Learning

A note on adiabatic theorem for Markov chains

Mixing Times of Plane Random Rhombus Tilings


Ch5. Markov Chain Monte Carlo

Transcription:

Flip dynamics on canonical cut and project tilings Thomas Fernique CNRS & Univ. Paris 13 M2 Pavages ENS Lyon November 5, 2015

Outline 1 Random tilings 2 Random sampling 3 Mixing time 4 Slow cooling

Outline 1 Random tilings 2 Random sampling 3 Mixing time 4 Slow cooling

The problem with local rules Reminder: local rules can enforce (some) aperiodic structures. Proposition Any local rules which allow only aperiodic tilings also allow finite patterns (called deceptions) that appear in none of these tilings. The existence of a solution does not tell how to solve the puzzle!

Back to real quasicrystals Real quasicrystals are quenched from high T : Cooled ribbon Melt Heating Cooling wheel Stability is governed by the minimization of the free energy F : F = E TS.

Random tilings of maximal entropy Entropy of a tiling T of a region R: S = log(# rearrangements of T ) # tiles in T = log(# tilings of R) # tiles to tile R. Two main issues: 1 Which region R does maximize the entropy? 2 How does a typical tiling T of R look like? Underlying question: is there some random order?

The 2 1 case Both issues are solved for tilings of the line by two type of tiles: 1 For a and b tiles of each type, the tiling has entropy ( ) a+b a. This is maximal for a = b. 2 The fluctuations of a tiling by n tiles are in Θ( n). A typical tiling thus looks like a line...

The 3 2 case (dimer tilings) Much harder, but powerful results exist: Kasteleyn matrix (1967) Lindström Gessel-Viennot lemma (1989) Cohn-Kenyon-Propp variational principle (2001)

Beyond the 3 2 case Only simulations...

The problems with random tilings At least two problems with random tilings: 1 Is it really easier to solve the puzzle? 2 Quenching has been dropped in favour of slow cooling. Why? We shall here focus on the second problem. It is worth first asking: How the previous pictures have been drawn?

Outline 1 Random tilings 2 Random sampling 3 Mixing time 4 Slow cooling

Philosophy Goal: draw an object at random in a (typically large) set. The distribution is prescribed (e.g., uniform). This is easy if one can index all the elements, as the Rubik s cube: Scrambling draw a number between 1 and 8! 3 7 12! 2 10. This may be harder in other cases (Ising model, Tilings model... ) Moreover, one could prefer a sampling which is physically realist instead of algorithmically efficient. A common solution: Markov chain methods.

Markov chain A Markov chain (X t ) is a memoryless random process: P(X t+1 = x X 1 = x 1,..., X t = x t ) = P(X t+1 = x X t = x t ). Here: finite state space Ω. Description by a transition matrix P. Acts on the distributions over Ω. Stationary distribution: π = Pπ. A Markov chain is said to be irreducible if the state space is strongly connected; aperiodic if the gcd of the cycles through any state is 1; Ergodic if it is both irreducible and aperiodic.

Convergence Total variation between two distributions over a space Ω: µ ν := max A Ω µ(a) ν(a) = 1 2 µ(x) ν(x). x Ω This allows to measure the distance to stationarity: d(t) := max x Ω Pt (x, ) π. Theorem (Exponential convergence) For any ergodic Markov chain, there is α < 1 s.t. d(t) Cst α t.

Flips and tiling spaces Whenever a vertex of a n d tiling belongs to exactly d + 1 tiles, translating each of them by the vector shared by the d other ones yields a new tiling. This elementary operation is called a flip. The tiling space associated with a region R is the graph whose vertices are the tilings of R; whose edges connect tilings which differ by a flip. We want to sample by performing a random walk on this graph.

Ergodicity Orient the flips and define the random walk which at each step 1 pick uniformly at random a vertex x; 2 choose a flip direction (coin tossing); 3 try to perform the flip around x. This is an aperiodic Markov chain (self-loops). Is it irreducible?

Ergodicity Orient the flips and define the random walk which at each step 1 pick uniformly at random a vertex x; 2 choose a flip direction (coin tossing); 3 try to perform the flip around x. This is an aperiodic Markov chain (self-loops). Is it irreducible? Theorem (Kenyon, 1993) The n 2 tilings of a simply connected region are flip-connected. Theorem (Desoutter-Destainville, 2005) The n d tilings of a simply connected region are flip-connected for d n 2, but not always for n 3 d 3 (cycle obstruction).

Beyond ergodicity A path of flips is direct if no two flips involve exactly the same tiles. Theorem (Bodini-Fernique-Rao-Rémila, 2011) The n 2 tilings of a simply connected region are connected by direct paths of flips for n 4, but not always for n 5.

Beyond ergodicity A path of flips is direct if no two flips involve exactly the same tiles. Theorem (Bodini-Fernique-Rao-Rémila, 2011) The n 2 tilings of a simply connected region are connected by direct paths of flips for n 4, but not always for n 5.

Eternity is really long, especially near the end The convergence is exponential, hence fast... as in Chernobyl! It is often important to be more precise: How many moves to scramble your Rubik s cube? How many steps to shuffle a deck of cards? How many flips to have a typical tiling? The point is: how does the exponent depend on the space size?

Outline 1 Random tilings 2 Random sampling 3 Mixing time 4 Slow cooling

Definition Mixing time: τ mix := min{t d(t) 1/4}. Arbitrary threshold? Theorem (Half-life) A Markov chain is two times closer to stationarity after τ mix steps. In other words: d(t) 2 t/τ mix (exponential convergence again). Problem: bound the mixing time as a function of the space size.

Eigenvalues and the spectral gap Theorem (Perron-Frobenius, 1907-1912) The largest eigenvalue of a non-negative irreducible matrix is unique and simple. An ergodic Markov Chain has a non-negative irreductible matrix. Let λ 1,..., λ n be its eigenvalues, ordered by decreasing moduli. Since it is stochastic, λ 1 = 1, and by Perron-Frobenius, λ 2 < 1. Assume it diagonalizable (it can be adapted for Jordan forms). By decomposing a vector p on an eigenbasis ( p 1,..., p n ), one gets P t p p 1 = λ t k p k λ 2 t p k. k 2 k 2 This gives the exponent of the convergence... but what is λ 2?

Coupling Alternative idea: a random walker is lost when he forgot its starting point; two random walkers are lost when they meet. Think about a random walk on two cliques connected by one edge.

Coupling Alternative idea: a random walker is lost when he forgot its starting point; two random walkers are lost when they meet. Think about a random walk on two cliques connected by one edge. Coupling: two random variables with equal marginal distributions. Coupling time: τ couple := min{t X t = Y t } (random variable). Theorem The mixing time is less than the expectation of any coupling time. The random variables can be correlated: make a good choice!

Contraction A typical way to bound the coupling time: Proposition (Contraction) Let (X t, Y t ) be a coupling and ϕ : Ω Ω {0,..., D}. If there is β < 1 such that E(ϕ(X t+1, Y t+1 ) (X t, Y t ) = (x, y)) βϕ(x, y), then τ mix log(d) 1 β.

The 2 1 case Theorem (Wilson, 2004) Let h(v) := v 1 v 2 and define ϕ(w, w ) := n k=0 ( k h(w 1 w k ) h(w 1 w k [π ) cos n 1 )]. 2 Then, for x and y such that h(x 1 x k ) h(y 1 y k ) for any k, E(ϕ(X t+1, Y t+1 ) (X t, Y t ) = (x, y)) = (1 β)ϕ(x, y), with β = 1 cos(π/n) n 1 π2 2n 3.

The 2 1 case Theorem (Wilson, 2004) Let h(v) := v 1 v 2 and define ϕ(w, w ) := n k=0 ( k h(w 1 w k ) h(w 1 w k [π ) cos n 1 )]. 2 Then, for x and y such that h(x 1 x k ) h(y 1 y k ) for any k, E(ϕ(X t+1, Y t+1 ) (X t, Y t ) = (x, y)) = (1 β)ϕ(x, y), with β = 1 cos(π/n) n 1 π2 2n 3. With ϕ(w, w ) n, this yields τ mix 2 π 2 n 3 log(n). Actually tight.

The 3 2 case and beyond One can rely on the 2 1 case up to a modification of the chain:

The 3 2 case and beyond One can rely on the 2 1 case up to a modification of the chain: This yields τ mix = Θ(n 2 log(n)) for this modified chain. One can derive τ mix = O(n 4 ) for the original chain. Simulations however suggest τ mix = Θ(n 2 log(n).

The 3 2 case and beyond One can rely on the 2 1 case up to a modification of the chain: This yields τ mix = Θ(n 2 log(n)) for this modified chain. One can derive τ mix = O(n 4 ) for the original chain. Simulations however suggest τ mix = Θ(n 2 log(n). Simulations actually suggest this bound for any n 2 tiling...

Coupling from the past Assume that we simultaneously run the chain on all the states. Assume that after some steps, all the states have coalesced. Is the obtained state randomly drawn?

Coupling from the past Assume that we simultaneously run the chain on all the states. Assume that after some steps, all the states have coalesced. Is the obtained state randomly drawn? And if we run the chain from t = and stop at t = 0?

Coupling from the past Assume that we simultaneously run the chain on all the states. Assume that after some steps, all the states have coalesced. Is the obtained state randomly drawn? And if we run the chain from t = and stop at t = 0? One can actually simulate increasing tails of the whole evolution until all the states have coalesced (for example from t = 2 k ). This is the coupling from the past method (Propp-Wilson, 1996).

Coupling from the past Assume that we simultaneously run the chain on all the states. Assume that after some steps, all the states have coalesced. Is the obtained state randomly drawn? And if we run the chain from t = and stop at t = 0? One can actually simulate increasing tails of the whole evolution until all the states have coalesced (for example from t = 2 k ). This is the coupling from the past method (Propp-Wilson, 1996). Sometimes, the coalescence of some chains, eventually modified, ensures the global coalescence (sandwiching or bounding chains).

Outline 1 Random tilings 2 Random sampling 3 Mixing time 4 Slow cooling

Bridgeman-Stockbarger method Minimization of F = E TS: from max. entropy to min. energy.

Flip mechanism for atomic diffusion Maximal entropy S (high T ): modeled by random tilings. Minimal energy E (low T ): modeled by tilings with local rules, with the energy being the number of occuring forbidden patterns. We model the transformation by flips: correspond to an observed mechanism of atomic diffusion. do not modify the entropy of a tiling but can lower its energy.

Random flips Markov chain on the tilings of a given region: 1 choose uniformly at random a vertex; 2 choose a flip to be performed; 3 perform it, if possible, with probability min(1, exp( E/T )). Ergodicity is ensured at T > 0 for n 1 and n 2 tilings. At fixed T, Boltzmann/Gibbs stationary distribution: π(x) = 1 exp( E(x)/T ). Z(T ) At T = : uniform distribution (random sampling). At T = 0: Dirac distribution (error-correcting chain).

The 2 1 case at T = 0 Tiling space: words with as many a as b. Energy: number of pairs of neighboor equal letters. At T = 0, the flips abab aabb and baba bbaa are forbidden. Any word is eventually corrected. How fast?

The 2 1 case at T = 0 Tiling space: words with as many a as b. Energy: number of pairs of neighboor equal letters. At T = 0, the flips abab aabb and baba bbaa are forbidden. Any word is eventually corrected. How fast? For ϕ(w) = v DF(w) v one shows E( ϕ(w) w)) 1 4n n. Theorem (Bodini-Fernique-Regnault, 2010) The coupling time, hence the mixing time, is O(n 3 ). It is conjectured to be Θ(n 3 ). At least, it is Ω(n 2 ) (diameter).

The 3 2 case at T = 0 Tiling space: patches with the boundary of a 6-fold planar tiling. Energy: number of pairs of neighboor equal tiles. Any tiling is eventually corrected. How fast? Theorem (Fernique-Regnault, 2010) The coupling time, hence the mixing time, is O(n 2 n) (for n tiles). It is conjectured to be Θ(n 2 ). At least, it is Ω(n n) (diameter).

Beyond the 3 2 case at T = 0 Tiling space: patches with the boundary of, e.g., a Penrose tiling. Energy: number of forbidden patterns, e.g., violated alternations.

Beyond the 3 2 case at T = 0 Tiling space: patches with the boundary of, e.g., a Penrose tiling. Energy: number of forbidden patterns, e.g., violated alternations.

Beyond the 3 2 case at T = 0 Tiling space: patches with the boundary of, e.g., a Penrose tiling. Energy: number of forbidden patterns, e.g., violated alternations.

Beyond the 3 2 case at T = 0 Tiling space: patches with the boundary of, e.g., a Penrose tiling. Energy: number of forbidden patterns, e.g., violated alternations.

Beyond the 3 2 case at T = 0 Tiling space: patches with the boundary of, e.g., a Penrose tiling. Energy: number of forbidden patterns, e.g., violated alternations. Simulations suggest a mixing time Θ(n 2 )...... but it is still not proven that tilings are eventually corrected!

Cooling schedule We considered Markov chains at two temperatures: T = : random sampling; T = 0: error-correcting chain. The chain at T = 0 aims to model the quasicrystallization but nothing really happens at T = 0 (frozen); flips allowed at T > 0 can fasten the correction (annealing). Optimal cooling schedule?

Some open questions Do Penrose tilings have maximal entropy? Arctic circle phenomena beyond dimer tilings? Error-correcting planarization of n d tilings? Mixing time of n d tilings at T = 0? at any T? Phase transition (with T ) in the mixing time? Optimal cooling schedule?