Mixing in Product Spaces. Elchanan Mossel

Size: px
Start display at page:

Download "Mixing in Product Spaces. Elchanan Mossel"

Transcription

1

2 Poincaré Recurrence Theorem Theorem (Poincaré, 1890) Let f : X X be a measure preserving transformation. Let E X measurable. Then P[x E : f n (x) / E, n > N(x)] = 0

3 Poincaré Recurrence Theorem Theorem (Poincaré, 1890) Let f : X X be a measure preserving transformation. Let E X measurable. Then P[x E : f n (x) / E, n > N(x)] = 0 One of the first results in Ergodic Theory. Long term mixing. This talk is about short term mixing.

4 Finite Markov Chains As a first example consider a Finite Markov chain. Let M be a k k doubly stochastic symmetric matrix. Pick X 0 uniformly at random from 1,..., k. Given X i = a, let X i+1 = b with probability M a,b.

5 Finite Markov Chains As a first example consider a Finite Markov chain. Let M be a k k doubly stochastic symmetric matrix. Pick X 0 uniformly at random from 1,..., k. Given X i = a, let X i+1 = b with probability M a,b. Theorem (Long Term Mixing for Markov Chains) Suppose that other than 1, all eigenvalues λ i of M satisfy λ i λ < 1. Then for any two sets A, B [k], it holds that P[X 0 A, X t B] P[A]P[B] λ t

6 Short Term Mixing for Markov Chains Theorem P[X 0 A, X 1 B] P[A]P[B] is upper bounded by λ P[A](1 P[A])P[B](1 P[B]) Shows: mixing in one step for large sets.

7 Short Term Mixing for Markov Chains Theorem P[X 0 A, X 1 B] P[A]P[B] is upper bounded by λ P[A](1 P[A])P[B](1 P[B]) Shows: mixing in one step for large sets. Proof: 1 A = P[A]1 + f, 1 B = P[B]1 + g, where f, g 1

8 Short Term Mixing for Markov Chains Theorem P[X 0 A, X 1 B] P[A]P[B] is upper bounded by λ P[A](1 P[A])P[B](1 P[B]) Shows: mixing in one step for large sets. Proof: 1 A = P[A]1 + f, 1 B = P[B]1 + g, where f, g 1 P[X 0 A, X 1 B] = 1 k (P[A]1 + f )t M(P[B]1 + g) = P[A]P[B] + 1 k f t Mg, 1 k f t Mg λ f 2 g 2 = λ P[A](1 P[A])P[B](1 P[B])

9 Short Term Mixing for Markov Chains Theorem P[X 0 A, X 1 B] P[A]P[B] is upper bounded by λ P[A](1 P[A])P[B](1 P[B]) Shows: mixing in one step for large sets. Proof: 1 A = P[A]1 + f, 1 B = P[B]1 + g, where f, g 1 P[X 0 A, X 1 B] = 1 k (P[A]1 + f )t M(P[B]1 + g) = P[A]P[B] + 1 k f t Mg, 1 k f t Mg λ f 2 g 2 = λ P[A](1 P[A])P[B](1 P[B]) Also called Expander Mixing Lemma. Used a lot in computer science, e.g. in (de)randomization.

10 The tensor property Consider (Y 1, Z 1 ),..., (Y n, Z n ) which are drawn independently from the distribution of (X 0, X 1 ). Equivalently, the transition matrix from Y = (Y 1,..., Y n ) to Z = (Z 1,..., Z n ) is M n.

11 The tensor property Consider (Y 1, Z 1 ),..., (Y n, Z n ) which are drawn independently from the distribution of (X 0, X 1 ). Equivalently, the transition matrix from Y = (Y 1,..., Y n ) to Z = (Z 1,..., Z n ) is M n. Thm = that for any sets A, B [k] n : P[Y A, Z B] P[A]P[B] λ P[A](1 P[A])P[B](1 P[B])

12 The tensor property Consider (Y 1, Z 1 ),..., (Y n, Z n ) which are drawn independently from the distribution of (X 0, X 1 ). Equivalently, the transition matrix from Y = (Y 1,..., Y n ) to Z = (Z 1,..., Z n ) is M n. Thm = that for any sets A, B [k] n : P[Y A, Z B] P[A]P[B] λ P[A](1 P[A])P[B](1 P[B]) Follows immediately from tensorization of the spectrum.

13 Log Sobolev inequalities Entropy, Log Sobolev and hyper-contraction A similar story could be told using more sophisticated analytic tools. Easier to work with Markov semi-groups T t = e tl.

14 Log Sobolev inequalities Entropy, Log Sobolev and hyper-contraction A similar story could be told using more sophisticated analytic tools. Easier to work with Markov semi-groups T t = e tl. Entropy, Dirchelet Form Ent(f ) = E(f log f ) Ef log Ef E(f, g) = E(fLg) = E(gLf ) = E(g, f ) = d dt EfT tg. t=0

15 Log Sobolev inequalities Entropy, Log Sobolev and hyper-contraction A similar story could be told using more sophisticated analytic tools. Easier to work with Markov semi-groups T t = e tl. Entropy, Dirchelet Form Ent(f ) = E(f log f ) Ef log Ef E(f, g) = E(fLg) = E(gLf ) = E(g, f ) = d dt EfT tg. t=0 Definition of Log-Sob p-logsob(c) f, Ent(f p ) Cp2 4(p 1) E(f p 1, f ) (p 0, 1) 1-logSob(C) f, Ent(f ) C 4 E(f, log f ) 0-logSob(C) f, Var(log f ) C 2 E(f, 1/f )

16 Log Sob. Inequalities and Hyper-Contraction Hyper-Contraction (Gross, Nelson ) r-logsob with constant C implies T t f p f q, t C 4 log p 1 q 1, 1 < p < q < r or r < q < p = E[g(X 0 )f (X t )] = E[gT t f g p Tf p g p f q If f = 1 A and g = 1 B, get: P[X 0 A, X t B] 1 A q 1 B p = P[A] 1/q P[B] 1/p, Now optimize over norms to get a better bound than CS.

17 Reverse-Hyper-Contraction Log-Sobolev and Rev. Hyper-Contraction(M-Oleszkiewicz-Sen-13) Let T t = e tl be a general Markov semi-group satisfying 2-Logsob with constant C or 1-Logsob inequality with constant C. Then for all q < p < 1, all positive f, g and all t C 4 holds that T t f q f p = E[g(X 0 )f (X t )] = E[gT t f ] g q f p log 1 q 1 p it

18 Short-Time Implications Theorem (M-Oleszkiewicz-Sen-13 ; Short-Time Implications) Let T t = e tl, where L satisfy 1 or 2-LogSob inequality with constant C. Let A, B Ω n with P[A] ɛ and P[B] ɛ. Then: P[X (0) A, X (t) B] ɛ 2 1 e 2t/C

19 Short-Time Implications Theorem (M-Oleszkiewicz-Sen-13 ; Short-Time Implications) Let T t = e tl, where L satisfy 1 or 2-LogSob inequality with constant C. Let A, B Ω n with P[A] ɛ and P[B] ɛ. Then: P[X (0) A, X (t) B] ɛ 2 1 e 2t/C Comments 1. Works for small sets too. 2. Tensorizes. 3. Some examples where it is (almost) tight. 4. Uses in social choice analysis, queuing theory.

20 Comment: typical application MCMC Long Time Behavior Log Sobolev inequalities play a major role in analyzing long term mixing of Markov chains, in particular in analysis of mixing times (Diaconis, Saloff-Coste etc.) Long Time Behavior The ε-total variation mixing time of a finite Markov chain is bounded by: 1 λ (log(1/π ) + log(1/ɛ)) 1 C (log log(1/π ) + log(1/ɛ)) for a continuous time Markov chain with spectral gap λ and 2-LogSob C.

21 Comment: typical application MCMC Long Time Behavior Log Sobolev inequalities play a major role in analyzing long term mixing of Markov chains, in particular in analysis of mixing times (Diaconis, Saloff-Coste etc.) Long Time Behavior The ε-total variation mixing time of a finite Markov chain is bounded by: 1 λ (log(1/π ) + log(1/ɛ)) 1 C (log log(1/π ) + log(1/ɛ)) for a continuous time Markov chain with spectral gap λ and 2-LogSob C.

22 What are these lectures about? High Dimensional Phenomena High dimensional mixing: mixing of product processes on product spaces Ω n with n large. Tight bounds For which processes, given measures a and b can we find precise upper/lower bounds for sup ( P[X 0 A, X t B] : P[A] = a, P[B] = b ) Interested in product space/processes of dimension n and answers as n. Most important examples / techniques from probability / analysis.

23 What are these lectures about? Mulit-step prcoesses How to bound P[X 0 A 0, X 1 A 1,..., X k A k ] for processes X 0,..., X k? Interested in product space/processes of dimension n and answers as n. Most important examples / techniques from additive combinatorics.

24 What are these lectures about? And more Theory that does both? Applications?

25 Today: tight bounds Borell s result. Open Problem: The Boolean cube. The state of affairs - partition into 3 parts or more.

26 Two Examples: Gaussian, Boolean Correlated pairs (M-O Donnell-Regev-Steif-Sudakov-05): Let x, y { 1, 1} n be e t correlated: x is chosen uniformly and y is T t correlated version. i.e. E[x i y i ] = e t for all i independently Let A, B { 1, 1} n 1/2 with P[A] ɛ and P[B] ɛ Then: P[x A, y B] ɛ 2 1 e t Easy to prove when A = B...

27 Two Examples: Gaussian, Boolean Correlated pairs (M-O Donnell-Regev-Steif-Sudakov-05): Let x, y { 1, 1} n be e t correlated: x is chosen uniformly and y is T t correlated version. i.e. E[x i y i ] = e t for all i independently Let A, B { 1, 1} n 1/2 with P[A] ɛ and P[B] ɛ Then: P[x A, y B] ɛ 2 1 e t Easy to prove when A = B... Gaussian Version Let x, y R n two Gaussian vectors: x N(0, 1), y N(0, 1), E[x i y j ] = e t δ i,j Let A, B R n with P[A] ɛ and P[B] ɛ Then: P[x A, y B] ɛ 2 1 e t

28 Borell s Result and Open Problems Borell (85): In Gaussian case the maximum and minimum of P[x A, y B] as a function of P[A] and P[B] is obtained for parallel half-spaces.

29 Borell s Result and Open Problems Borell (85): In Gaussian case the maximum and minimum of P[x A, y B] as a function of P[A] and P[B] is obtained for parallel half-spaces. Do not know what is the optimum in { 1, 1} n. In particular:

30 Borell s Result and Open Problems Borell (85): In Gaussian case the maximum and minimum of P[x A, y B] as a function of P[A] and P[B] is obtained for parallel half-spaces. Do not know what is the optimum in { 1, 1} n. In particular: Open Problem: lim min(p[x A, Y B] : A, B { 1, n 1}n, P[A] = P[B] = 1/4) and similarly for max.

31 Borell s Result and Open Problems Borell (85): In Gaussian case the maximum and minimum of P[x A, y B] as a function of P[A] and P[B] is obtained for parallel half-spaces. Do not know what is the optimum in { 1, 1} n. In particular: Open Problem: lim min(p[x A, Y B] : A, B { 1, n 1}n, P[A] = P[B] = 1/4) and similarly for max. Partition to 3 or more parts even in Gaussian space.

32 If there is time before the break... A cute proof of a special case of Borell s result. Connections to social choice Theory.

33 Simple Example 1 Cosmic coin problem(m-o Donnell-05): x { 1, 1} n uniform. (y i ) m 1 conditionally independent given x. Each pair (x, y i ) is ρ-correlated. Problem: What is the largest P[y 1 A,... y m A] can be?

34 Simple Example 2 (y i,j ) 1 i<j m is an exchangeable collection of vectors in { 1, 1} n. If I J = 1 then y I, y J are 1/3 correlated. Otherwise independent. Why? If n voters rank alternatives uniformly at random, the pairwise preferences between alternatives will be given by the collection y.

35 Full support finite Ω using hyper-contraction Thm: More General Reverse Hypercontractivity Theorem (M-Oleszkiewicz-Sen-13) Let a the measure Ψ over a finite Ω k satisfy min x1,...,x k Ω Pr[X 1 = x,..., X k = x k ] = α > 0 and have equal marginals.

36 Full support finite Ω using hyper-contraction Thm: More General Reverse Hypercontractivity Theorem (M-Oleszkiewicz-Sen-13) Let a the measure Ψ over a finite Ω k satisfy min x1,...,x k Ω Pr[X 1 = x,..., X k = x k ] = α > 0 and have equal marginals. Consider the distribution Ψ n and let A 1,..., A k Ω n, µ(a i ) µ. Then: Pr[X 1 A 1,... X k A k ] µ O( 1 α), where (X 1 (i),..., X k (i)) are i.i.d. according to Ψ. Note This is a key tool of analyzing the examples above as well as many others.

37 Notation Distributed according to P := P n. Tuples X i are i.i.d. according to P. The marginals of P are π j. Vectors X (j) are distributed according to π j := π n j. X X 1 X 2... X i... X n X (1) X (1) 1 X (1) 2 X (2) X (2) 1 X (2) 2. X (j)... X (j) 1 X (j) 2.. X (1) i X (2) i. X (j) i. X (l) X (l) X (l) 1 X (l) 2 i X (1) n X (2) n. X n (j). X n (l)

38 Lower Bounds We are mostly interested in two types of lower bounds: Set hitting: Lower bounds on in terms of P[A 1 ],..., P[A k ] P[X 1 A 1,..., X k A k ] Same set hitting: Lower bounds on in terms of P[A]. P[X 1 A,..., X k A] Set hitting will require something... - e.g. X 1 = X 2 =... = X k.

39 Gaussian Bounds Borell (85) k = 2 - parallel half-spaces are optimal (also Isaksson-Mossel, Neeman) By a Reverse Brascamp-Lieb inq. (Ledoux, Chen-Dafnis-Paouris 14-15) for A,... C R n : P[U A,..., Z C] (P[A] P[C]) 1/(1 ρ2), where ρ is the second eigenvalue of Σ. Doesn t require independence of coordinates

40 Full Support Case Thm: More General Reverse Hypercontractivity Theorem (M-Oleszkiewicz-Sen-13) Let a the measure Ψ over a finite Ω k satisfy min x1,...,x k Ω Pr[X 1 = x,..., X k = x k ] = α > 0 and have equal marginals.

41 Full Support Case Thm: More General Reverse Hypercontractivity Theorem (M-Oleszkiewicz-Sen-13) Let a the measure Ψ over a finite Ω k satisfy min x1,...,x k Ω Pr[X 1 = x,..., X k = x k ] = α > 0 and have equal marginals. Then: Pr[X 1 A 1,... X k A k ] µ O( 1 α), where (X 1 (i),..., X k (i)) are i.i.d. according to Ψ.

42 Non full support? What if the support of Ω is not full? Do we care?

43 Non full support? What if the support of Ω is not full? Do we care? Maybe: This is what additive combinatorics is all about. In particular: finite cominatorics in finite field models (Green ). Many other applications in combinatorics and computer science.

44 Additive combinatorics perspective Example: Theorem (Finite Field Roth Theorem) Y, R be chosen uniformly at random at F n 3. Then for every µ > 0 there exists c(µ) > 0, N(µ) such that if n N(µ) and A F n 3 satisfies P[A] µ, then: P[Y A, Y + R A, Y + 2R A] c(µ).

45 Additive combinatorics perspective Example: Theorem (Finite Field Roth Theorem) Y, R be chosen uniformly at random at F n 3. Then for every µ > 0 there exists c(µ) > 0, N(µ) such that if n N(µ) and A F n 3 Why is this true? satisfies P[A] µ, then: P[Y A, Y + R A, Y + 2R A] c(µ).

46 Fourier Obstructions Theorem (Finite Field Roth Theorem - Analysis) Let Y, R be chosen uniformly at random at F n 3. Let A, B, C F n 3 then P[Y A, Y + R B, Y + 2R C] P[A]P[B]P[C] Â

47 Fourier Obstructions Theorem (Finite Field Roth Theorem - Analysis) Let Y, R be chosen uniformly at random at F n 3. Let A, B, C F n 3 then P[Y A, Y + R B, Y + 2R C] P[A]P[B]P[C] Â Only obstruction to uniformity is linear structure

48 Fourier Obstructions Theorem (Finite Field Roth Theorem - Analysis) Let Y, R be chosen uniformly at random at F n 3. Let A, B, C F n 3 then P[Y A, Y + R B, Y + 2R C] P[A]P[B]P[C] Â Only obstruction to uniformity is linear structure If A = B = C, high Fourier coefficient = can restrict to linear subspace with higher denisty

49 Fourier Obstructions Theorem (Finite Field Roth Theorem - Analysis) Let Y, R be chosen uniformly at random at F n 3. Let A, B, C F n 3 then P[Y A, Y + R B, Y + 2R C] P[A]P[B]P[C] Â Only obstruction to uniformity is linear structure If A = B = C, high Fourier coefficient = can restrict to linear subspace with higher denisty Density increase arguments...

50 Higher Order Arithmetic Obstructions Furstenberg-Weiss (80s): For longer arithmetic progressions, obstructions other than Fourier. Gowers: Obstructions can be identified using the Gowers norms. Again - use obstruction to your benefit. Thm: (Gowers 08; Rodel and Skokan 04,06): If q is prime and l q then for every µ > 0 there exists c(µ) > 0, N(µ) such that if n N(µ) and A Fq n satisfies P[A] µ, then: P[Y A, Y + R A,..., Y + (l 1)R A] c(µ), where A, R F n q are chosen uniformly at random.

51 Higher Order Arithmetic Obstructions Furstenberg-Weiss (80s): For longer arithmetic progressions, obstructions other than Fourier. Gowers: Obstructions can be identified using the Gowers norms. Again - use obstruction to your benefit. Thm: (Gowers 08; Rodel and Skokan 04,06): If q is prime and l q then for every µ > 0 there exists c(µ) > 0, N(µ) such that if n N(µ) and A Fq n satisfies P[A] µ, then: P[Y A, Y + R A,..., Y + (l 1)R A] c(µ), where A, R F n q are chosen uniformly at random. Question: Is the additive structure necessary?

52 Obstruction to Chaos Consider the support of Ω as a graph G with vertex V = all atoms with non-zero weight and edges between any two atoms that differ in one coordinate. We say that ρ < 1 if the graph G is connected. More formally: Definition { ρ(p, S, T ) := sup Cov[f (X (S) ), g(x (T ) )] f : Ω (S) R, g : Ω (T ) R, } Var[f (X (S) )] = Var[g(X (T ) )] = 1. The correlation of P is ρ(p) := max j [l] ρ (P, {j}, [l] \ {j}).

53 The quest for a unifying theory Is there one theory that explains both the noisy examples and the additive theory?

54 Example Let X be uniform in F n 3. Let Y i = X i or X i + 1 with probability 1/2 independently for each coordinate. Theorem = P[X A, Y a] c(p[a]). Motivation from understanding parallel repetition. Does not follow from hyper-contraction nor does it follow from additive techniques...

55 A General Result Theorem (+ Hazla, Holenstein) Suppose (X, Y ) is distributed in a finite Ω 2 such that: α = min a P[X = Y = a] > 0. P[X = a] = P[Y = a] for all a. Then for any set A Ω n with P X n[a] = P Y n[a] µ it holds that P[X A, Y A] c(α, µ) > 0 Our c is pretty bad: c = 1/ exp(exp(exp(1/(µ) D ))), D = D(α)

56 A General Result Theorem (+ Hazla, Holenstein) Suppose (X, Y ) is distributed in a finite Ω 2 such that: α = min a P[X = Y = a] > 0. P[X = a] = P[Y = a] for all a. Then for any set A Ω n with P X n[a] = P Y n[a] µ it holds that P[X A, Y A] c(α, µ) > 0 Our c is pretty bad: c = 1/ exp(exp(exp(1/(µ) D ))), D = D(α) Related to the fact that the proof is interesting: 1 Lose in Regularity Lemma type arguments. 2 Lose in Invariance transforming the problem to a Gaussian problem.

57 A Markov Chain Theorem and a general process theorem Theorem[+Hazla, Holenstein] X i, Y i, Z i,..., W i be a Markov chain over Ω with min x Ω Pr[X i = Y i = Z i =... W i = x] = β > 0 and uniform marginals. Let A Ω n, µ(a) = µ > 0. Pr[X A Y A Z A,..., W A] f (µ, β) > 0. Theorem[+Hazla, Holenstein] X i, Y i, Z i,..., W i be distributed over Ω k with min x Ω Pr[X i = Y i = Z i =... W i = x] = β > 0 and uniform marginals. Suppose further that ρ(x i, Y i,..., W i ) < 1. Let A Ω n, µ(a) = µ > 0. Pr[X A Y A Z A,..., W A] f (µ, β) > 0.

58 The condition ρ < 1 Weaker than full support.

59 The condition ρ < 1 Weaker than full support. Does not hold in arithmetic setups.

60 The condition ρ < 1 Weaker than full support. Does not hold in arithmetic setups. ρ = 1 iff the support of Ψ is connected with respect to changing one coordinate at a time.

61 The condition ρ < 1 Weaker than full support. Does not hold in arithmetic setups. ρ = 1 iff the support of Ψ is connected with respect to changing one coordinate at a time. Example: (x, y) F3 2 where y = x, x + 1 has ρ < 1 but not full support.

62 Open Problems Still searching for unified theory. Concrete Example: Suppose Ψ is uniform over {(0, 0, 0), (1, 1, 1), (2, 2, 2), (0, 1, 2), (1, 2, 0), (2, 0, 1)} ρ = 1 but not arithmetic. Do not understand.

63 Questions??

64 Questions?? Thank you!

Reverse Hyper Contractive Inequalities: What? Why?? When??? How????

Reverse Hyper Contractive Inequalities: What? Why?? When??? How???? Reverse Hyper Contractive Inequalities: What? Why?? When??? How???? UC Berkeley Warsaw Cambridge May 21, 2012 Hyper-Contractive Inequalities The noise (Bonami-Beckner) semi-group on { 1, 1} n 1 2 Def 1:

More information

Heat Flows, Geometric and Functional Inequalities

Heat Flows, Geometric and Functional Inequalities Heat Flows, Geometric and Functional Inequalities M. Ledoux Institut de Mathématiques de Toulouse, France heat flow and semigroup interpolations Duhamel formula (19th century) pde, probability, dynamics

More information

Probability Models of Information Exchange on Networks Lecture 3

Probability Models of Information Exchange on Networks Lecture 3 Probability Models of Information Exchange on Networks Lecture 3 Elchanan Mossel UC Berkeley All Right Reserved The Bayesian View of the Jury Theorem Recall: we assume 0/1 with prior probability (0.5,0.5).

More information

Hypercontractivity of spherical averages in Hamming space

Hypercontractivity of spherical averages in Hamming space Hypercontractivity of spherical averages in Hamming space Yury Polyanskiy Department of EECS MIT yp@mit.edu Allerton Conference, Oct 2, 2014 Yury Polyanskiy Hypercontractivity in Hamming space 1 Hypercontractivity:

More information

Notes 6 : First and second moment methods

Notes 6 : First and second moment methods Notes 6 : First and second moment methods Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Roc, Sections 2.1-2.3]. Recall: THM 6.1 (Markov s inequality) Let X be a non-negative

More information

Arrow Theorem. Elchanan Mossel. Draft All rights reserved

Arrow Theorem. Elchanan Mossel. Draft All rights reserved Arrow Theorem Elchanan Mossel Draft All rights reserved Condorcet Paradox n voters are to choose between 3 alternatives. Condorcet: Is there a rational way to do it? More specifically, for majority vote:

More information

Lecture 11. Probability Theory: an Overveiw

Lecture 11. Probability Theory: an Overveiw Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the

More information

Boolean Functions: Influence, threshold and noise

Boolean Functions: Influence, threshold and noise Boolean Functions: Influence, threshold and noise Einstein Institute of Mathematics Hebrew University of Jerusalem Based on recent joint works with Jean Bourgain, Jeff Kahn, Guy Kindler, Nathan Keller,

More information

Stein s Method for concentration inequalities

Stein s Method for concentration inequalities Number of Triangles in Erdős-Rényi random graph, UC Berkeley joint work with Sourav Chatterjee, UC Berkeley Cornell Probability Summer School July 7, 2009 Number of triangles in Erdős-Rényi random graph

More information

The Green-Tao theorem and a relative Szemerédi theorem

The Green-Tao theorem and a relative Szemerédi theorem The Green-Tao theorem and a relative Szemerédi theorem Yufei Zhao Massachusetts Institute of Technology Based on joint work with David Conlon (Oxford) and Jacob Fox (MIT) Green Tao Theorem (arxiv 2004;

More information

Hypercontractivity, Maximal Correlation and Non-interactive Simulation

Hypercontractivity, Maximal Correlation and Non-interactive Simulation Hypercontractivity, Maximal Correlation and Non-interactive Simulation Dept. Electrical Engineering, Stanford University Dec 4, 2014 Non-interactive Simulation Scenario: non-interactive simulation: Alice

More information

Concentration inequalities for non-lipschitz functions

Concentration inequalities for non-lipschitz functions Concentration inequalities for non-lipschitz functions University of Warsaw Berkeley, October 1, 2013 joint work with Radosław Adamczak (University of Warsaw) Gaussian concentration (Sudakov-Tsirelson,

More information

Non Linear Invariance & Applications. Elchanan Mossel. U.C. Berkeley

Non Linear Invariance & Applications. Elchanan Mossel. U.C. Berkeley Non Linear Invariance & Applications Elchanan Mossel U.C. Berkeley http://stat.berkeley.edu/~mossel 1 Talk Plan Probability and Gaussian Geometry: Non linear invariance for low influence functions Gaussian

More information

Solutions to Problem Set 5

Solutions to Problem Set 5 UC Berkeley, CS 74: Combinatorics and Discrete Probability (Fall 00 Solutions to Problem Set (MU 60 A family of subsets F of {,,, n} is called an antichain if there is no pair of sets A and B in F satisfying

More information

Probability Models of Information Exchange on Networks Lecture 1

Probability Models of Information Exchange on Networks Lecture 1 Probability Models of Information Exchange on Networks Lecture 1 Elchanan Mossel UC Berkeley All Rights Reserved Motivating Questions How are collective decisions made by: people / computational agents

More information

FKN Theorem on the biased cube

FKN Theorem on the biased cube FKN Theorem on the biased cube Piotr Nayar Abstract In this note we consider Boolean functions defined on the discrete cube { γ, γ 1 } n equipped with a product probability measure µ n, where µ = βδ γ

More information

Workshop on Discrete Harmonic Analysis Newton Institute, March 2011

Workshop on Discrete Harmonic Analysis Newton Institute, March 2011 High frequency criteria for Boolean functions (with an application to percolation) Christophe Garban ENS Lyon and CNRS Workshop on Discrete Harmonic Analysis Newton Institute, March 2011 C. Garban (ENS

More information

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27 Probability Review Yutian Li Stanford University January 18, 2018 Yutian Li (Stanford University) Probability Review January 18, 2018 1 / 27 Outline 1 Elements of probability 2 Random variables 3 Multiple

More information

Modern Discrete Probability Spectral Techniques

Modern Discrete Probability Spectral Techniques Modern Discrete Probability VI - Spectral Techniques Background Sébastien Roch UW Madison Mathematics December 22, 2014 1 Review 2 3 4 Mixing time I Theorem (Convergence to stationarity) Consider a finite

More information

18.175: Lecture 2 Extension theorems, random variables, distributions

18.175: Lecture 2 Extension theorems, random variables, distributions 18.175: Lecture 2 Extension theorems, random variables, distributions Scott Sheffield MIT Outline Extension theorems Characterizing measures on R d Random variables Outline Extension theorems Characterizing

More information

Quick Tour of Basic Probability Theory and Linear Algebra

Quick Tour of Basic Probability Theory and Linear Algebra Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions

More information

Rigorous approximation of invariant measures for IFS Joint work April 8, with 2016 S. 1 Galat / 21

Rigorous approximation of invariant measures for IFS Joint work April 8, with 2016 S. 1 Galat / 21 Rigorous approximation of invariant measures for IFS Joint work with S. Galatolo e I. Nisoli Maurizio Monge maurizio.monge@im.ufrj.br Universidade Federal do Rio de Janeiro April 8, 2016 Rigorous approximation

More information

On Extracting Common Random Bits From Correlated Sources on Large Alphabets

On Extracting Common Random Bits From Correlated Sources on Large Alphabets University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 3-2014 On Extracting Common Random Bits From Correlated Sources on Large Alphabets Siu On Chan Elchanan Mossel University

More information

Lecture 5: January 30

Lecture 5: January 30 CS71 Randomness & Computation Spring 018 Instructor: Alistair Sinclair Lecture 5: January 30 Disclaimer: These notes have not been subjected to the usual scrutiny accorded to formal publications. They

More information

Weak and strong moments of l r -norms of log-concave vectors

Weak and strong moments of l r -norms of log-concave vectors Weak and strong moments of l r -norms of log-concave vectors Rafał Latała based on the joint work with Marta Strzelecka) University of Warsaw Minneapolis, April 14 2015 Log-concave measures/vectors A measure

More information

Non-interactive Simulation of Joint Distributions: The Hirschfeld-Gebelein-Rényi Maximal Correlation and the Hypercontractivity Ribbon

Non-interactive Simulation of Joint Distributions: The Hirschfeld-Gebelein-Rényi Maximal Correlation and the Hypercontractivity Ribbon Non-interactive Simulation of Joint Distributions: The Hirschfeld-Gebelein-Rényi Maximal Correlation and the Hypercontractivity Ribbon Sudeep Kamath and Venkat Anantharam EECS Department University of

More information

Letting p shows that {B t } t 0. Definition 0.5. For λ R let δ λ : A (V ) A (V ) be defined by. 1 = g (symmetric), and. 3. g

Letting p shows that {B t } t 0. Definition 0.5. For λ R let δ λ : A (V ) A (V ) be defined by. 1 = g (symmetric), and. 3. g 4 Contents.1 Lie group p variation results Suppose G, d) is a group equipped with a left invariant metric, i.e. Let a := d e, a), then d ca, cb) = d a, b) for all a, b, c G. d a, b) = d e, a 1 b ) = a

More information

curvature, mixing, and entropic interpolation Simons Feb-2016 and CSE 599s Lecture 13

curvature, mixing, and entropic interpolation Simons Feb-2016 and CSE 599s Lecture 13 curvature, mixing, and entropic interpolation Simons Feb-2016 and CSE 599s Lecture 13 James R. Lee University of Washington Joint with Ronen Eldan (Weizmann) and Joseph Lehec (Paris-Dauphine) Markov chain

More information

Disjointness and Additivity

Disjointness and Additivity Midterm 2: Format Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten

More information

Input-queued switches: Scheduling algorithms for a crossbar switch. EE 384X Packet Switch Architectures 1

Input-queued switches: Scheduling algorithms for a crossbar switch. EE 384X Packet Switch Architectures 1 Input-queued switches: Scheduling algorithms for a crossbar switch EE 84X Packet Switch Architectures Overview Today s lecture - the input-buffered switch architecture - the head-of-line blocking phenomenon

More information

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and

More information

Lecture 14: October 22

Lecture 14: October 22 CS294 Markov Chain Monte Carlo: Foundations & Applications Fall 2009 Lecture 14: October 22 Lecturer: Alistair Sinclair Scribes: Alistair Sinclair Disclaimer: These notes have not been subjected to the

More information

Szemerédi s regularity lemma revisited. Lewis Memorial Lecture March 14, Terence Tao (UCLA)

Szemerédi s regularity lemma revisited. Lewis Memorial Lecture March 14, Terence Tao (UCLA) Szemerédi s regularity lemma revisited Lewis Memorial Lecture March 14, 2008 Terence Tao (UCLA) 1 Finding models of large dense graphs Suppose we are given a large dense graph G = (V, E), where V is a

More information

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten

More information

25.1 Markov Chain Monte Carlo (MCMC)

25.1 Markov Chain Monte Carlo (MCMC) CS880: Approximations Algorithms Scribe: Dave Andrzejewski Lecturer: Shuchi Chawla Topic: Approx counting/sampling, MCMC methods Date: 4/4/07 The previous lecture showed that, for self-reducible problems,

More information

215 Problem 1. (a) Define the total variation distance µ ν tv for probability distributions µ, ν on a finite set S. Show that

215 Problem 1. (a) Define the total variation distance µ ν tv for probability distributions µ, ν on a finite set S. Show that 15 Problem 1. (a) Define the total variation distance µ ν tv for probability distributions µ, ν on a finite set S. Show that µ ν tv = (1/) x S µ(x) ν(x) = x S(µ(x) ν(x)) + where a + = max(a, 0). Show that

More information

Lecture 35: December The fundamental statistical distances

Lecture 35: December The fundamental statistical distances 36-705: Intermediate Statistics Fall 207 Lecturer: Siva Balakrishnan Lecture 35: December 4 Today we will discuss distances and metrics between distributions that are useful in statistics. I will be lose

More information

Lecture 28: April 26

Lecture 28: April 26 CS271 Randomness & Computation Spring 2018 Instructor: Alistair Sinclair Lecture 28: April 26 Disclaimer: These notes have not been subjected to the usual scrutiny accorded to formal publications. They

More information

Notes on Gaussian processes and majorizing measures

Notes on Gaussian processes and majorizing measures Notes on Gaussian processes and majorizing measures James R. Lee 1 Gaussian processes Consider a Gaussian process {X t } for some index set T. This is a collection of jointly Gaussian random variables,

More information

Uncertainty Quantification in Computational Science

Uncertainty Quantification in Computational Science DTU 2010 - Lecture I Uncertainty Quantification in Computational Science Jan S Hesthaven Brown University Jan.Hesthaven@Brown.edu Objective of lectures The main objective of these lectures are To offer

More information

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M. Lecture 10 1 Ergodic decomposition of invariant measures Let T : (Ω, F) (Ω, F) be measurable, and let M denote the space of T -invariant probability measures on (Ω, F). Then M is a convex set, although

More information

The expansion of random regular graphs

The expansion of random regular graphs The expansion of random regular graphs David Ellis Introduction Our aim is now to show that for any d 3, almost all d-regular graphs on {1, 2,..., n} have edge-expansion ratio at least c d d (if nd is

More information

CSE 291: Fourier analysis Chapter 2: Social choice theory

CSE 291: Fourier analysis Chapter 2: Social choice theory CSE 91: Fourier analysis Chapter : Social choice theory 1 Basic definitions We can view a boolean function f : { 1, 1} n { 1, 1} as a means to aggregate votes in a -outcome election. Common examples are:

More information

The coupling method - Simons Counting Complexity Bootcamp, 2016

The coupling method - Simons Counting Complexity Bootcamp, 2016 The coupling method - Simons Counting Complexity Bootcamp, 2016 Nayantara Bhatnagar (University of Delaware) Ivona Bezáková (Rochester Institute of Technology) January 26, 2016 Techniques for bounding

More information

On Reparametrization and the Gibbs Sampler

On Reparametrization and the Gibbs Sampler On Reparametrization and the Gibbs Sampler Jorge Carlos Román Department of Mathematics Vanderbilt University James P. Hobert Department of Statistics University of Florida March 2014 Brett Presnell Department

More information

Lecture I: Asymptotics for large GUE random matrices

Lecture I: Asymptotics for large GUE random matrices Lecture I: Asymptotics for large GUE random matrices Steen Thorbjørnsen, University of Aarhus andom Matrices Definition. Let (Ω, F, P) be a probability space and let n be a positive integer. Then a random

More information

10-704: Information Processing and Learning Fall Lecture 9: Sept 28

10-704: Information Processing and Learning Fall Lecture 9: Sept 28 10-704: Information Processing and Learning Fall 2016 Lecturer: Siheng Chen Lecture 9: Sept 28 Note: These notes are based on scribed notes from Spring15 offering of this course. LaTeX template courtesy

More information

Rapid Introduction to Machine Learning/ Deep Learning

Rapid Introduction to Machine Learning/ Deep Learning Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian

More information

Functional Inequalities for Gaussian and Log-Concave Probability Measures

Functional Inequalities for Gaussian and Log-Concave Probability Measures Functional Inequalities for Gaussian and Log-Concave Probability Measures Ewain Gwynne Adviser: Professor Elton Hsu Northwestern University A thesis submitted for a Bachelor s degree in Mathematics (with

More information

Fourier analysis of boolean functions in quantum computation

Fourier analysis of boolean functions in quantum computation Fourier analysis of boolean functions in quantum computation Ashley Montanaro Centre for Quantum Information and Foundations, Department of Applied Mathematics and Theoretical Physics, University of Cambridge

More information

Coupling AMS Short Course

Coupling AMS Short Course Coupling AMS Short Course January 2010 Distance If µ and ν are two probability distributions on a set Ω, then the total variation distance between µ and ν is Example. Let Ω = {0, 1}, and set Then d TV

More information

STA205 Probability: Week 8 R. Wolpert

STA205 Probability: Week 8 R. Wolpert INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and

More information

1 Sequences of events and their limits

1 Sequences of events and their limits O.H. Probability II (MATH 2647 M15 1 Sequences of events and their limits 1.1 Monotone sequences of events Sequences of events arise naturally when a probabilistic experiment is repeated many times. For

More information

High-dimensional distributions with convexity properties

High-dimensional distributions with convexity properties High-dimensional distributions with convexity properties Bo az Klartag Tel-Aviv University A conference in honor of Charles Fefferman, Princeton, May 2009 High-Dimensional Distributions We are concerned

More information

Lecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016

Lecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016 Lecture 15: MCMC Sanjeev Arora Elad Hazan COS 402 Machine Learning and Artificial Intelligence Fall 2016 Course progress Learning from examples Definition + fundamental theorem of statistical learning,

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( ) Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca October 22nd, 2014 E. Tanré (INRIA - Team Tosca) Mathematical

More information

Lecture 7: Stochastic Dynamic Programing and Markov Processes

Lecture 7: Stochastic Dynamic Programing and Markov Processes Lecture 7: Stochastic Dynamic Programing and Markov Processes Florian Scheuer References: SLP chapters 9, 10, 11; LS chapters 2 and 6 1 Examples 1.1 Neoclassical Growth Model with Stochastic Technology

More information

Applied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books.

Applied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books. Applied Analysis APPM 44: Final exam 1:3pm 4:pm, Dec. 14, 29. Closed books. Problem 1: 2p Set I = [, 1]. Prove that there is a continuous function u on I such that 1 ux 1 x sin ut 2 dt = cosx, x I. Define

More information

Probability & Computing

Probability & Computing Probability & Computing Stochastic Process time t {X t t 2 T } state space Ω X t 2 state x 2 discrete time: T is countable T = {0,, 2,...} discrete space: Ω is finite or countably infinite X 0,X,X 2,...

More information

Chapter 9: Basic of Hypercontractivity

Chapter 9: Basic of Hypercontractivity Analysis of Boolean Functions Prof. Ryan O Donnell Chapter 9: Basic of Hypercontractivity Date: May 6, 2017 Student: Chi-Ning Chou Index Problem Progress 1 Exercise 9.3 (Tightness of Bonami Lemma) 2/2

More information

1 Ways to Describe a Stochastic Process

1 Ways to Describe a Stochastic Process purdue university cs 59000-nmc networks & matrix computations LECTURE NOTES David F. Gleich September 22, 2011 Scribe Notes: Debbie Perouli 1 Ways to Describe a Stochastic Process We will use the biased

More information

Notes on Probability

Notes on Probability Notes on Probability Mark Schmidt January 7, 2017 1 Probabilites Consider an event A that may or may not happen. For example, if we roll a dice then we may or may not roll a 6. We use the notation p(a)

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued

Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and

More information

The Green-Tao theorem and a relative Szemerédi theorem

The Green-Tao theorem and a relative Szemerédi theorem The Green-Tao theorem and a relative Szemerédi theorem Yufei Zhao Massachusetts Institute of Technology Joint work with David Conlon (Oxford) and Jacob Fox (MIT) Simons Institute December 2013 Green Tao

More information

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < + Random Walks: WEEK 2 Recurrence and transience Consider the event {X n = i for some n > 0} by which we mean {X = i}or{x 2 = i,x i}or{x 3 = i,x 2 i,x i},. Definition.. A state i S is recurrent if P(X n

More information

MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016

MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016 MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016 Lecture 14: Information Theoretic Methods Lecturer: Jiaming Xu Scribe: Hilda Ibriga, Adarsh Barik, December 02, 2016 Outline f-divergence

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

Functional Analysis, Math 7320 Lecture Notes from August taken by Yaofeng Su

Functional Analysis, Math 7320 Lecture Notes from August taken by Yaofeng Su Functional Analysis, Math 7320 Lecture Notes from August 30 2016 taken by Yaofeng Su 1 Essentials of Topology 1.1 Continuity Next we recall a stronger notion of continuity: 1.1.1 Definition. Let (X, d

More information

Characterization of cutoff for reversible Markov chains

Characterization of cutoff for reversible Markov chains Characterization of cutoff for reversible Markov chains Yuval Peres Joint work with Riddhi Basu and Jonathan Hermon 23 Feb 2015 Joint work with Riddhi Basu and Jonathan Hermon Characterization of cutoff

More information

2. The averaging process

2. The averaging process 2. The averaging process David Aldous July 10, 2012 The two models in lecture 1 has only rather trivial interaction. In this lecture I discuss a simple FMIE process with less trivial interaction. It is

More information

Lecture Notes 3 Convergence (Chapter 5)

Lecture Notes 3 Convergence (Chapter 5) Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Approximation & Complexity

Approximation & Complexity Summer school on semidefinite optimization Approximation & Complexity David Steurer Cornell University Part 1 September 6, 2012 Overview Part 1 Unique Games Conjecture & Basic SDP Part 2 SDP Hierarchies:

More information

A Characterization of Locally Testable Affine-Invariant Properties via Decomposition Theorems

A Characterization of Locally Testable Affine-Invariant Properties via Decomposition Theorems A Characterization of Locally Testable Affine-Invariant Properties via Decomposition Theorems Yuichi Yoshida National Institute of Informatics and Preferred Infrastructure, Inc June 1, 2014 Yuichi Yoshida

More information

Information Theory and Statistics Lecture 3: Stationary ergodic processes

Information Theory and Statistics Lecture 3: Stationary ergodic processes Information Theory and Statistics Lecture 3: Stationary ergodic processes Łukasz Dębowski ldebowsk@ipipan.waw.pl Ph. D. Programme 2013/2014 Measurable space Definition (measurable space) Measurable space

More information

Exercises Measure Theoretic Probability

Exercises Measure Theoretic Probability Exercises Measure Theoretic Probability 2002-2003 Week 1 1. Prove the folloing statements. (a) The intersection of an arbitrary family of d-systems is again a d- system. (b) The intersection of an arbitrary

More information

EE514A Information Theory I Fall 2013

EE514A Information Theory I Fall 2013 EE514A Information Theory I Fall 2013 K. Mohan, Prof. J. Bilmes University of Washington, Seattle Department of Electrical Engineering Fall Quarter, 2013 http://j.ee.washington.edu/~bilmes/classes/ee514a_fall_2013/

More information

Math 456: Mathematical Modeling. Tuesday, March 6th, 2018

Math 456: Mathematical Modeling. Tuesday, March 6th, 2018 Math 456: Mathematical Modeling Tuesday, March 6th, 2018 Markov Chains: Exit distributions and the Strong Markov Property Tuesday, March 6th, 2018 Last time 1. Weighted graphs. 2. Existence of stationary

More information

LECTURE 3. Last time:

LECTURE 3. Last time: LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE Most of the material in this lecture is covered in [Bertsekas & Tsitsiklis] Sections 1.3-1.5

More information

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16 EE539R: Problem Set 4 Assigned: 3/08/6, Due: 07/09/6. Cover and Thomas: Problem 3.5 Sets defined by probabilities: Define the set C n (t = {x n : P X n(x n 2 nt } (a We have = P X n(x n P X n(x n 2 nt

More information

Stochastic Problems. 1 Examples. 1.1 Neoclassical Growth Model with Stochastic Technology. 1.2 A Model of Job Search

Stochastic Problems. 1 Examples. 1.1 Neoclassical Growth Model with Stochastic Technology. 1.2 A Model of Job Search Stochastic Problems References: SLP chapters 9, 10, 11; L&S chapters 2 and 6 1 Examples 1.1 Neoclassical Growth Model with Stochastic Technology Production function y = Af k where A is random Let A s t

More information

Probability Background

Probability Background Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second

More information

1 Math 241A-B Homework Problem List for F2015 and W2016

1 Math 241A-B Homework Problem List for F2015 and W2016 1 Math 241A-B Homework Problem List for F2015 W2016 1.1 Homework 1. Due Wednesday, October 7, 2015 Notation 1.1 Let U be any set, g be a positive function on U, Y be a normed space. For any f : U Y let

More information

A Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones

A Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones A Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones Abstract We present a regularity lemma for Boolean functions f : {, } n {, } based on noisy influence, a measure of how locally correlated

More information

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time

More information

Spectral properties of Markov operators in Markov chain Monte Carlo

Spectral properties of Markov operators in Markov chain Monte Carlo Spectral properties of Markov operators in Markov chain Monte Carlo Qian Qin Advisor: James P. Hobert October 2017 1 Introduction Markov chain Monte Carlo (MCMC) is an indispensable tool in Bayesian statistics.

More information

MAA704, Perron-Frobenius theory and Markov chains.

MAA704, Perron-Frobenius theory and Markov chains. November 19, 2013 Lecture overview Today we will look at: Permutation and graphs. Perron frobenius for non-negative. Stochastic, and their relation to theory. Hitting and hitting probabilities of chain.

More information

BALANCING GAUSSIAN VECTORS. 1. Introduction

BALANCING GAUSSIAN VECTORS. 1. Introduction BALANCING GAUSSIAN VECTORS KEVIN P. COSTELLO Abstract. Let x 1,... x n be independent normally distributed vectors on R d. We determine the distribution function of the minimum norm of the 2 n vectors

More information

5 Birkhoff s Ergodic Theorem

5 Birkhoff s Ergodic Theorem 5 Birkhoff s Ergodic Theorem Birkhoff s Ergodic Theorem extends the validity of Kolmogorov s strong law to the class of stationary sequences of random variables. Stationary sequences occur naturally even

More information

Consistency of the maximum likelihood estimator for general hidden Markov models

Consistency of the maximum likelihood estimator for general hidden Markov models Consistency of the maximum likelihood estimator for general hidden Markov models Jimmy Olsson Centre for Mathematical Sciences Lund University Nordstat 2012 Umeå, Sweden Collaborators Hidden Markov models

More information

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 1 Entropy Since this course is about entropy maximization,

More information

An exponential separation between quantum and classical one-way communication complexity

An exponential separation between quantum and classical one-way communication complexity An exponential separation between quantum and classical one-way communication complexity Ashley Montanaro Centre for Quantum Information and Foundations, Department of Applied Mathematics and Theoretical

More information

Roth s Theorem on 3-term Arithmetic Progressions

Roth s Theorem on 3-term Arithmetic Progressions Roth s Theorem on 3-term Arithmetic Progressions Mustazee Rahman 1 Introduction This article is a discussion about the proof of a classical theorem of Roth s regarding the existence of three term arithmetic

More information

Logarithmic Sobolev Inequalities

Logarithmic Sobolev Inequalities Logarithmic Sobolev Inequalities M. Ledoux Institut de Mathématiques de Toulouse, France logarithmic Sobolev inequalities what they are, some history analytic, geometric, optimal transportation proofs

More information

Lecture Introduction. 2 Brief Recap of Lecture 10. CS-621 Theory Gems October 24, 2012

Lecture Introduction. 2 Brief Recap of Lecture 10. CS-621 Theory Gems October 24, 2012 CS-62 Theory Gems October 24, 202 Lecture Lecturer: Aleksander Mądry Scribes: Carsten Moldenhauer and Robin Scheibler Introduction In Lecture 0, we introduced a fundamental object of spectral graph theory:

More information

Inference for Stochastic Processes

Inference for Stochastic Processes Inference for Stochastic Processes Robert L. Wolpert Revised: June 19, 005 Introduction A stochastic process is a family {X t } of real-valued random variables, all defined on the same probability space

More information

µ X (A) = P ( X 1 (A) )

µ X (A) = P ( X 1 (A) ) 1 STOCHASTIC PROCESSES This appendix provides a very basic introduction to the language of probability theory and stochastic processes. We assume the reader is familiar with the general measure and integration

More information

Lecture 10: Semi-Markov Type Processes

Lecture 10: Semi-Markov Type Processes Lecture 1: Semi-Markov Type Processes 1. Semi-Markov processes (SMP) 1.1 Definition of SMP 1.2 Transition probabilities for SMP 1.3 Hitting times and semi-markov renewal equations 2. Processes with semi-markov

More information