Mixing in Product Spaces. Elchanan Mossel
|
|
- Edgar O’Neal’
- 5 years ago
- Views:
Transcription
1
2 Poincaré Recurrence Theorem Theorem (Poincaré, 1890) Let f : X X be a measure preserving transformation. Let E X measurable. Then P[x E : f n (x) / E, n > N(x)] = 0
3 Poincaré Recurrence Theorem Theorem (Poincaré, 1890) Let f : X X be a measure preserving transformation. Let E X measurable. Then P[x E : f n (x) / E, n > N(x)] = 0 One of the first results in Ergodic Theory. Long term mixing. This talk is about short term mixing.
4 Finite Markov Chains As a first example consider a Finite Markov chain. Let M be a k k doubly stochastic symmetric matrix. Pick X 0 uniformly at random from 1,..., k. Given X i = a, let X i+1 = b with probability M a,b.
5 Finite Markov Chains As a first example consider a Finite Markov chain. Let M be a k k doubly stochastic symmetric matrix. Pick X 0 uniformly at random from 1,..., k. Given X i = a, let X i+1 = b with probability M a,b. Theorem (Long Term Mixing for Markov Chains) Suppose that other than 1, all eigenvalues λ i of M satisfy λ i λ < 1. Then for any two sets A, B [k], it holds that P[X 0 A, X t B] P[A]P[B] λ t
6 Short Term Mixing for Markov Chains Theorem P[X 0 A, X 1 B] P[A]P[B] is upper bounded by λ P[A](1 P[A])P[B](1 P[B]) Shows: mixing in one step for large sets.
7 Short Term Mixing for Markov Chains Theorem P[X 0 A, X 1 B] P[A]P[B] is upper bounded by λ P[A](1 P[A])P[B](1 P[B]) Shows: mixing in one step for large sets. Proof: 1 A = P[A]1 + f, 1 B = P[B]1 + g, where f, g 1
8 Short Term Mixing for Markov Chains Theorem P[X 0 A, X 1 B] P[A]P[B] is upper bounded by λ P[A](1 P[A])P[B](1 P[B]) Shows: mixing in one step for large sets. Proof: 1 A = P[A]1 + f, 1 B = P[B]1 + g, where f, g 1 P[X 0 A, X 1 B] = 1 k (P[A]1 + f )t M(P[B]1 + g) = P[A]P[B] + 1 k f t Mg, 1 k f t Mg λ f 2 g 2 = λ P[A](1 P[A])P[B](1 P[B])
9 Short Term Mixing for Markov Chains Theorem P[X 0 A, X 1 B] P[A]P[B] is upper bounded by λ P[A](1 P[A])P[B](1 P[B]) Shows: mixing in one step for large sets. Proof: 1 A = P[A]1 + f, 1 B = P[B]1 + g, where f, g 1 P[X 0 A, X 1 B] = 1 k (P[A]1 + f )t M(P[B]1 + g) = P[A]P[B] + 1 k f t Mg, 1 k f t Mg λ f 2 g 2 = λ P[A](1 P[A])P[B](1 P[B]) Also called Expander Mixing Lemma. Used a lot in computer science, e.g. in (de)randomization.
10 The tensor property Consider (Y 1, Z 1 ),..., (Y n, Z n ) which are drawn independently from the distribution of (X 0, X 1 ). Equivalently, the transition matrix from Y = (Y 1,..., Y n ) to Z = (Z 1,..., Z n ) is M n.
11 The tensor property Consider (Y 1, Z 1 ),..., (Y n, Z n ) which are drawn independently from the distribution of (X 0, X 1 ). Equivalently, the transition matrix from Y = (Y 1,..., Y n ) to Z = (Z 1,..., Z n ) is M n. Thm = that for any sets A, B [k] n : P[Y A, Z B] P[A]P[B] λ P[A](1 P[A])P[B](1 P[B])
12 The tensor property Consider (Y 1, Z 1 ),..., (Y n, Z n ) which are drawn independently from the distribution of (X 0, X 1 ). Equivalently, the transition matrix from Y = (Y 1,..., Y n ) to Z = (Z 1,..., Z n ) is M n. Thm = that for any sets A, B [k] n : P[Y A, Z B] P[A]P[B] λ P[A](1 P[A])P[B](1 P[B]) Follows immediately from tensorization of the spectrum.
13 Log Sobolev inequalities Entropy, Log Sobolev and hyper-contraction A similar story could be told using more sophisticated analytic tools. Easier to work with Markov semi-groups T t = e tl.
14 Log Sobolev inequalities Entropy, Log Sobolev and hyper-contraction A similar story could be told using more sophisticated analytic tools. Easier to work with Markov semi-groups T t = e tl. Entropy, Dirchelet Form Ent(f ) = E(f log f ) Ef log Ef E(f, g) = E(fLg) = E(gLf ) = E(g, f ) = d dt EfT tg. t=0
15 Log Sobolev inequalities Entropy, Log Sobolev and hyper-contraction A similar story could be told using more sophisticated analytic tools. Easier to work with Markov semi-groups T t = e tl. Entropy, Dirchelet Form Ent(f ) = E(f log f ) Ef log Ef E(f, g) = E(fLg) = E(gLf ) = E(g, f ) = d dt EfT tg. t=0 Definition of Log-Sob p-logsob(c) f, Ent(f p ) Cp2 4(p 1) E(f p 1, f ) (p 0, 1) 1-logSob(C) f, Ent(f ) C 4 E(f, log f ) 0-logSob(C) f, Var(log f ) C 2 E(f, 1/f )
16 Log Sob. Inequalities and Hyper-Contraction Hyper-Contraction (Gross, Nelson ) r-logsob with constant C implies T t f p f q, t C 4 log p 1 q 1, 1 < p < q < r or r < q < p = E[g(X 0 )f (X t )] = E[gT t f g p Tf p g p f q If f = 1 A and g = 1 B, get: P[X 0 A, X t B] 1 A q 1 B p = P[A] 1/q P[B] 1/p, Now optimize over norms to get a better bound than CS.
17 Reverse-Hyper-Contraction Log-Sobolev and Rev. Hyper-Contraction(M-Oleszkiewicz-Sen-13) Let T t = e tl be a general Markov semi-group satisfying 2-Logsob with constant C or 1-Logsob inequality with constant C. Then for all q < p < 1, all positive f, g and all t C 4 holds that T t f q f p = E[g(X 0 )f (X t )] = E[gT t f ] g q f p log 1 q 1 p it
18 Short-Time Implications Theorem (M-Oleszkiewicz-Sen-13 ; Short-Time Implications) Let T t = e tl, where L satisfy 1 or 2-LogSob inequality with constant C. Let A, B Ω n with P[A] ɛ and P[B] ɛ. Then: P[X (0) A, X (t) B] ɛ 2 1 e 2t/C
19 Short-Time Implications Theorem (M-Oleszkiewicz-Sen-13 ; Short-Time Implications) Let T t = e tl, where L satisfy 1 or 2-LogSob inequality with constant C. Let A, B Ω n with P[A] ɛ and P[B] ɛ. Then: P[X (0) A, X (t) B] ɛ 2 1 e 2t/C Comments 1. Works for small sets too. 2. Tensorizes. 3. Some examples where it is (almost) tight. 4. Uses in social choice analysis, queuing theory.
20 Comment: typical application MCMC Long Time Behavior Log Sobolev inequalities play a major role in analyzing long term mixing of Markov chains, in particular in analysis of mixing times (Diaconis, Saloff-Coste etc.) Long Time Behavior The ε-total variation mixing time of a finite Markov chain is bounded by: 1 λ (log(1/π ) + log(1/ɛ)) 1 C (log log(1/π ) + log(1/ɛ)) for a continuous time Markov chain with spectral gap λ and 2-LogSob C.
21 Comment: typical application MCMC Long Time Behavior Log Sobolev inequalities play a major role in analyzing long term mixing of Markov chains, in particular in analysis of mixing times (Diaconis, Saloff-Coste etc.) Long Time Behavior The ε-total variation mixing time of a finite Markov chain is bounded by: 1 λ (log(1/π ) + log(1/ɛ)) 1 C (log log(1/π ) + log(1/ɛ)) for a continuous time Markov chain with spectral gap λ and 2-LogSob C.
22 What are these lectures about? High Dimensional Phenomena High dimensional mixing: mixing of product processes on product spaces Ω n with n large. Tight bounds For which processes, given measures a and b can we find precise upper/lower bounds for sup ( P[X 0 A, X t B] : P[A] = a, P[B] = b ) Interested in product space/processes of dimension n and answers as n. Most important examples / techniques from probability / analysis.
23 What are these lectures about? Mulit-step prcoesses How to bound P[X 0 A 0, X 1 A 1,..., X k A k ] for processes X 0,..., X k? Interested in product space/processes of dimension n and answers as n. Most important examples / techniques from additive combinatorics.
24 What are these lectures about? And more Theory that does both? Applications?
25 Today: tight bounds Borell s result. Open Problem: The Boolean cube. The state of affairs - partition into 3 parts or more.
26 Two Examples: Gaussian, Boolean Correlated pairs (M-O Donnell-Regev-Steif-Sudakov-05): Let x, y { 1, 1} n be e t correlated: x is chosen uniformly and y is T t correlated version. i.e. E[x i y i ] = e t for all i independently Let A, B { 1, 1} n 1/2 with P[A] ɛ and P[B] ɛ Then: P[x A, y B] ɛ 2 1 e t Easy to prove when A = B...
27 Two Examples: Gaussian, Boolean Correlated pairs (M-O Donnell-Regev-Steif-Sudakov-05): Let x, y { 1, 1} n be e t correlated: x is chosen uniformly and y is T t correlated version. i.e. E[x i y i ] = e t for all i independently Let A, B { 1, 1} n 1/2 with P[A] ɛ and P[B] ɛ Then: P[x A, y B] ɛ 2 1 e t Easy to prove when A = B... Gaussian Version Let x, y R n two Gaussian vectors: x N(0, 1), y N(0, 1), E[x i y j ] = e t δ i,j Let A, B R n with P[A] ɛ and P[B] ɛ Then: P[x A, y B] ɛ 2 1 e t
28 Borell s Result and Open Problems Borell (85): In Gaussian case the maximum and minimum of P[x A, y B] as a function of P[A] and P[B] is obtained for parallel half-spaces.
29 Borell s Result and Open Problems Borell (85): In Gaussian case the maximum and minimum of P[x A, y B] as a function of P[A] and P[B] is obtained for parallel half-spaces. Do not know what is the optimum in { 1, 1} n. In particular:
30 Borell s Result and Open Problems Borell (85): In Gaussian case the maximum and minimum of P[x A, y B] as a function of P[A] and P[B] is obtained for parallel half-spaces. Do not know what is the optimum in { 1, 1} n. In particular: Open Problem: lim min(p[x A, Y B] : A, B { 1, n 1}n, P[A] = P[B] = 1/4) and similarly for max.
31 Borell s Result and Open Problems Borell (85): In Gaussian case the maximum and minimum of P[x A, y B] as a function of P[A] and P[B] is obtained for parallel half-spaces. Do not know what is the optimum in { 1, 1} n. In particular: Open Problem: lim min(p[x A, Y B] : A, B { 1, n 1}n, P[A] = P[B] = 1/4) and similarly for max. Partition to 3 or more parts even in Gaussian space.
32 If there is time before the break... A cute proof of a special case of Borell s result. Connections to social choice Theory.
33 Simple Example 1 Cosmic coin problem(m-o Donnell-05): x { 1, 1} n uniform. (y i ) m 1 conditionally independent given x. Each pair (x, y i ) is ρ-correlated. Problem: What is the largest P[y 1 A,... y m A] can be?
34 Simple Example 2 (y i,j ) 1 i<j m is an exchangeable collection of vectors in { 1, 1} n. If I J = 1 then y I, y J are 1/3 correlated. Otherwise independent. Why? If n voters rank alternatives uniformly at random, the pairwise preferences between alternatives will be given by the collection y.
35 Full support finite Ω using hyper-contraction Thm: More General Reverse Hypercontractivity Theorem (M-Oleszkiewicz-Sen-13) Let a the measure Ψ over a finite Ω k satisfy min x1,...,x k Ω Pr[X 1 = x,..., X k = x k ] = α > 0 and have equal marginals.
36 Full support finite Ω using hyper-contraction Thm: More General Reverse Hypercontractivity Theorem (M-Oleszkiewicz-Sen-13) Let a the measure Ψ over a finite Ω k satisfy min x1,...,x k Ω Pr[X 1 = x,..., X k = x k ] = α > 0 and have equal marginals. Consider the distribution Ψ n and let A 1,..., A k Ω n, µ(a i ) µ. Then: Pr[X 1 A 1,... X k A k ] µ O( 1 α), where (X 1 (i),..., X k (i)) are i.i.d. according to Ψ. Note This is a key tool of analyzing the examples above as well as many others.
37 Notation Distributed according to P := P n. Tuples X i are i.i.d. according to P. The marginals of P are π j. Vectors X (j) are distributed according to π j := π n j. X X 1 X 2... X i... X n X (1) X (1) 1 X (1) 2 X (2) X (2) 1 X (2) 2. X (j)... X (j) 1 X (j) 2.. X (1) i X (2) i. X (j) i. X (l) X (l) X (l) 1 X (l) 2 i X (1) n X (2) n. X n (j). X n (l)
38 Lower Bounds We are mostly interested in two types of lower bounds: Set hitting: Lower bounds on in terms of P[A 1 ],..., P[A k ] P[X 1 A 1,..., X k A k ] Same set hitting: Lower bounds on in terms of P[A]. P[X 1 A,..., X k A] Set hitting will require something... - e.g. X 1 = X 2 =... = X k.
39 Gaussian Bounds Borell (85) k = 2 - parallel half-spaces are optimal (also Isaksson-Mossel, Neeman) By a Reverse Brascamp-Lieb inq. (Ledoux, Chen-Dafnis-Paouris 14-15) for A,... C R n : P[U A,..., Z C] (P[A] P[C]) 1/(1 ρ2), where ρ is the second eigenvalue of Σ. Doesn t require independence of coordinates
40 Full Support Case Thm: More General Reverse Hypercontractivity Theorem (M-Oleszkiewicz-Sen-13) Let a the measure Ψ over a finite Ω k satisfy min x1,...,x k Ω Pr[X 1 = x,..., X k = x k ] = α > 0 and have equal marginals.
41 Full Support Case Thm: More General Reverse Hypercontractivity Theorem (M-Oleszkiewicz-Sen-13) Let a the measure Ψ over a finite Ω k satisfy min x1,...,x k Ω Pr[X 1 = x,..., X k = x k ] = α > 0 and have equal marginals. Then: Pr[X 1 A 1,... X k A k ] µ O( 1 α), where (X 1 (i),..., X k (i)) are i.i.d. according to Ψ.
42 Non full support? What if the support of Ω is not full? Do we care?
43 Non full support? What if the support of Ω is not full? Do we care? Maybe: This is what additive combinatorics is all about. In particular: finite cominatorics in finite field models (Green ). Many other applications in combinatorics and computer science.
44 Additive combinatorics perspective Example: Theorem (Finite Field Roth Theorem) Y, R be chosen uniformly at random at F n 3. Then for every µ > 0 there exists c(µ) > 0, N(µ) such that if n N(µ) and A F n 3 satisfies P[A] µ, then: P[Y A, Y + R A, Y + 2R A] c(µ).
45 Additive combinatorics perspective Example: Theorem (Finite Field Roth Theorem) Y, R be chosen uniformly at random at F n 3. Then for every µ > 0 there exists c(µ) > 0, N(µ) such that if n N(µ) and A F n 3 Why is this true? satisfies P[A] µ, then: P[Y A, Y + R A, Y + 2R A] c(µ).
46 Fourier Obstructions Theorem (Finite Field Roth Theorem - Analysis) Let Y, R be chosen uniformly at random at F n 3. Let A, B, C F n 3 then P[Y A, Y + R B, Y + 2R C] P[A]P[B]P[C] Â
47 Fourier Obstructions Theorem (Finite Field Roth Theorem - Analysis) Let Y, R be chosen uniformly at random at F n 3. Let A, B, C F n 3 then P[Y A, Y + R B, Y + 2R C] P[A]P[B]P[C] Â Only obstruction to uniformity is linear structure
48 Fourier Obstructions Theorem (Finite Field Roth Theorem - Analysis) Let Y, R be chosen uniformly at random at F n 3. Let A, B, C F n 3 then P[Y A, Y + R B, Y + 2R C] P[A]P[B]P[C] Â Only obstruction to uniformity is linear structure If A = B = C, high Fourier coefficient = can restrict to linear subspace with higher denisty
49 Fourier Obstructions Theorem (Finite Field Roth Theorem - Analysis) Let Y, R be chosen uniformly at random at F n 3. Let A, B, C F n 3 then P[Y A, Y + R B, Y + 2R C] P[A]P[B]P[C] Â Only obstruction to uniformity is linear structure If A = B = C, high Fourier coefficient = can restrict to linear subspace with higher denisty Density increase arguments...
50 Higher Order Arithmetic Obstructions Furstenberg-Weiss (80s): For longer arithmetic progressions, obstructions other than Fourier. Gowers: Obstructions can be identified using the Gowers norms. Again - use obstruction to your benefit. Thm: (Gowers 08; Rodel and Skokan 04,06): If q is prime and l q then for every µ > 0 there exists c(µ) > 0, N(µ) such that if n N(µ) and A Fq n satisfies P[A] µ, then: P[Y A, Y + R A,..., Y + (l 1)R A] c(µ), where A, R F n q are chosen uniformly at random.
51 Higher Order Arithmetic Obstructions Furstenberg-Weiss (80s): For longer arithmetic progressions, obstructions other than Fourier. Gowers: Obstructions can be identified using the Gowers norms. Again - use obstruction to your benefit. Thm: (Gowers 08; Rodel and Skokan 04,06): If q is prime and l q then for every µ > 0 there exists c(µ) > 0, N(µ) such that if n N(µ) and A Fq n satisfies P[A] µ, then: P[Y A, Y + R A,..., Y + (l 1)R A] c(µ), where A, R F n q are chosen uniformly at random. Question: Is the additive structure necessary?
52 Obstruction to Chaos Consider the support of Ω as a graph G with vertex V = all atoms with non-zero weight and edges between any two atoms that differ in one coordinate. We say that ρ < 1 if the graph G is connected. More formally: Definition { ρ(p, S, T ) := sup Cov[f (X (S) ), g(x (T ) )] f : Ω (S) R, g : Ω (T ) R, } Var[f (X (S) )] = Var[g(X (T ) )] = 1. The correlation of P is ρ(p) := max j [l] ρ (P, {j}, [l] \ {j}).
53 The quest for a unifying theory Is there one theory that explains both the noisy examples and the additive theory?
54 Example Let X be uniform in F n 3. Let Y i = X i or X i + 1 with probability 1/2 independently for each coordinate. Theorem = P[X A, Y a] c(p[a]). Motivation from understanding parallel repetition. Does not follow from hyper-contraction nor does it follow from additive techniques...
55 A General Result Theorem (+ Hazla, Holenstein) Suppose (X, Y ) is distributed in a finite Ω 2 such that: α = min a P[X = Y = a] > 0. P[X = a] = P[Y = a] for all a. Then for any set A Ω n with P X n[a] = P Y n[a] µ it holds that P[X A, Y A] c(α, µ) > 0 Our c is pretty bad: c = 1/ exp(exp(exp(1/(µ) D ))), D = D(α)
56 A General Result Theorem (+ Hazla, Holenstein) Suppose (X, Y ) is distributed in a finite Ω 2 such that: α = min a P[X = Y = a] > 0. P[X = a] = P[Y = a] for all a. Then for any set A Ω n with P X n[a] = P Y n[a] µ it holds that P[X A, Y A] c(α, µ) > 0 Our c is pretty bad: c = 1/ exp(exp(exp(1/(µ) D ))), D = D(α) Related to the fact that the proof is interesting: 1 Lose in Regularity Lemma type arguments. 2 Lose in Invariance transforming the problem to a Gaussian problem.
57 A Markov Chain Theorem and a general process theorem Theorem[+Hazla, Holenstein] X i, Y i, Z i,..., W i be a Markov chain over Ω with min x Ω Pr[X i = Y i = Z i =... W i = x] = β > 0 and uniform marginals. Let A Ω n, µ(a) = µ > 0. Pr[X A Y A Z A,..., W A] f (µ, β) > 0. Theorem[+Hazla, Holenstein] X i, Y i, Z i,..., W i be distributed over Ω k with min x Ω Pr[X i = Y i = Z i =... W i = x] = β > 0 and uniform marginals. Suppose further that ρ(x i, Y i,..., W i ) < 1. Let A Ω n, µ(a) = µ > 0. Pr[X A Y A Z A,..., W A] f (µ, β) > 0.
58 The condition ρ < 1 Weaker than full support.
59 The condition ρ < 1 Weaker than full support. Does not hold in arithmetic setups.
60 The condition ρ < 1 Weaker than full support. Does not hold in arithmetic setups. ρ = 1 iff the support of Ψ is connected with respect to changing one coordinate at a time.
61 The condition ρ < 1 Weaker than full support. Does not hold in arithmetic setups. ρ = 1 iff the support of Ψ is connected with respect to changing one coordinate at a time. Example: (x, y) F3 2 where y = x, x + 1 has ρ < 1 but not full support.
62 Open Problems Still searching for unified theory. Concrete Example: Suppose Ψ is uniform over {(0, 0, 0), (1, 1, 1), (2, 2, 2), (0, 1, 2), (1, 2, 0), (2, 0, 1)} ρ = 1 but not arithmetic. Do not understand.
63 Questions??
64 Questions?? Thank you!
Reverse Hyper Contractive Inequalities: What? Why?? When??? How????
Reverse Hyper Contractive Inequalities: What? Why?? When??? How???? UC Berkeley Warsaw Cambridge May 21, 2012 Hyper-Contractive Inequalities The noise (Bonami-Beckner) semi-group on { 1, 1} n 1 2 Def 1:
More informationHeat Flows, Geometric and Functional Inequalities
Heat Flows, Geometric and Functional Inequalities M. Ledoux Institut de Mathématiques de Toulouse, France heat flow and semigroup interpolations Duhamel formula (19th century) pde, probability, dynamics
More informationProbability Models of Information Exchange on Networks Lecture 3
Probability Models of Information Exchange on Networks Lecture 3 Elchanan Mossel UC Berkeley All Right Reserved The Bayesian View of the Jury Theorem Recall: we assume 0/1 with prior probability (0.5,0.5).
More informationHypercontractivity of spherical averages in Hamming space
Hypercontractivity of spherical averages in Hamming space Yury Polyanskiy Department of EECS MIT yp@mit.edu Allerton Conference, Oct 2, 2014 Yury Polyanskiy Hypercontractivity in Hamming space 1 Hypercontractivity:
More informationNotes 6 : First and second moment methods
Notes 6 : First and second moment methods Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Roc, Sections 2.1-2.3]. Recall: THM 6.1 (Markov s inequality) Let X be a non-negative
More informationArrow Theorem. Elchanan Mossel. Draft All rights reserved
Arrow Theorem Elchanan Mossel Draft All rights reserved Condorcet Paradox n voters are to choose between 3 alternatives. Condorcet: Is there a rational way to do it? More specifically, for majority vote:
More informationLecture 11. Probability Theory: an Overveiw
Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the
More informationBoolean Functions: Influence, threshold and noise
Boolean Functions: Influence, threshold and noise Einstein Institute of Mathematics Hebrew University of Jerusalem Based on recent joint works with Jean Bourgain, Jeff Kahn, Guy Kindler, Nathan Keller,
More informationStein s Method for concentration inequalities
Number of Triangles in Erdős-Rényi random graph, UC Berkeley joint work with Sourav Chatterjee, UC Berkeley Cornell Probability Summer School July 7, 2009 Number of triangles in Erdős-Rényi random graph
More informationThe Green-Tao theorem and a relative Szemerédi theorem
The Green-Tao theorem and a relative Szemerédi theorem Yufei Zhao Massachusetts Institute of Technology Based on joint work with David Conlon (Oxford) and Jacob Fox (MIT) Green Tao Theorem (arxiv 2004;
More informationHypercontractivity, Maximal Correlation and Non-interactive Simulation
Hypercontractivity, Maximal Correlation and Non-interactive Simulation Dept. Electrical Engineering, Stanford University Dec 4, 2014 Non-interactive Simulation Scenario: non-interactive simulation: Alice
More informationConcentration inequalities for non-lipschitz functions
Concentration inequalities for non-lipschitz functions University of Warsaw Berkeley, October 1, 2013 joint work with Radosław Adamczak (University of Warsaw) Gaussian concentration (Sudakov-Tsirelson,
More informationNon Linear Invariance & Applications. Elchanan Mossel. U.C. Berkeley
Non Linear Invariance & Applications Elchanan Mossel U.C. Berkeley http://stat.berkeley.edu/~mossel 1 Talk Plan Probability and Gaussian Geometry: Non linear invariance for low influence functions Gaussian
More informationSolutions to Problem Set 5
UC Berkeley, CS 74: Combinatorics and Discrete Probability (Fall 00 Solutions to Problem Set (MU 60 A family of subsets F of {,,, n} is called an antichain if there is no pair of sets A and B in F satisfying
More informationProbability Models of Information Exchange on Networks Lecture 1
Probability Models of Information Exchange on Networks Lecture 1 Elchanan Mossel UC Berkeley All Rights Reserved Motivating Questions How are collective decisions made by: people / computational agents
More informationFKN Theorem on the biased cube
FKN Theorem on the biased cube Piotr Nayar Abstract In this note we consider Boolean functions defined on the discrete cube { γ, γ 1 } n equipped with a product probability measure µ n, where µ = βδ γ
More informationWorkshop on Discrete Harmonic Analysis Newton Institute, March 2011
High frequency criteria for Boolean functions (with an application to percolation) Christophe Garban ENS Lyon and CNRS Workshop on Discrete Harmonic Analysis Newton Institute, March 2011 C. Garban (ENS
More informationProbability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27
Probability Review Yutian Li Stanford University January 18, 2018 Yutian Li (Stanford University) Probability Review January 18, 2018 1 / 27 Outline 1 Elements of probability 2 Random variables 3 Multiple
More informationModern Discrete Probability Spectral Techniques
Modern Discrete Probability VI - Spectral Techniques Background Sébastien Roch UW Madison Mathematics December 22, 2014 1 Review 2 3 4 Mixing time I Theorem (Convergence to stationarity) Consider a finite
More information18.175: Lecture 2 Extension theorems, random variables, distributions
18.175: Lecture 2 Extension theorems, random variables, distributions Scott Sheffield MIT Outline Extension theorems Characterizing measures on R d Random variables Outline Extension theorems Characterizing
More informationQuick Tour of Basic Probability Theory and Linear Algebra
Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions
More informationRigorous approximation of invariant measures for IFS Joint work April 8, with 2016 S. 1 Galat / 21
Rigorous approximation of invariant measures for IFS Joint work with S. Galatolo e I. Nisoli Maurizio Monge maurizio.monge@im.ufrj.br Universidade Federal do Rio de Janeiro April 8, 2016 Rigorous approximation
More informationOn Extracting Common Random Bits From Correlated Sources on Large Alphabets
University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 3-2014 On Extracting Common Random Bits From Correlated Sources on Large Alphabets Siu On Chan Elchanan Mossel University
More informationLecture 5: January 30
CS71 Randomness & Computation Spring 018 Instructor: Alistair Sinclair Lecture 5: January 30 Disclaimer: These notes have not been subjected to the usual scrutiny accorded to formal publications. They
More informationWeak and strong moments of l r -norms of log-concave vectors
Weak and strong moments of l r -norms of log-concave vectors Rafał Latała based on the joint work with Marta Strzelecka) University of Warsaw Minneapolis, April 14 2015 Log-concave measures/vectors A measure
More informationNon-interactive Simulation of Joint Distributions: The Hirschfeld-Gebelein-Rényi Maximal Correlation and the Hypercontractivity Ribbon
Non-interactive Simulation of Joint Distributions: The Hirschfeld-Gebelein-Rényi Maximal Correlation and the Hypercontractivity Ribbon Sudeep Kamath and Venkat Anantharam EECS Department University of
More informationLetting p shows that {B t } t 0. Definition 0.5. For λ R let δ λ : A (V ) A (V ) be defined by. 1 = g (symmetric), and. 3. g
4 Contents.1 Lie group p variation results Suppose G, d) is a group equipped with a left invariant metric, i.e. Let a := d e, a), then d ca, cb) = d a, b) for all a, b, c G. d a, b) = d e, a 1 b ) = a
More informationcurvature, mixing, and entropic interpolation Simons Feb-2016 and CSE 599s Lecture 13
curvature, mixing, and entropic interpolation Simons Feb-2016 and CSE 599s Lecture 13 James R. Lee University of Washington Joint with Ronen Eldan (Weizmann) and Joseph Lehec (Paris-Dauphine) Markov chain
More informationDisjointness and Additivity
Midterm 2: Format Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationInput-queued switches: Scheduling algorithms for a crossbar switch. EE 384X Packet Switch Architectures 1
Input-queued switches: Scheduling algorithms for a crossbar switch EE 84X Packet Switch Architectures Overview Today s lecture - the input-buffered switch architecture - the head-of-line blocking phenomenon
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationLecture 14: October 22
CS294 Markov Chain Monte Carlo: Foundations & Applications Fall 2009 Lecture 14: October 22 Lecturer: Alistair Sinclair Scribes: Alistair Sinclair Disclaimer: These notes have not been subjected to the
More informationSzemerédi s regularity lemma revisited. Lewis Memorial Lecture March 14, Terence Tao (UCLA)
Szemerédi s regularity lemma revisited Lewis Memorial Lecture March 14, 2008 Terence Tao (UCLA) 1 Finding models of large dense graphs Suppose we are given a large dense graph G = (V, E), where V is a
More informationMidterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley
Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More information25.1 Markov Chain Monte Carlo (MCMC)
CS880: Approximations Algorithms Scribe: Dave Andrzejewski Lecturer: Shuchi Chawla Topic: Approx counting/sampling, MCMC methods Date: 4/4/07 The previous lecture showed that, for self-reducible problems,
More information215 Problem 1. (a) Define the total variation distance µ ν tv for probability distributions µ, ν on a finite set S. Show that
15 Problem 1. (a) Define the total variation distance µ ν tv for probability distributions µ, ν on a finite set S. Show that µ ν tv = (1/) x S µ(x) ν(x) = x S(µ(x) ν(x)) + where a + = max(a, 0). Show that
More informationLecture 35: December The fundamental statistical distances
36-705: Intermediate Statistics Fall 207 Lecturer: Siva Balakrishnan Lecture 35: December 4 Today we will discuss distances and metrics between distributions that are useful in statistics. I will be lose
More informationLecture 28: April 26
CS271 Randomness & Computation Spring 2018 Instructor: Alistair Sinclair Lecture 28: April 26 Disclaimer: These notes have not been subjected to the usual scrutiny accorded to formal publications. They
More informationNotes on Gaussian processes and majorizing measures
Notes on Gaussian processes and majorizing measures James R. Lee 1 Gaussian processes Consider a Gaussian process {X t } for some index set T. This is a collection of jointly Gaussian random variables,
More informationUncertainty Quantification in Computational Science
DTU 2010 - Lecture I Uncertainty Quantification in Computational Science Jan S Hesthaven Brown University Jan.Hesthaven@Brown.edu Objective of lectures The main objective of these lectures are To offer
More informationLecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.
Lecture 10 1 Ergodic decomposition of invariant measures Let T : (Ω, F) (Ω, F) be measurable, and let M denote the space of T -invariant probability measures on (Ω, F). Then M is a convex set, although
More informationThe expansion of random regular graphs
The expansion of random regular graphs David Ellis Introduction Our aim is now to show that for any d 3, almost all d-regular graphs on {1, 2,..., n} have edge-expansion ratio at least c d d (if nd is
More informationCSE 291: Fourier analysis Chapter 2: Social choice theory
CSE 91: Fourier analysis Chapter : Social choice theory 1 Basic definitions We can view a boolean function f : { 1, 1} n { 1, 1} as a means to aggregate votes in a -outcome election. Common examples are:
More informationThe coupling method - Simons Counting Complexity Bootcamp, 2016
The coupling method - Simons Counting Complexity Bootcamp, 2016 Nayantara Bhatnagar (University of Delaware) Ivona Bezáková (Rochester Institute of Technology) January 26, 2016 Techniques for bounding
More informationOn Reparametrization and the Gibbs Sampler
On Reparametrization and the Gibbs Sampler Jorge Carlos Román Department of Mathematics Vanderbilt University James P. Hobert Department of Statistics University of Florida March 2014 Brett Presnell Department
More informationLecture I: Asymptotics for large GUE random matrices
Lecture I: Asymptotics for large GUE random matrices Steen Thorbjørnsen, University of Aarhus andom Matrices Definition. Let (Ω, F, P) be a probability space and let n be a positive integer. Then a random
More information10-704: Information Processing and Learning Fall Lecture 9: Sept 28
10-704: Information Processing and Learning Fall 2016 Lecturer: Siheng Chen Lecture 9: Sept 28 Note: These notes are based on scribed notes from Spring15 offering of this course. LaTeX template courtesy
More informationRapid Introduction to Machine Learning/ Deep Learning
Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian
More informationFunctional Inequalities for Gaussian and Log-Concave Probability Measures
Functional Inequalities for Gaussian and Log-Concave Probability Measures Ewain Gwynne Adviser: Professor Elton Hsu Northwestern University A thesis submitted for a Bachelor s degree in Mathematics (with
More informationFourier analysis of boolean functions in quantum computation
Fourier analysis of boolean functions in quantum computation Ashley Montanaro Centre for Quantum Information and Foundations, Department of Applied Mathematics and Theoretical Physics, University of Cambridge
More informationCoupling AMS Short Course
Coupling AMS Short Course January 2010 Distance If µ and ν are two probability distributions on a set Ω, then the total variation distance between µ and ν is Example. Let Ω = {0, 1}, and set Then d TV
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More information1 Sequences of events and their limits
O.H. Probability II (MATH 2647 M15 1 Sequences of events and their limits 1.1 Monotone sequences of events Sequences of events arise naturally when a probabilistic experiment is repeated many times. For
More informationHigh-dimensional distributions with convexity properties
High-dimensional distributions with convexity properties Bo az Klartag Tel-Aviv University A conference in honor of Charles Fefferman, Princeton, May 2009 High-Dimensional Distributions We are concerned
More informationLecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016
Lecture 15: MCMC Sanjeev Arora Elad Hazan COS 402 Machine Learning and Artificial Intelligence Fall 2016 Course progress Learning from examples Definition + fundamental theorem of statistical learning,
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationMathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )
Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca October 22nd, 2014 E. Tanré (INRIA - Team Tosca) Mathematical
More informationLecture 7: Stochastic Dynamic Programing and Markov Processes
Lecture 7: Stochastic Dynamic Programing and Markov Processes Florian Scheuer References: SLP chapters 9, 10, 11; LS chapters 2 and 6 1 Examples 1.1 Neoclassical Growth Model with Stochastic Technology
More informationApplied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books.
Applied Analysis APPM 44: Final exam 1:3pm 4:pm, Dec. 14, 29. Closed books. Problem 1: 2p Set I = [, 1]. Prove that there is a continuous function u on I such that 1 ux 1 x sin ut 2 dt = cosx, x I. Define
More informationProbability & Computing
Probability & Computing Stochastic Process time t {X t t 2 T } state space Ω X t 2 state x 2 discrete time: T is countable T = {0,, 2,...} discrete space: Ω is finite or countably infinite X 0,X,X 2,...
More informationChapter 9: Basic of Hypercontractivity
Analysis of Boolean Functions Prof. Ryan O Donnell Chapter 9: Basic of Hypercontractivity Date: May 6, 2017 Student: Chi-Ning Chou Index Problem Progress 1 Exercise 9.3 (Tightness of Bonami Lemma) 2/2
More information1 Ways to Describe a Stochastic Process
purdue university cs 59000-nmc networks & matrix computations LECTURE NOTES David F. Gleich September 22, 2011 Scribe Notes: Debbie Perouli 1 Ways to Describe a Stochastic Process We will use the biased
More informationNotes on Probability
Notes on Probability Mark Schmidt January 7, 2017 1 Probabilites Consider an event A that may or may not happen. For example, if we roll a dice then we may or may not roll a 6. We use the notation p(a)
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued
Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and
More informationThe Green-Tao theorem and a relative Szemerédi theorem
The Green-Tao theorem and a relative Szemerédi theorem Yufei Zhao Massachusetts Institute of Technology Joint work with David Conlon (Oxford) and Jacob Fox (MIT) Simons Institute December 2013 Green Tao
More informationNote that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +
Random Walks: WEEK 2 Recurrence and transience Consider the event {X n = i for some n > 0} by which we mean {X = i}or{x 2 = i,x i}or{x 3 = i,x 2 i,x i},. Definition.. A state i S is recurrent if P(X n
More informationMGMT 69000: Topics in High-dimensional Data Analysis Falll 2016
MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016 Lecture 14: Information Theoretic Methods Lecturer: Jiaming Xu Scribe: Hilda Ibriga, Adarsh Barik, December 02, 2016 Outline f-divergence
More informationErgodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.
Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions
More informationFunctional Analysis, Math 7320 Lecture Notes from August taken by Yaofeng Su
Functional Analysis, Math 7320 Lecture Notes from August 30 2016 taken by Yaofeng Su 1 Essentials of Topology 1.1 Continuity Next we recall a stronger notion of continuity: 1.1.1 Definition. Let (X, d
More informationCharacterization of cutoff for reversible Markov chains
Characterization of cutoff for reversible Markov chains Yuval Peres Joint work with Riddhi Basu and Jonathan Hermon 23 Feb 2015 Joint work with Riddhi Basu and Jonathan Hermon Characterization of cutoff
More information2. The averaging process
2. The averaging process David Aldous July 10, 2012 The two models in lecture 1 has only rather trivial interaction. In this lecture I discuss a simple FMIE process with less trivial interaction. It is
More informationLecture Notes 3 Convergence (Chapter 5)
Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationApproximation & Complexity
Summer school on semidefinite optimization Approximation & Complexity David Steurer Cornell University Part 1 September 6, 2012 Overview Part 1 Unique Games Conjecture & Basic SDP Part 2 SDP Hierarchies:
More informationA Characterization of Locally Testable Affine-Invariant Properties via Decomposition Theorems
A Characterization of Locally Testable Affine-Invariant Properties via Decomposition Theorems Yuichi Yoshida National Institute of Informatics and Preferred Infrastructure, Inc June 1, 2014 Yuichi Yoshida
More informationInformation Theory and Statistics Lecture 3: Stationary ergodic processes
Information Theory and Statistics Lecture 3: Stationary ergodic processes Łukasz Dębowski ldebowsk@ipipan.waw.pl Ph. D. Programme 2013/2014 Measurable space Definition (measurable space) Measurable space
More informationExercises Measure Theoretic Probability
Exercises Measure Theoretic Probability 2002-2003 Week 1 1. Prove the folloing statements. (a) The intersection of an arbitrary family of d-systems is again a d- system. (b) The intersection of an arbitrary
More informationEE514A Information Theory I Fall 2013
EE514A Information Theory I Fall 2013 K. Mohan, Prof. J. Bilmes University of Washington, Seattle Department of Electrical Engineering Fall Quarter, 2013 http://j.ee.washington.edu/~bilmes/classes/ee514a_fall_2013/
More informationMath 456: Mathematical Modeling. Tuesday, March 6th, 2018
Math 456: Mathematical Modeling Tuesday, March 6th, 2018 Markov Chains: Exit distributions and the Strong Markov Property Tuesday, March 6th, 2018 Last time 1. Weighted graphs. 2. Existence of stationary
More informationLECTURE 3. Last time:
LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE Most of the material in this lecture is covered in [Bertsekas & Tsitsiklis] Sections 1.3-1.5
More informationEE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16
EE539R: Problem Set 4 Assigned: 3/08/6, Due: 07/09/6. Cover and Thomas: Problem 3.5 Sets defined by probabilities: Define the set C n (t = {x n : P X n(x n 2 nt } (a We have = P X n(x n P X n(x n 2 nt
More informationStochastic Problems. 1 Examples. 1.1 Neoclassical Growth Model with Stochastic Technology. 1.2 A Model of Job Search
Stochastic Problems References: SLP chapters 9, 10, 11; L&S chapters 2 and 6 1 Examples 1.1 Neoclassical Growth Model with Stochastic Technology Production function y = Af k where A is random Let A s t
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More information1 Math 241A-B Homework Problem List for F2015 and W2016
1 Math 241A-B Homework Problem List for F2015 W2016 1.1 Homework 1. Due Wednesday, October 7, 2015 Notation 1.1 Let U be any set, g be a positive function on U, Y be a normed space. For any f : U Y let
More informationA Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones
A Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones Abstract We present a regularity lemma for Boolean functions f : {, } n {, } based on noisy influence, a measure of how locally correlated
More informationMarkov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains
Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time
More informationSpectral properties of Markov operators in Markov chain Monte Carlo
Spectral properties of Markov operators in Markov chain Monte Carlo Qian Qin Advisor: James P. Hobert October 2017 1 Introduction Markov chain Monte Carlo (MCMC) is an indispensable tool in Bayesian statistics.
More informationMAA704, Perron-Frobenius theory and Markov chains.
November 19, 2013 Lecture overview Today we will look at: Permutation and graphs. Perron frobenius for non-negative. Stochastic, and their relation to theory. Hitting and hitting probabilities of chain.
More informationBALANCING GAUSSIAN VECTORS. 1. Introduction
BALANCING GAUSSIAN VECTORS KEVIN P. COSTELLO Abstract. Let x 1,... x n be independent normally distributed vectors on R d. We determine the distribution function of the minimum norm of the 2 n vectors
More information5 Birkhoff s Ergodic Theorem
5 Birkhoff s Ergodic Theorem Birkhoff s Ergodic Theorem extends the validity of Kolmogorov s strong law to the class of stationary sequences of random variables. Stationary sequences occur naturally even
More informationConsistency of the maximum likelihood estimator for general hidden Markov models
Consistency of the maximum likelihood estimator for general hidden Markov models Jimmy Olsson Centre for Mathematical Sciences Lund University Nordstat 2012 Umeå, Sweden Collaborators Hidden Markov models
More informationLecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016
Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 1 Entropy Since this course is about entropy maximization,
More informationAn exponential separation between quantum and classical one-way communication complexity
An exponential separation between quantum and classical one-way communication complexity Ashley Montanaro Centre for Quantum Information and Foundations, Department of Applied Mathematics and Theoretical
More informationRoth s Theorem on 3-term Arithmetic Progressions
Roth s Theorem on 3-term Arithmetic Progressions Mustazee Rahman 1 Introduction This article is a discussion about the proof of a classical theorem of Roth s regarding the existence of three term arithmetic
More informationLogarithmic Sobolev Inequalities
Logarithmic Sobolev Inequalities M. Ledoux Institut de Mathématiques de Toulouse, France logarithmic Sobolev inequalities what they are, some history analytic, geometric, optimal transportation proofs
More informationLecture Introduction. 2 Brief Recap of Lecture 10. CS-621 Theory Gems October 24, 2012
CS-62 Theory Gems October 24, 202 Lecture Lecturer: Aleksander Mądry Scribes: Carsten Moldenhauer and Robin Scheibler Introduction In Lecture 0, we introduced a fundamental object of spectral graph theory:
More informationInference for Stochastic Processes
Inference for Stochastic Processes Robert L. Wolpert Revised: June 19, 005 Introduction A stochastic process is a family {X t } of real-valued random variables, all defined on the same probability space
More informationµ X (A) = P ( X 1 (A) )
1 STOCHASTIC PROCESSES This appendix provides a very basic introduction to the language of probability theory and stochastic processes. We assume the reader is familiar with the general measure and integration
More informationLecture 10: Semi-Markov Type Processes
Lecture 1: Semi-Markov Type Processes 1. Semi-Markov processes (SMP) 1.1 Definition of SMP 1.2 Transition probabilities for SMP 1.3 Hitting times and semi-markov renewal equations 2. Processes with semi-markov
More information