ECON 2530b: International Finance

Size: px
Start display at page:

Download "ECON 2530b: International Finance"

Transcription

1 ECON 2530b: International Finance Compiled by Vu T. Chau 1 Non-neutrality of Nominal Exchange Rate 1.1 Mussa s Puzzle(1986) RER can be defined as the relative price of a country s consumption basket in terms of another country s consumption basket: or, in logs: RER f /d = NER f /d P d P f rer f /d = ner f /d + p where p is the log ratio of domestic price index to foreign price index. 1 An appreciation (depreciation) in the RER can therefore be caused by either (1) appreciation (depreciation) of the nominal exchange rate, or (2) inflation (deflation). Relative PPP: RER is constant, hence movements in NER must be offset by movement of P d /P f. Furthermore, exchange rate regimes (fixed or floating) should not matter. Empirical findings contradict these implications of relative PPP: Systemic differences in movements of RER under fixed and floating regime (RER is more volatile under floating NER regime). Under floating regime, RER co-moves closely with NER, hence contradicting the neutrality of NER. RER is very persistent with long half-lives. 1.2 Cavallo, Neiman, and Rigobon (2013) Law of One Price (LOP) holds well for thousands of goods in currency Union, but not well outside. LOP holds well for dollarized countries, but not among countries that peg to the dollar. Therefore, having the same currency is important in getting LOP! Variable markups matter! (IPIs). 1 This price index can be calculated using consumer price indices (CPIs), producer price indices (PPIs), or import price indices 1

2 2 Lebesgue Integration and Convergence Theorem Theorem 2.1 (Monotone Convergence Theorem). Let g, f 1, f 2,... be measurable functions such that g L 1, g f 1 a.e. and f n f a.e. Then: f n dµ f dµ Lemma 2.2 (Fatou s Lemma). Let g, f 1, f 2,... be measurable functions such that g L 1, f n g a.e. for n. Then: lim inf f n dµ lim inf f n dµ n n Theorem 2.3 (Lebesgue s Dominated Convergence Theorem). Let g, f 1, f 2,... be measurable functions such that g L 1, f n g n a.e. n, and f n f a.e. Then, we have f n dµ = f dµ and lim n lim n f f n = 0 If µ is a probability measure, then µ f is a probability measure too, and called the distribution of f. Proposition 2.4 (Push-forward measure 2 ). Let (Ω, F, µ) be a measure space, and (Ω, F ) a measurable space. f : (Ω, F) (Ω, F ) a measurable function. Then: µ f (B) = µ[ f 1 (B)] defines a measure on (Ω, F ), and g f dµ = gdµ Ω Ω f Definition (Expectation, Variance, and Covariance). We define: EX = Ω XdP for X L1. Var(X) = Ω (X EX)2 dp = E[X 2 ] (EX) 2 Cov(X, Y) = E[(X EX)(Y EY)] = E[XY] (EX)(EY) So far, we are saying that given a measure space and a random variable, we can characterize the distribution of that variable by the cdf F X (x) = (P X 1 )[X x]. Now we define the inverse function of F X : Definition (Right-quantile function). Define q X : (0, 1) R by q X (u) = sup{x : F X (x) u}. We note that q X = d X. So far we have linked the relationship between pointwise convergence and convergence in L 1 using the two convergence theorems. The following section explores more types of convergence, and the notion of uniform integrable. 3 Types of Convergence and Uniform Integrability Definition. Let X 1, X 2,... be randome variables on a probability space (Ω, F, P). X n X a.s. N F such that P[N] = 0 and lim n X n (Ω) = X(Ω) for ω Ω \ N. 3 X n L p X X n X p 0 as n. 4. X n P X P[ Xn X ] > ɛ 0 as n for all ɛ > 0. X n D X E[ f (Xn )] E[ f (X)] as n for all bounded continuous functions f : R R. Definition (Uniform Integrable). A family of random variables (X i ) i I on a probability space is uniformly integrable if sup i I X i >c X i dp 0 as c. 2 Take g(x) = x, we end up with EX = Ω XdP = R xd(p X 1 ) 3 Almost surely convergence is pointwise convergence at all points except for a set of measure zero. 4 Recall that f p = (E[ f p ]) 1/p. Note that when p = 1 we get convergence in L 1, which we discussed extensively earlier. 2

3 It turns out there is a very simple condition for uniform integrability: if the sequence of random variables is bounded by an integrable RV X. This is stated clearly below: Lemma 3.1 (Conditions for Uniform Integrability). If (X i ) i I is a family of random variables on a probability space (Ω, F, P) such that X i X for some X L 1 (Ω, F, P) and i I, then (X i ) i I is uniformly integrable. It follows that finitely many random variables X 1, X 2,..., X n L 1 (Ω, F, P) are uniformly integrable. Now we turn to the necessary and sufficient condition for uniform integrability. In simple words, a family of RVs is uniformly integrable if (1) each RV in the family is L 1 bounded, and (2) the L 1 norm of each RV vanishes on any small enough set. Particularly: Theorem 3.2 (Necessary and Sufficient Conditions for Uniform Integrability). Let (X i ) i I be a family of RVs. This family is uniformly integrable if: (L 1 boundedness) sup i I X i L 1 < ; and, (vanishes with small domain) For any ɛ > 0, δ > 0 s.t. E[ X i 1 A ] < ɛ for all i and set A F with P[A] < δ. Below is a major result: Theorem 3.3 (Uniform Integrability + Convergence in Probability = Convergence in L 1 ). Let X 1, X 2,... P be RVs on (Ω, F, P), X n X, and {Xn } n 1 uniformly integrable. Then X is integrable (EX < ), X n X L 1 0, and, as a result, EX n EX. Recall that we have a sufficient condition for uniform integrability before, which is (1) (X i ) has to be L 1 bounded, and (2) a really long and complicated condition. The following lemma and its corollary helps simplify the second condition: Lemma 3.4 (de la Vallee - Poussin). A family of RVs (X i ) i I on (Ω, F, P) is uniformly integrable if and only if there exists ϕ : R + R + such that lim x ϕ(x) x = and sup i I E[ϕ( X i )] <. ϕ can always be chosen to be convex and non-decreasing. Corollary 3.5 (L p bounded (p > 1) implies uniform integrability). A family of random variables (X i ) i I defined on (Ω, F, P) is uniformly integrable if sup i I X i L p < for some p (1, ]. Using Jensen s inequality 5, we can prove the following result: Lemma 3.6. Let X be a RV. Then X L p X L q for all 1 p q. 4 Weak Law of Large Numbers The law of large number states that the sample mean of n uncorrelated variables with the same mean and finite second moment will converge to the population mean in L 2 and in probability. Theorem 4.1 (Weak Law of Large Number). Let X 1, X 2,..., X n be uncorrelated RVs in L 2 s.t. E[X n ] = m n, and sup n E[Xn] 2 <. Then n 1 n X i m in L 2 (and hence, in probability). i=1 5 Distributions and Densities 5.1 Discrete distributions Definition (Dirac measure). Let Ω be a set, ω Ω. The Dirac measure δ ω : 2 Ω {0, 1} is given by: { 1 if ω A, δ ω 0 else. 5 Let X be a RV and ϕ : R R a convex function such that X, ϕ L 1. Then: E[ϕ(X)] ϕ(e[x]) 3

4 Let µ be a measure on R d. Definition. µ is called discrete if µ = n 1 p n δ xn for {x n } n 1 and p 1, p 2, Some examples of discrete distributions on R: (1) Discrete uniform distribution: µ(n) = N 1, n = 1, 2,..., N. (2) Bernoulli: µ(0) = 1 p, µ(1) = p for p [0, 1]. (3) Binomial: µ(n) = ( N n )pn (1 p) N n for n = 0, 1,..., N. (4) Poisson: µ(n) = e λ λn n!, n = 0, 1,... (5) Geometric: µ(n) = (1 p) n p, n = 1, 2, Continuous Distributions Definition. µ is called continuous if µ(a) is continuous in R d, for A = (a 1, b 1 ] (a2, b2]... (a d, b d ]. Definition. µ is called absolutely continuous if there exists a Borel measurable function f : R d R + such that: µ ((a 1, b 1 ]... (a d, b d ]) = b1 a 1... bd a d f (y 1, y 2,..., y d )dy d...dy 1 f is called a density and µ[a] = A f (x)dx A B(R) d. Also, for any g measurable, we have Examples: Rd gdµ(x) = (1) Normal distribution (or, Gaussian distribution): f (x) = g(x) f (x)dx Rd 1 (x µ)2 exp{ 2πσ 2 2σ 2 } (2) Uniform on [a, b]: f (x) = b a 1 for x [a, b] and zero outside. (3) Exponential: f (x) = λe λx, for x 0, λ > 0. 6 (memoryless) (4) Bilateral exponential: f (x) = 1 2 λe λ x for x R, λ > 0. (5) Cauchy distribution: It is the distribution of a ratio of two standard normal variables. The density is: θ f (x) = π(x 2 + θ 2 ) This distribution is pathological in the sense that it has no moment (everything, mean, variance, etc. is undefined). 6 Characteristic Function Definition (Characteristic Function). Let X be a d dimensional RV with F(x 1,..., x d ) = P(X 1 x 1,..., X d x d ). The characteristic function ψ X : R d C of X is given by: [ ] ψ X (u) = E e iut X = eiutx df(x) R d Some properties of characteristic function: Lemma 6.1. Let X be a d dimensional RV. Then: 6 This is the continuous analogue of geometric distribution. It describes the time between events in a Poisson process. 4

5 (1) ψ X (u) 1 (2) ψ X ( u) = ψ X (u) = ψ X (u) (3) If ψ X is real-valued, then X and X has the same distribution. (4) ψ X (u) is uniformly continuous in R d. 7 (5) If Y = ax + b, then the characteristic function of Y is ψ Y = e iutb ψ X (au). 8 (6) The characteristic function of Z = X + Y where X, Y are independent RVs is ψ Z (u) = ψ X (u)ψ Y (u). 6.1 Common Characteristic Function (1) Dirac delta measure P[X = x] = 1: ψ X (u) = exp(iu T x). (2) Bernoulli: P[X = 0] = 1 p, P[X = 1] = p. Then: ψ X (u) = E[e iux ] = (1 p)e iu 0 + pe iu 1 = (1 p) + pe iu. (3) Poisson distribution ( f (n) = e λ λn n! ):9 ψ X (u) = n=0 f (n)eiun = e λ [λe iu ] n n=0 n! = e λ(1 eiu ) (4) Gaussian distribution X N (µ, σ 2 ): ψ X (u) = exp(iuµ 1 2 σ2 u 2 ). Standard normal (µ = 0, σ = 1), then µ X (u) = e 1 2 u Inversion Formula One way to check if a probability measure admits a density is to see if the characteristic function is integrable. If that s the case, the theorem below helps uncover the original density: Theorem 6.2 (Inversion Formula). Let ψ X (u) = e iux dµ such that R ψ X(u) du <, then a density f (x) exists and satisfies f (x) = 1 e iux ψ(u)du 2π R 6.3 Levy s Continuity Theorem We have that the distribution of a random variable X is uniquely determined by the characteristic function ψ x due to the inversion vormula. This applies to a vector of RVs as well. Theorem 6.3 (Levy s Continuity Theorem). Let X n be RVs defined on probability spaces (Ω n, F n, P n ) with characteristic function ψ n (u), n 1. Then convergence in distribution of (X n ) n 1 implies pointwise convergence in the characteristic functions, and vice versa. More rigorously: (1) If X n d X for some RV X, then ψ n (u) ψ(u) for all u R. (2) If ψ n (u) ψ(u) for all u R, and ψ(u) is continuous at u = 0, then ψ is the characteristic function of some RV X, and X n d X. 6.4 Central Limit Theorem Theorem 6.4 (CLT). Let X 1, X 2,..., X n be RVs on (Ω, F, P) such that E[X 2 ] < and σ n = Var(X n ) > 0 1 for n. Then, n i=1 n X i E[X i ] σ(x i ) d Z where Z N (0, 1). 7 Quick proof: Fix ɛ > 0. ψ X (u + t) ψ X (u) = E[exp(iu T X)(exp(it T X) 1)] E[exp(it T X) 1] 0 as t 0 8 ψ Y (u) = E[exp(iu T Y)] = E[[exp(i(au) T X)]] = exp(iu T b)ψ X (au) 9 Recall that e x = n=0 xn n! 5

6 7 Dynamical System ( ) Definition. A random vector X R d is normal (or Gaussian) if and only ψ X (u) = exp iu T µ 1 2 ut Cu for some µ R d and C symmetric positive semidefinite. Notation: X N d (µ, C). Note that when d = 1 we get the characteristic function ψ X (u) = exp(iuµ 1 2 σ2 u 2 ) of a normal distribution mean µ and variance σ 2. Importantly, if C is not invertible, then the density doesn t exist. Proposition 7.1. Let µ R d and C R d d be a symmetric positive semidedinite matrix. Then: (1) There exists A such that A 2 = C If Z N d (0, I d ), then X = µ + AZ N d (µ, C). (2) The components of X are independent if and only if they are uncorrelated. 12 (3) If C is invertible, then X has a density: { } 1 1 f (x) = (2π) d/2 det(c) exp 2 (X µ)t C 1 (X µ) If C is not-invertible, then A is not invertible, and AR d is a strict subspace of R d, and X cannot have a density. (4) For all V R k and M R k d, Y = V + MX N d (V + Mµ, MCM T ) Proposition 7.2. A d dimensional random vector X is normal if and only if v T X is normal for all v R d. 7.1 Gaussian Process A Gaussian process is a process in which any finite set samples in time (or space) is normally distributed. Formally: Definition (Gaussian Process). Let I be a non-empty set. A family (X i ) i I of RVs on a probability space is called a Gaussian process if (X i1, X i2,..., X id ) is a finite dimensional Gaussian for all d and i 1, i 2,..., i d I. As we move to a (possibly) continuous index set I (for example, time [0, T]), we need to extend the notion of covariance matrix C from the last session. Definition (Symmetric, positive semidefiniteness of function). A function C : I 2 R is symmetric if C(i, j) = C(j, i) for all i, j I. It is positive semidefinite if [C(i k, i l )] d k,l=1 is positive semidefinite for all d and (i 1,..., i d ) I d. For a Gaussian process (X i ) i I, denote µ(i) = E[X i ] and C X (i, j) = cov(x i, X j ) for i, j I. Then, C X is symmetric and positive semidefinite by definition (since any finite sample follows multivariate normal, its covariance matrix must be positive semidefinite). However, provided a mean function µ X (i) and covariance function C X (i, j), a Gaussian process with that mean and covariance function is guaranteed to exist: Theorem 7.3 (Existence of Gaussian process). Let I be a non-empty set, µ : I R an arbitrary function, and C : I 2 R an arbitrary symmetric and positive semidefinite function. Then there exists a Gaussian process (X i ) i I with µ X = µ and C X = C. Examples of Gaussian processes: 10 Linear algebra note: C is symmetric, then the followings are equivalent: (1) C is positive semidefinite; (2) C has all nonnegative eigenvalues; (3) all principal minors of C are nonnegative; and (4) There exists A such that A T A = C. 11 Connection between definiteness and principal minors: Let D k and k denotes leading principal minor and principal minor of order k respectively. Positive definiteness = D k > 0 for all D k ; negative definiteness = ( 1) k D k > 0 for all D k ; positive semidefiniteness = Delta k 0 for all k ; negative semidefiniteness = ( 1) k k 0 for all k. 12 Generally, zero correlation does not imply independence. However, if two variables are jointly normal, then correlation does imply independence. Note however that two variables which are separately normal may still not have this property, since there is no guarantee that the joint distribution is normal. 6

7 (1) White noise: Gaussian process with µ X (i) = 0 for i I and C X (i, j) = 1 i=j for i, j I. 13 (2) Brownian motion: µ X (i) = 0 for i, and C X (i, j) = (i j) for i, j I. Properties of a Brownian motion: E[X 2 0 ] = 0 0 = 0 X 0 = 0 a.s. Increments are normally distributed: For t > s, X t X s = ( 1 1 ) ( ) X s N (0, t s). X t ( Xs X t ) (( ) ( )) 0 s s N 2,, so 0 s t Increments are independent: For s < t < u < v:cov(x v X u, X t X s ) = cov(x v, X t ) + cov(x u, X s ) cov(x v, X s ) cov(x u, X t ) = t + s s t = 0. Since increments are distributed jointly Gaussian, this also implies they are independent. 8 Martingales 8.1 Conditional Expectation Definition (Absolute Continuous Measure). Let µ 1, µ 2 be two measures on (Ω, F). µ 2 is said to be absolutely continuous wrt µ 1 if µ 1 [A] = 0 implies µ 2 [A] = 0 for all A F. 14 Notation: µ 1 >> µ 2. Note that µ 2 = A f dµ 1 = E[1 A f ] defines a measure that is absolutely continuous wrt µ 1. This means that given a function, then we can generate a corresponding absolutely continuous measure. The other direction is guaranteed by the Radon-Nikodyn theorem, which posits that given a measure, we can find such a function f : Theorem 8.1 (Radon-Nikodym). Let µ 1, µ 2 be 2 measures on (Ω, F) such that µ 2 << µ 1 and µ 1 is σ finite. Then, there exists f : Ω R + { } such that µ 2 [A] = A f dµ 1 = E[1 A f ] for all A F. We use the remark above and the R-N theorem to guarantee the existence of a conditional expectation which is defined below: Definition (Conditional Expectation). Let (Ω, F, P) be a probability space, and G a sub-algebra of F. The conditional expectation of X L 1 (Ω, F, P) given G is a RV Y which satisfies: Y is G measurable, Y L 1 (Ω, G, P). E[1 A Y] = E[1 A X] for all A G. Notation: Y = E[X G]. We also define expectation wrt another RV: E[X Z] = E[X σ(z)] Proposition 8.2. Such a conditional expectation always exists uniquely (up to a.e. equal). The idea is that: (1) given F, P, and X, we can generate another measure P 2 = E[1 A f ] << P for A G (and P 2 [A] = 0 for A G). This measure, therefore, is limited onto G only. Then, applies Radon-Nikodym on (Ω, G) guarantees the existence of Y L 1 (Ω, G) such that P 2 [A] = E[1 A Y] for all A G. It follows that E[1 A Y] = E[1 A X] for all A G. Properties of conditional expectation: Proposition 8.3 (Properties of Conditional Expectation). We have: (1) (monotonicity) E[X G] E[Y G] if X Y a.e. (2) (Pulling out what s known) E[XY G] = YE[X G] if Y is G measurable. 13 This means that the variance of the noise at each point in time is var(x i ) = 1, and the noises are independent: cov(x i, X j ) = 0 (note: they are distributed normally, so zero correlation implies independence). 14 We can understand that µ 2 is a more narrowly defined measure, and the expectation on mu 2 will be the conditional expectation viewed from µ 1. 7

8 (3) (Irrelevant information) E[X G] = E[X] if X is independent of G. (4) (Jensen s Inequality) E[ϕ(X) G] ϕ(e[x G]) for ϕ : R R, ϕ L 1 (Ω, F, P) convex. (5) (Conditional MCT) If Y X 1 X 2... for Y, X 1, X 2,... L 1 (Ω, F, P) then lim n E[X n G] = E[lim n X n G] a.s. (6) (Conditional Fatou) If X n Y for some Y L 1 (Ω, F, P), then E[lim inf n X n G] lim inf n E[X n G]. (7) (Conditional DCT) If X n Y for some Y L 1 (Ω, F, P), then lim n E[X n G] = E[lim n X n G]. 8.2 Martingales Definition. (X n ) n 0 is a martingale wrt (F n ) n 0 if X n L 1 (Ω, F n, P) and E[X n+1 F n ] = X n. Definition. (X n ) n 0 is a supermartingale wrt (F n ) n 0 if X n L 1 (Ω, F n, P) and E[X n+1 F n ] X n. Definition. (X n ) n 0 is a submartingale wrt (F n ) n 0 if X n L 1 (Ω, F n, P) and E[X n+1 F n ] X n. Proposition 8.4. X n = E[X F n ] defines a uniformly integrable martingale. Definition (Martingale Transform). Let (X n ) n 0 and (V n ) n 0 { be 2 stochastic processes on (Ω, F, P). Define i=1 n X n = X n X n 1 for n 1. Define a new process (V X) n = V i X i if n 1 0 if n = 0. A process (V X) n such as above is called a Martingale transform. If (X n ) is a martingale, and (V n ) is predictable, then the transform is also a martingale. More rigorously: Theorem 8.5. Let (X n ) n 0 is a martingale wrt (F n ) n 0, (V n ) n 0 is predictable wrt (F n ) n 0. If (V X) n is L 1 bounded (i.e. sup n E (V X) n < ), then (V X n ) n 0 is a (F n ) martingale. Definition (Stopping Time 15 ). A stopping time wrt (F n ) n 0 on (Ω, F, P) is a RV τ : Ω N {+ } such that {τ = n} = {ω Ω : τ(ω) = n} F n Definition (Stopping Algebra). Given a stopping time τ, the algebra of that stopping time is the collection of events such that for each event, the decision to stop for each state of that event depends solely on the information available up to then: (F τ ) = {A F : A {τ n} F n }. It is easy to verify that this is a σ algebra. Lemma 8.6. If τ, σ are stopping times, then so are τ + σ, τ σ, τ σ. Also, F τ σ = F τ F σ. If τ σ, then F τ F σ. Below we define X τ n as a stopped process: it takes value of X n as normal until the stopping time, then stay constant thereafter. This process will be shown to be a martingale also. Lemma 8.7. (X n ) n 0 a stochastic process adapted to (F n ) n 0, and τ is a (F n ) stopping time. Then: the random variable X τ 1 {τ< } is F τ measurable. X τ n = X n τ is (F n ) adapted. Corollary 8.8. Let X n be (F n )-martingale, τ a (F n ) stopping time. (F n ) n 0. In particular, if τ is bounded, then E[X τ ] = E[X 0 ]. Then (X τ n) n 0 is a martingale wrt This implies that the expected value of any stopped process under a bounded stopping time is the same as the expected value of the process at time 0. All this result boils down to a major theorem: Doob s Optional Stopping theorem. This theorem posits that given a martingale process, a bounded stopping time in which the decision to stop in every state depends only the information up to that time. Then, that stopping time is optional in the sense that it will not alter the expected value at stopping time. More formally: 15 We can think of τ(ω) as a stopping decision: the decision to stop at any time τ is based solely on the information up to τ. This definition is equivalent to {τ n} F n for all n. 8

9 Theorem 8.9 (Doob s Optional Stopping). Let (X n ) n 0 be a (F n ) martingale, and σ < τ N two bounded stopping times wrt (F n ) n 0. Then E[X τ F σ ] = X σ. Taking σ(ω) = 0 for ω, we have E[X τ ] = X 0. Theorem 8.10 (Doob s Decomposition Theorem). Let (X n ) n 0 be a sub-martingale wrt (F n ) n 0. Then there exists a unique decomposition X n = M n + A n in which (M n ) n 0 is a martingale and (A n ) n 0 is a non-decreasing predictable process. Before, we have that transforming a martingale using a predictable process yields another martingale. We now transform submartingale: Corollary 8.11 (Submartingale Transformation). Let (X n ) n 0 be a submartingale, (V n ) n 0 be a nonnegative predictable process wrt (F n ) n 0, and E[ (V X) n ] < for n. Then (V X n ) n 0 is a submartingale. Proposition Let (X n ) n 0 be a submartingale wrt (F n ) n 0, and ϕ : R R convex, ϕ(x n ) L 1 for all n. If X n is a martingale or ϕ is non-decreasing, then ϕ(x n ) is a martingale. Let (X n ) n 0 be a submartingale wrt (F n ) n 0 and an interval [a, b] Let β n (a, b) denotes the number of times (X n ) n crosses b from below before n. It is easy to show that the time in which (X n ) n crosses from below is a stopping time. We also have an upper bound on the β n (a, b): Theorem 8.13 (Upcrossing Inequality). If (X n ) n 0 is a submartingale, then E[β n (a, b)] E[(X n a) + ] b a. This upcrossing inequality is used to prove the following convergence theorem of sub-martingale: Theorem 8.14 (Sub/super-martingale Convergence). (X n ) n 0 is L 1 bounded submartingale. Then there exists X such that X n X a.s., and X L 1 sup n X n L 1 Similarly, an L 1 bounded supermartingale converges almost surely. Corollary If (X n ) n 0 is a uniformly integrable submartingale, then X n X a.s. and in L 1. More over, E[X F n ] X n for all n. Similarly, a uniformly integrable supermartingale converges a.s. and in L 1. Corollary A submartingale (X n ) n 0 that is bounded above converges a.s. Similarly, a supermartingale that is bounded from below converges a.s. Corollary Similarly, an L p bounded super/sub-martingale (p > 1) converges a.s. and in L Corollary 8.18 (Convergence of Conditional Expectation Martingale). Let X L 1 (Ω, F, P) and (F n ) n 0 a filtration. Denote X n = E[X F n ], and X = E[X F ] where F = n F n = σ( n F n ). Then X n X a.s. and in L 1. Theorem 8.19 (Reverse Martingale Convergence). Let G 0 G 1... be a decreasing sequence of σ algebras on a probability space (Ω, F, P) and X L 1 (Ω, F, P). Then E[X G n ] E[X G] a.s. and in L 1, where G = n 0 G n. The Reverse Martingale Convergence theorem is useful to prove the strong law of large number. Lemma Let X 1, X 2,... be i.i.d. RV in L 1 and define S n n i=1 X i. Then E[X i S n ] = S n n. Theorem 8.21 (STRONG LAW OR LARGE NUMBER). Let X 1, X 2,... be i.i.d. RV in L 1 (Ω, F, P). Then 1 n n i=1 X i E[X i ] a.s. and in L 1. Theorem 8.22 (Doob s Maximal Inequality for Probabilities). Let (X n ) n 0 be a sub-martingale and λ > 0. Then: λp[max k n X k λ] E[X n 1 {maxk n X k λ}] E[X + n ] Corollary If (X n ) n 0 is a martingale, then P[max k n X k λ] E X n p λ p. Theorem 8.24 (Doob s L p maximal inequality). Let (X n ) n 0 be a non-negative submartingale. max k n X k L p p p 1 X n L p and max k n X k L 1 e 1 e (1 + X n log + X n L 1). 16 This is straightforward, because L p bounded for p > 1 implies uniform integrability. Then: 9

10 9 Stochastic Kernel and Regular Conditional Probabilities Definition. Let (Ω, F) and (Ω, F ) be two measurable spaces. A mapping K : Ω Ω R + {± } is a stochastic kernel from (Ω, F) to (Ω, F ) if: ω K(ω, A) is F measurable for A F. A K(ω, A is a measure on (Ω, F ) for all ω Ω. 10 Markov Chains Let I be a countable state space. A vector (λ i ) i I is a distribution if λ i 0 and i λ i = 1. A matrix (P ij ) i,j I is called stochastic if all its rows are distributions. Definition. A stochastic process (X n ) n 0 on a probability space (Ω, F, P) is a Markov chain with initial distribution λ and transition matrix p if: (i) P[X 0 = i] = λ i (ii) P[X n+1 = i n+1 X n = i n,..., X 0 = i 0 ] = p in i n+1 Notation: (X n ) n 0 is Markov(λ, p). Theorem 10.1 (Equivalent definition of Markov Chain). A discrete time stochastic process (X n ) n 0 is Markov(λ, p) if and only if P[X n+1 = i n+1 X n = i n,..., X 0 = i 0 ] = λ i0 p i0 i 1...p in 1 i n for all n 0 and i 0, i 1,..., i n I. Theorem 10.2 (Markov Property). Let (X n ) n 0 be the Markov(λ, p). If P[X m = i] > 0, then conditional on X m = i, (X m+n ) n 0 is Markov(δ i, p) and is independent of (X 0,..., X m ) Note that the state space I decomposes into communicating classes. A communicating class C I is called closed if i j for i C implies j C. i I is called absorbing if {i} is a closed communicating class. A Markov chain or transition matrix is called irreducible if I is a communicating class. Definition. Let A I. Let H A = inf{n : X n A} be the hitting time of the subset A. h A = P i [H A < ] the probability that A will be reached in finite time, and k A = E i [H A ] the expected hitting time given that we start at i. Theorem The vector of hitting probabilities (h A i ) i I is the minimal non-negative solution to the system { 1 if i A j I p ij h A j if i A Theorem The vector of hitting time expectation (k A i ) i I is the minimal non-negative solution to the system of linear equations: { 1 if 0 A 1 + j I p ij k A j if i A Theorem 10.5 (Strong Markov Property). Let (X n ) n 0 be Markov(λ, p) and a stopping time τ wrt (F X n ) n 0. Let i I be such that P[τ < and X τ = i] > 0. Then, conditional on τ < and X τ = i, (X τ+n ) is Markov(δ i, p) and independent of F X τ. Definition (Recurrent and Transient State). A state is recurrent if P i [X n = i for infinitely many n] = 1. A state is transient if P i [X n = i for infinitely many n = 0. Define the hitting times τi 0 = 0, τi 1 = τ i = inf{n 1 : X n = i}, τi 2 = inf{n > τi 1 : X n = i}, etc. Also, define the time difference between two hittings Si k = τi k τ k 1 i if τ k 1 i < and 0 otherwise. Lemma For k = 2, 3,..., conditional on τ k 1 i ] = P i [τ i = n]. <, S k i is independent of F X τ k 1 i and P[S k i = n τ k 1 i < 10

11 This lemma is a direct consequence of the Strong Markov Property. Define V i = n 0 1 Xn =i (number of visits to i), and f i = P i [τ i < ]. Lemma E i [V i ] = n 0 p (n) ii. Also, P i [V i > k] = fi k, for k = 1, 2,... Theorem 10.8 (Determining Recurrent or Transient). If P i [τ i < ] = 1, then i is recurrent and n 0 p n ii =. If P i [τ i < ] < 1, then i is transient and n 0 p n ii <. In particular, every state is either transient or recurrent. Theorem The states of a given communicating class are either all recurrent or transient. Therefore, one can speak of a recurrent/transient class. Theorem Every recurrent class is closed. Theorem Every finite closed class is recurrent. Theorem If p is an irreducible and recurrent. Then, for every initial distribution λ, P[τ j < ] = 1 for all j I Equilibrium Distribution Definition. A measure λ on I is said to be invariant wrt a transition matrix p if λp = λ. Theorem If (X n ) n 0 is Markov(λ, p) and λ is invariant wrt p, then for every m 1, (X m+n ) n 0 is again Markov(λ, p). Theorem Let I be finite. Assume there exists i I such that p n ij Π j as n for all j. Then (Π j ) j I is an invariant distribution. Definition. For i, k I, set γ k i E k [ τ k 1 n=0 1 X n =i] (starting at k, number of visits to i before visiting k again). Theorem Let p be an irreducible recurrent transition matrix. Then: (1) γ k k = 1. (2) (γ k i ) i I solves γ k p = γ k (3) 0 < γ k i < for all i I. Theorem Let p be an irreducible transition matrix and λ an invariant measure such that λ k = 1. Then λ γ k. In addition, if p is recurrent, then λ = γ k. 11

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal

More information

1. Stochastic Processes and filtrations

1. Stochastic Processes and filtrations 1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S

More information

Math 6810 (Probability) Fall Lecture notes

Math 6810 (Probability) Fall Lecture notes Math 6810 (Probability) Fall 2012 Lecture notes Pieter Allaart University of North Texas September 23, 2012 2 Text: Introduction to Stochastic Calculus with Applications, by Fima C. Klebaner (3rd edition),

More information

A D VA N C E D P R O B A B I L - I T Y

A D VA N C E D P R O B A B I L - I T Y A N D R E W T U L L O C H A D VA N C E D P R O B A B I L - I T Y T R I N I T Y C O L L E G E T H E U N I V E R S I T Y O F C A M B R I D G E Contents 1 Conditional Expectation 5 1.1 Discrete Case 6 1.2

More information

Part II Probability and Measure

Part II Probability and Measure Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

Probability Theory. Richard F. Bass

Probability Theory. Richard F. Bass Probability Theory Richard F. Bass ii c Copyright 2014 Richard F. Bass Contents 1 Basic notions 1 1.1 A few definitions from measure theory............. 1 1.2 Definitions............................. 2

More information

Useful Probability Theorems

Useful Probability Theorems Useful Probability Theorems Shiu-Tang Li Finished: March 23, 2013 Last updated: November 2, 2013 1 Convergence in distribution Theorem 1.1. TFAE: (i) µ n µ, µ n, µ are probability measures. (ii) F n (x)

More information

18.175: Lecture 3 Integration

18.175: Lecture 3 Integration 18.175: Lecture 3 Scott Sheffield MIT Outline Outline Recall definitions Probability space is triple (Ω, F, P) where Ω is sample space, F is set of events (the σ-algebra) and P : F [0, 1] is the probability

More information

Selected Exercises on Expectations and Some Probability Inequalities

Selected Exercises on Expectations and Some Probability Inequalities Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ

More information

Advanced Probability

Advanced Probability Advanced Probability Perla Sousi October 10, 2011 Contents 1 Conditional expectation 1 1.1 Discrete case.................................. 3 1.2 Existence and uniqueness............................ 3 1

More information

P (A G) dp G P (A G)

P (A G) dp G P (A G) First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume

More information

Lecture 6 Basic Probability

Lecture 6 Basic Probability Lecture 6: Basic Probability 1 of 17 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 6 Basic Probability Probability spaces A mathematical setup behind a probabilistic

More information

Brownian Motion and Conditional Probability

Brownian Motion and Conditional Probability Math 561: Theory of Probability (Spring 2018) Week 10 Brownian Motion and Conditional Probability 10.1 Standard Brownian Motion (SBM) Brownian motion is a stochastic process with both practical and theoretical

More information

Lecture 12. F o s, (1.1) F t := s>t

Lecture 12. F o s, (1.1) F t := s>t Lecture 12 1 Brownian motion: the Markov property Let C := C(0, ), R) be the space of continuous functions mapping from 0, ) to R, in which a Brownian motion (B t ) t 0 almost surely takes its value. Let

More information

Stochastic Processes. Winter Term Paolo Di Tella Technische Universität Dresden Institut für Stochastik

Stochastic Processes. Winter Term Paolo Di Tella Technische Universität Dresden Institut für Stochastik Stochastic Processes Winter Term 2016-2017 Paolo Di Tella Technische Universität Dresden Institut für Stochastik Contents 1 Preliminaries 5 1.1 Uniform integrability.............................. 5 1.2

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information

6.1 Moment Generating and Characteristic Functions

6.1 Moment Generating and Characteristic Functions Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,

More information

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have 362 Problem Hints and Solutions sup g n (ω, t) g(ω, t) sup g(ω, s) g(ω, t) µ n (ω). t T s,t: s t 1/n By the uniform continuity of t g(ω, t) on [, T], one has for each ω that µ n (ω) as n. Two applications

More information

Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales

Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales Prakash Balachandran Department of Mathematics Duke University April 2, 2008 1 Review of Discrete-Time

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Probability Theory II. Spring 2016 Peter Orbanz

Probability Theory II. Spring 2016 Peter Orbanz Probability Theory II Spring 2016 Peter Orbanz Contents Chapter 1. Martingales 1 1.1. Martingales indexed by partially ordered sets 1 1.2. Martingales from adapted processes 4 1.3. Stopping times and

More information

Stochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes

Stochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes BMS Basic Course Stochastic Processes II/ Wahrscheinlichkeitstheorie III Michael Scheutzow Lecture Notes Technische Universität Berlin Sommersemester 218 preliminary version October 12th 218 Contents

More information

Random Process Lecture 1. Fundamentals of Probability

Random Process Lecture 1. Fundamentals of Probability Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus

More information

4 Expectation & the Lebesgue Theorems

4 Expectation & the Lebesgue Theorems STA 205: Probability & Measure Theory Robert L. Wolpert 4 Expectation & the Lebesgue Theorems Let X and {X n : n N} be random variables on a probability space (Ω,F,P). If X n (ω) X(ω) for each ω Ω, does

More information

Exercises Measure Theoretic Probability

Exercises Measure Theoretic Probability Exercises Measure Theoretic Probability 2002-2003 Week 1 1. Prove the folloing statements. (a) The intersection of an arbitrary family of d-systems is again a d- system. (b) The intersection of an arbitrary

More information

Probability and Measure

Probability and Measure Probability and Measure Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Convergence of Random Variables 1. Convergence Concepts 1.1. Convergence of Real

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 9 10/2/2013. Conditional expectations, filtration and martingales

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 9 10/2/2013. Conditional expectations, filtration and martingales MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 9 10/2/2013 Conditional expectations, filtration and martingales Content. 1. Conditional expectations 2. Martingales, sub-martingales

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1 Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be

More information

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M. Lecture 10 1 Ergodic decomposition of invariant measures Let T : (Ω, F) (Ω, F) be measurable, and let M denote the space of T -invariant probability measures on (Ω, F). Then M is a convex set, although

More information

I. ANALYSIS; PROBABILITY

I. ANALYSIS; PROBABILITY ma414l1.tex Lecture 1. 12.1.2012 I. NLYSIS; PROBBILITY 1. Lebesgue Measure and Integral We recall Lebesgue measure (M411 Probability and Measure) λ: defined on intervals (a, b] by λ((a, b]) := b a (so

More information

ELEMENTS OF PROBABILITY THEORY

ELEMENTS OF PROBABILITY THEORY ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable

More information

1: PROBABILITY REVIEW

1: PROBABILITY REVIEW 1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following

More information

3 Integration and Expectation

3 Integration and Expectation 3 Integration and Expectation 3.1 Construction of the Lebesgue Integral Let (, F, µ) be a measure space (not necessarily a probability space). Our objective will be to define the Lebesgue integral R fdµ

More information

Probability and Measure

Probability and Measure Chapter 4 Probability and Measure 4.1 Introduction In this chapter we will examine probability theory from the measure theoretic perspective. The realisation that measure theory is the foundation of probability

More information

Lectures 22-23: Conditional Expectations

Lectures 22-23: Conditional Expectations Lectures 22-23: Conditional Expectations 1.) Definitions Let X be an integrable random variable defined on a probability space (Ω, F 0, P ) and let F be a sub-σ-algebra of F 0. Then the conditional expectation

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Gaussian vectors and central limit theorem

Gaussian vectors and central limit theorem Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables

More information

Probability Background

Probability Background Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second

More information

STOCHASTIC MODELS FOR WEB 2.0

STOCHASTIC MODELS FOR WEB 2.0 STOCHASTIC MODELS FOR WEB 2.0 VIJAY G. SUBRAMANIAN c 2011 by Vijay G. Subramanian. All rights reserved. Permission is hereby given to freely print and circulate copies of these notes so long as the notes

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

Product measure and Fubini s theorem

Product measure and Fubini s theorem Chapter 7 Product measure and Fubini s theorem This is based on [Billingsley, Section 18]. 1. Product spaces Suppose (Ω 1, F 1 ) and (Ω 2, F 2 ) are two probability spaces. In a product space Ω = Ω 1 Ω

More information

Gaussian, Markov and stationary processes

Gaussian, Markov and stationary processes Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November

More information

Inference for Stochastic Processes

Inference for Stochastic Processes Inference for Stochastic Processes Robert L. Wolpert Revised: June 19, 005 Introduction A stochastic process is a family {X t } of real-valued random variables, all defined on the same probability space

More information

Lecture 2: Repetition of probability theory and statistics

Lecture 2: Repetition of probability theory and statistics Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:

More information

MATHS 730 FC Lecture Notes March 5, Introduction

MATHS 730 FC Lecture Notes March 5, Introduction 1 INTRODUCTION MATHS 730 FC Lecture Notes March 5, 2014 1 Introduction Definition. If A, B are sets and there exists a bijection A B, they have the same cardinality, which we write as A, #A. If there exists

More information

MATH 418: Lectures on Conditional Expectation

MATH 418: Lectures on Conditional Expectation MATH 418: Lectures on Conditional Expectation Instructor: r. Ed Perkins, Notes taken by Adrian She Conditional expectation is one of the most useful tools of probability. The Radon-Nikodym theorem enables

More information

Notes on Measure, Probability and Stochastic Processes. João Lopes Dias

Notes on Measure, Probability and Stochastic Processes. João Lopes Dias Notes on Measure, Probability and Stochastic Processes João Lopes Dias Departamento de Matemática, ISEG, Universidade de Lisboa, Rua do Quelhas 6, 1200-781 Lisboa, Portugal E-mail address: jldias@iseg.ulisboa.pt

More information

CHAPTER 1. Martingales

CHAPTER 1. Martingales CHAPTER 1 Martingales The basic limit theorems of probability, such as the elementary laws of large numbers and central limit theorems, establish that certain averages of independent variables converge

More information

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets

More information

Lecture 22 Girsanov s Theorem

Lecture 22 Girsanov s Theorem Lecture 22: Girsanov s Theorem of 8 Course: Theory of Probability II Term: Spring 25 Instructor: Gordan Zitkovic Lecture 22 Girsanov s Theorem An example Consider a finite Gaussian random walk X n = n

More information

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past. 1 Markov chain: definition Lecture 5 Definition 1.1 Markov chain] A sequence of random variables (X n ) n 0 taking values in a measurable state space (S, S) is called a (discrete time) Markov chain, if

More information

Lectures for APM 541: Stochastic Modeling in Biology. Jay Taylor

Lectures for APM 541: Stochastic Modeling in Biology. Jay Taylor Lectures for APM 541: Stochastic Modeling in Biology Jay Taylor November 3, 2011 Contents 1 Distributions, Expectations, and Random Variables 4 1.1 Probability Spaces...................................

More information

Lecture 22: Variance and Covariance

Lecture 22: Variance and Covariance EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce

More information

Modern Discrete Probability Branching processes

Modern Discrete Probability Branching processes Modern Discrete Probability IV - Branching processes Review Sébastien Roch UW Madison Mathematics November 15, 2014 1 Basic definitions 2 3 4 Galton-Watson branching processes I Definition A Galton-Watson

More information

Martingale Theory for Finance

Martingale Theory for Finance Martingale Theory for Finance Tusheng Zhang October 27, 2015 1 Introduction 2 Probability spaces and σ-fields 3 Integration with respect to a probability measure. 4 Conditional expectation. 5 Martingales.

More information

STOR 635 Notes (S13)

STOR 635 Notes (S13) STOR 635 Notes (S13) Jimmy Jin UNC-Chapel Hill Last updated: 1/14/14 Contents 1 Measure theory and probability basics 2 1.1 Algebras and measure.......................... 2 1.2 Integration................................

More information

1 Stochastic Dynamic Programming

1 Stochastic Dynamic Programming 1 Stochastic Dynamic Programming Formally, a stochastic dynamic program has the same components as a deterministic one; the only modification is to the state transition equation. When events in the future

More information

Math 635: An Introduction to Brownian Motion and Stochastic Calculus

Math 635: An Introduction to Brownian Motion and Stochastic Calculus First Prev Next Go To Go Back Full Screen Close Quit 1 Math 635: An Introduction to Brownian Motion and Stochastic Calculus 1. Introduction and review 2. Notions of convergence and results from measure

More information

CONVERGENCE OF RANDOM SERIES AND MARTINGALES

CONVERGENCE OF RANDOM SERIES AND MARTINGALES CONVERGENCE OF RANDOM SERIES AND MARTINGALES WESLEY LEE Abstract. This paper is an introduction to probability from a measuretheoretic standpoint. After covering probability spaces, it delves into the

More information

PROBABILITY THEORY II

PROBABILITY THEORY II Ruprecht-Karls-Universität Heidelberg Institut für Angewandte Mathematik Prof. Dr. Jan JOHANNES Outline of the lecture course PROBABILITY THEORY II Summer semester 2016 Preliminary version: April 21, 2016

More information

(A n + B n + 1) A n + B n

(A n + B n + 1) A n + B n 344 Problem Hints and Solutions Solution for Problem 2.10. To calculate E(M n+1 F n ), first note that M n+1 is equal to (A n +1)/(A n +B n +1) with probability M n = A n /(A n +B n ) and M n+1 equals

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

1 Presessional Probability

1 Presessional Probability 1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows. Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage

More information

µ X (A) = P ( X 1 (A) )

µ X (A) = P ( X 1 (A) ) 1 STOCHASTIC PROCESSES This appendix provides a very basic introduction to the language of probability theory and stochastic processes. We assume the reader is familiar with the general measure and integration

More information

Universal examples. Chapter The Bernoulli process

Universal examples. Chapter The Bernoulli process Chapter 1 Universal examples 1.1 The Bernoulli process First description: Bernoulli random variables Y i for i = 1, 2, 3,... independent with P [Y i = 1] = p and P [Y i = ] = 1 p. Second description: Binomial

More information

JUSTIN HARTMANN. F n Σ.

JUSTIN HARTMANN. F n Σ. BROWNIAN MOTION JUSTIN HARTMANN Abstract. This paper begins to explore a rigorous introduction to probability theory using ideas from algebra, measure theory, and other areas. We start with a basic explanation

More information

Errata for FIRST edition of A First Look at Rigorous Probability

Errata for FIRST edition of A First Look at Rigorous Probability Errata for FIRST edition of A First Look at Rigorous Probability NOTE: the corrections below (plus many more improvements) were all encorporated into the second edition: J.S. Rosenthal, A First Look at

More information

STA 711: Probability & Measure Theory Robert L. Wolpert

STA 711: Probability & Measure Theory Robert L. Wolpert STA 711: Probability & Measure Theory Robert L. Wolpert 6 Independence 6.1 Independent Events A collection of events {A i } F in a probability space (Ω,F,P) is called independent if P[ i I A i ] = P[A

More information

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A )

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A ) 6. Brownian Motion. stochastic process can be thought of in one of many equivalent ways. We can begin with an underlying probability space (Ω, Σ, P) and a real valued stochastic process can be defined

More information

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition Filtrations, Markov Processes and Martingales Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition David pplebaum Probability and Statistics Department,

More information

We will briefly look at the definition of a probability space, probability measures, conditional probability and independence of probability events.

We will briefly look at the definition of a probability space, probability measures, conditional probability and independence of probability events. 1 Probability 1.1 Probability spaces We will briefly look at the definition of a probability space, probability measures, conditional probability and independence of probability events. Definition 1.1.

More information

Continuous Random Variables

Continuous Random Variables 1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables

More information

Probability: Handout

Probability: Handout Probability: Handout Klaus Pötzelberger Vienna University of Economics and Business Institute for Statistics and Mathematics E-mail: Klaus.Poetzelberger@wu.ac.at Contents 1 Axioms of Probability 3 1.1

More information

1 Martingales. Martingales. (Ω, B, P ) is a probability space.

1 Martingales. Martingales. (Ω, B, P ) is a probability space. Martingales January 8, 206 Debdee Pati Martingales (Ω, B, P ) is a robability sace. Definition. (Filtration) filtration F = {F n } n 0 is a collection of increasing sub-σfields such that for m n, we have

More information

Notes on Stochastic Calculus

Notes on Stochastic Calculus Notes on Stochastic Calculus David Nualart Kansas University nualart@math.ku.edu 1 Stochastic Processes 1.1 Probability Spaces and Random Variables In this section we recall the basic vocabulary and results

More information

1.1 Review of Probability Theory

1.1 Review of Probability Theory 1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,

More information

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539 Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory

More information

3. Review of Probability and Statistics

3. Review of Probability and Statistics 3. Review of Probability and Statistics ECE 830, Spring 2014 Probabilistic models will be used throughout the course to represent noise, errors, and uncertainty in signal processing problems. This lecture

More information

Probability Theory I: Syllabus and Exercise

Probability Theory I: Syllabus and Exercise Probability Theory I: Syllabus and Exercise Narn-Rueih Shieh **Copyright Reserved** This course is suitable for those who have taken Basic Probability; some knowledge of Real Analysis is recommended( will

More information

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get 18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in

More information

Math 832 Fall University of Wisconsin at Madison. Instructor: David F. Anderson

Math 832 Fall University of Wisconsin at Madison. Instructor: David F. Anderson Math 832 Fall 2013 University of Wisconsin at Madison Instructor: David F. Anderson Pertinent information Instructor: David Anderson Office: Van Vleck 617 email: anderson@math.wisc.edu Office hours: Mondays

More information

Applications of Ito s Formula

Applications of Ito s Formula CHAPTER 4 Applications of Ito s Formula In this chapter, we discuss several basic theorems in stochastic analysis. Their proofs are good examples of applications of Itô s formula. 1. Lévy s martingale

More information

Jump Processes. Richard F. Bass

Jump Processes. Richard F. Bass Jump Processes Richard F. Bass ii c Copyright 214 Richard F. Bass Contents 1 Poisson processes 1 1.1 Definitions............................. 1 1.2 Stopping times.......................... 3 1.3 Markov

More information

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2) 14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence

More information

Lecture 5. 1 Chung-Fuchs Theorem. Tel Aviv University Spring 2011

Lecture 5. 1 Chung-Fuchs Theorem. Tel Aviv University Spring 2011 Random Walks and Brownian Motion Tel Aviv University Spring 20 Instructor: Ron Peled Lecture 5 Lecture date: Feb 28, 20 Scribe: Yishai Kohn In today's lecture we return to the Chung-Fuchs theorem regarding

More information

Verona Course April Lecture 1. Review of probability

Verona Course April Lecture 1. Review of probability Verona Course April 215. Lecture 1. Review of probability Viorel Barbu Al.I. Cuza University of Iaşi and the Romanian Academy A probability space is a triple (Ω, F, P) where Ω is an abstract set, F is

More information

Stochastic integration. P.J.C. Spreij

Stochastic integration. P.J.C. Spreij Stochastic integration P.J.C. Spreij this version: April 22, 29 Contents 1 Stochastic processes 1 1.1 General theory............................... 1 1.2 Stopping times...............................

More information

4 Sums of Independent Random Variables

4 Sums of Independent Random Variables 4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables

More information

Lecture 17 Brownian motion as a Markov process

Lecture 17 Brownian motion as a Markov process Lecture 17: Brownian motion as a Markov process 1 of 14 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 17 Brownian motion as a Markov process Brownian motion is

More information

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample

More information

17. Convergence of Random Variables

17. Convergence of Random Variables 7. Convergence of Random Variables In elementary mathematics courses (such as Calculus) one speaks of the convergence of functions: f n : R R, then lim f n = f if lim f n (x) = f(x) for all x in R. This

More information

7 Convergence in R d and in Metric Spaces

7 Convergence in R d and in Metric Spaces STA 711: Probability & Measure Theory Robert L. Wolpert 7 Convergence in R d and in Metric Spaces A sequence of elements a n of R d converges to a limit a if and only if, for each ǫ > 0, the sequence a

More information

Notes 15 : UI Martingales

Notes 15 : UI Martingales Notes 15 : UI Martingales Math 733 - Fall 2013 Lecturer: Sebastien Roch References: [Wil91, Chapter 13, 14], [Dur10, Section 5.5, 5.6, 5.7]. 1 Uniform Integrability We give a characterization of L 1 convergence.

More information

Probability and Distributions

Probability and Distributions Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated

More information

Elementary Probability. Exam Number 38119

Elementary Probability. Exam Number 38119 Elementary Probability Exam Number 38119 2 1. Introduction Consider any experiment whose result is unknown, for example throwing a coin, the daily number of customers in a supermarket or the duration of

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information