Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability
|
|
- Eric Gardner
- 5 years ago
- Views:
Transcription
1 Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with probability 1 (w.p.1), to the random variable X as n iff Definition iti 1.2. X n converges in probability to the random variable X as n iff for every ε>0 1 2 Four different convergence concepts Definition 1.3. X n converges in r-mean to the random variable X as n iff Convergence in probability Definition 1.2. X n converges in probability to the random variable X as n iff for every ε>0 Definition 1.4. X n converges in distribution to the random variable X as n iff where C(F X ) is the continuity set of F X. p Notation. X n X as n. In situations where the limiting distribution is degenerate, that is, the limiting random variable X is a constant, convergence in probability is (in statistics) also known as consistency. Chebyshev s inequality. Let X be a random variable with mean μ and finite variance σ 2. Then 3 4 1
2 Problem a Let X₁,X₂, be i.i.d. Pa(1,2)-distributed random variables, and set p Y n = min{x₁,x₂,,x n }. Show that Y n 1 as n. Since it follows that The weak law of large numbers The weak law of large numbers. Let X 1, X 2, be a sequence of i.i.d. random variables with mean μ and finite variance σ 2 and set S n =X+X X+X n. Then The distribution function of Y n is therefore given by and so (for any ε>0) Proof. The statement is a simple consequence of Chebyshev s inequality. It thus follows that it follows that 5 6 Example: Consistency of S 2 Consistency of S 2. Let X 1, X 2, be a sequence of i.i.d. random variables with mean μ and finite variance σ 2. Define the sample variance by Convergence in probability: Extension Theorem 6.7 Theorem 6.7.Suppose that X 1, X 2, converges in probability to a constant a and that h is a continuous function. Then Proof. h is continuous, so given ε>0 there exists a δ>0 such that Since (prove this!) The continuity of h thus makes sure that it follows from Chebyshev s s inequality that and since X n converges in probability to X and thus, a sufficient condition for S 2 p σ 2 is that Var(S 2 ) 0 as n
3 Convergence in probability: Extension Exercise Exercise Suppose that X 1, X 2, converges in probability to a random variable X and that h is a continuous function. Then Convergence in probability: Extension Exercise We therefore divide the sample space into two disjoint subsets Proof. To prove this we use the fact that on a closed interval (compact set) any continuous function h is actually uniformly continuous. To this we use the fact that for any given η>0 there exists an A such that and it now follows that Since h for any A is uniformly continuous on [-A,A] we can (on this interval) for any given η>0 and ε>0 find a δ>0 and an m such that for any n>m 9 10 Almost sure convergence Definition 1.1. X n converges almost surely, or with probability one, to the random variable X as n iff Example: Almost sure convergence Let the sample space S be [0,1] with the uniform distribution. Define random variables X n (s) = s+s n and X(s) = s. As n we have that a.s. Notation. X n X as n. When to prove that X n converges (or fails to converge) almost surely we can use the following result. a.s. X n X as n iff, ε>0 and 0<δ<1, n 0 such that, n>n 0, which means that But since Pr([0,1)) = 1 it follows by Definition 1.1 that
4 Relationship: Almost sure convergence and convergence in probability Comparison of Definitions 1.1 and 1.2. We have that Example: Convergence in probability but not almost sure convergence Let the sample space S be [0,1] with the uniform distribution. Define X(s) = s and the sequence X 1,X 2, by Since for m>n, etc. It is clear that for any 0<ε<1 it is proven that p where I n is the interval related to X n. It is thus clear that t X n X as n. However, since X n (s) alternates between s and s+1 infinitely often, that is it is clear that X n does not converge to X almost surely as n Relationship: Convergence in r-mean and convergence in probability Convergence in distribution (and relationships between concepts) Definition 1.3. X n converges in r-mean to the random variable X as n iff Definition 1.4. X n converges in distribution to the random variable X as n iff r Notation. X n X as n. Convergence in r-mean is stronger convergence concept than convergence in probability. By Markov s inequality (for any ε>0) d where C(F X ) is the continuity set of F X. Notation. X n X as n. Convergence in distribution is the weakest concept of the four but also the most useful. The (complete) relationships can be described as which implies that where all implications are strict
5 Important results concerning limits Consider two functions f and g, such that Then Problem b Let X₁,X₂, be i.i.d. Pa(1,2)-distributed random variables, and set Y n = min{x₁,x₂,,x n }. Show that U n = n(y n -1) converges in distribution as n, and determine the limit distribution. Since If h is a continuous function then it follows that The single most important limit (in its most general form) is the following: Let a n a as n. Then d It is thus clear that U n X as n where X Exp(1/2) Convergence via transforms Important results concerning limits (when using Taylor series expansion) Theorem 4.1. Let X, X₁,X₂, be nonnegative, integer-valued random variables. Then The Taylor series expansion of a function f that is infinitely differentiable in a neighborhood of x=a is the power series Theorem 4.2. Let X₁,X₂, be random variables for which the mgf s exist for h<t<h for some h>0, and suppose that X is a random variable whose mgf ψ X (t) exists for h 1 t h 1 where 0<h 1 <h. If then which implies that Some terms in the expansion might be insignificant in the limit. For such terms we can use the o -concept. The function f(x) is said to be little-o of g(x) if f(x)/g(x) 0 as x 0 and we write
6 Problem Let X₁,X₂, be i.i.d. random variables with mean μ<, and let N n Ge(p n ), 0<p n <1, independent of X₁,X₂,. Determine the limit distribution of Problem We now note that for a general probability distribution with mean μ (where the mgf exists) it holds that as n if p n 0 as n. It follows from Theorem that that is and dfrom Theorem we have thatt Since ψ X (p n t) ψ X (0) = 1 as n it therefore follows that 21 d and so it is clear that Y n Exp(μ) as n. 22 The weak law of large numbers Revisited The weak law of large numbers (LLN). Let X 1, X 2, be a sequence of i.i.d. random variables with mean μ and mgf ψ X (t). Then The Central Limit Theorem The Central Limit Theorem (CLT). Let X 1, X 2, be i.i.d. random variables with mean μ, variance σ 2, and mgf ψ X (t) and set S n = X 1 +X 2 + +X n. Then Proof. Proof. Because of linear properties of moment generating functions it is no restriction to let μ=0 and σ 2 =1. This means that as n. It is clear that, or equvalently,
7 and therefore we get that The Central Limit Theorem Convergence of sums of sequences of random variables Theorem Let X₁,X₂, and Y₁,Y₂, be sequences of random variables. Then as n, and we are done since this is the moment generating function of N(0,1). Theorem Let X₁,X₂, X₂ and Y₁,Y₂, Y₂ be sequences of random variables. Suppose further that X n and Y n are independent for all n and that X and Y are independent. Then Slutsky s theorem (or Cramér s theorem) Theorem 6.5. Let X₁,X₂, and Y₁,Y₂, be sequences of random variables. Suppose that Exercise Let X₁,X₂, be i.i.d. Be(p)-distributed random variables where 0<p<1. We would like to construct a confidence interval for the population proportion p. What about the random behavior of the sample proportion? Set S n = X 1 +X 2 + +X n and consider Y₁,Y₂, where Y n =S n /n. Since where a is a constant. Then it follows by the Central Limit Theorem (CLT) that which, for instance, implies that
8 Exercise Hence, in the denominator we have to replace p(1-p) with Y n (1-Y n ), that is Exercise Now it follows by the third and the fourth result of Theorem 6.5 (Slutsky) that is to be replaced by With the aid of Slutsky s theorem we can prove that CLT still works. First it follows by the law of large numbers that Since the square root is a continuous function, it follows by Theorem 6.7 that Finally, it follows by the fourth result of Theorem 6.5 (Slutsky) that and therefore by the second result of Theorem 6.5 (Slutsky) we have that 29 as n. It is thus clear that the approximation is still valid for sufficiently large sample sizes. 30 Problem Let X₁,X₂, be positive i.i.d. random variables with mean μ and variance σ 2 <, and set S n = X 1 +X 2 + +X n. Determine the limit distribution of So therefore we have that Problem In order to use the central limit theorem we rewrite the expression as Hence, it follows from Theorem 6.5 (Slutsky) that The expression in the numerator meets the requirements of the central limit theorem, and the expression in the denominator meets the requirements of the law of large numbers. since a linear function of a normal random variable also is normal
Lecture 21: Convergence of transformations and generating a random variable
Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationLimiting Distributions
We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationIf g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get
18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in
More informationCS145: Probability & Computing
CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,
More informationEconomics 583: Econometric Theory I A Primer on Asymptotics
Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationLecture Notes 3 Convergence (Chapter 5)
Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let
More informationCOMPSCI 240: Reasoning Under Uncertainty
COMPSCI 240: Reasoning Under Uncertainty Andrew Lan and Nic Herndon University of Massachusetts at Amherst Spring 2019 Lecture 20: Central limit theorem & The strong law of large numbers Markov and Chebyshev
More informationThe space complexity of approximating the frequency moments
The space complexity of approximating the frequency moments Felix Biermeier November 24, 2015 1 Overview Introduction Approximations of frequency moments lower bounds 2 Frequency moments Problem Estimate
More informationExercises in Extreme value theory
Exercises in Extreme value theory 2016 spring semester 1. Show that L(t) = logt is a slowly varying function but t ǫ is not if ǫ 0. 2. If the random variable X has distribution F with finite variance,
More informationElements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley
Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the
More informationModule 3. Function of a Random Variable and its distribution
Module 3 Function of a Random Variable and its distribution 1. Function of a Random Variable Let Ω, F, be a probability space and let be random variable defined on Ω, F,. Further let h: R R be a given
More information1 Exercises for lecture 1
1 Exercises for lecture 1 Exercise 1 a) Show that if F is symmetric with respect to µ, and E( X )
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More informationHomework 11. Solutions
Homework 11. Solutions Problem 2.3.2. Let f n : R R be 1/n times the characteristic function of the interval (0, n). Show that f n 0 uniformly and f n µ L = 1. Why isn t it a counterexample to the Lebesgue
More informationChapter 5. Measurable Functions
Chapter 5. Measurable Functions 1. Measurable Functions Let X be a nonempty set, and let S be a σ-algebra of subsets of X. Then (X, S) is a measurable space. A subset E of X is said to be measurable if
More informationRegression and Statistical Inference
Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF
More informationSTAT 7032 Probability Spring Wlodek Bryc
STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationOn the convergence of sequences of random variables: A primer
BCAM May 2012 1 On the convergence of sequences of random variables: A primer Armand M. Makowski ECE & ISR/HyNet University of Maryland at College Park armand@isr.umd.edu BCAM May 2012 2 A sequence a :
More informationlim F n(x) = F(x) will not use either of these. In particular, I m keeping reserved for implies. ) Note:
APPM/MATH 4/5520, Fall 2013 Notes 9: Convergence in Distribution and the Central Limit Theorem Definition: Let {X n } be a sequence of random variables with cdfs F n (x) = P(X n x). Let X be a random variable
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued
Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and
More informationPRACTICE PROBLEM SET
PRACTICE PROBLEM SET NOTE: On the exam, you will have to show all your work (unless told otherwise), so write down all your steps and justify them. Exercise. Solve the following inequalities: () x < 3
More informationLarge Sample Properties of Estimators in the Classical Linear Regression Model
Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in
More informationCHAPTER 3: LARGE SAMPLE THEORY
CHAPTER 3 LARGE SAMPLE THEORY 1 CHAPTER 3: LARGE SAMPLE THEORY CHAPTER 3 LARGE SAMPLE THEORY 2 Introduction CHAPTER 3 LARGE SAMPLE THEORY 3 Why large sample theory studying small sample property is usually
More informationChapter-2 Relations and Functions. Miscellaneous
1 Chapter-2 Relations and Functions Miscellaneous Question 1: The relation f is defined by The relation g is defined by Show that f is a function and g is not a function. The relation f is defined as It
More information6.1 Moment Generating and Characteristic Functions
Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,
More informationChapter 7: Special Distributions
This chater first resents some imortant distributions, and then develos the largesamle distribution theory which is crucial in estimation and statistical inference Discrete distributions The Bernoulli
More informationElements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley
Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the
More information8 Laws of large numbers
8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable
More informationStochastic Models (Lecture #4)
Stochastic Models (Lecture #4) Thomas Verdebout Université libre de Bruxelles (ULB) Today Today, our goal will be to discuss limits of sequences of rv, and to study famous limiting results. Convergence
More information1 Adeles over Q. 1.1 Absolute values
1 Adeles over Q 1.1 Absolute values Definition 1.1.1 (Absolute value) An absolute value on a field F is a nonnegative real valued function on F which satisfies the conditions: (i) x = 0 if and only if
More information9. Series representation for analytic functions
9. Series representation for analytic functions 9.. Power series. Definition: A power series is the formal expression S(z) := c n (z a) n, a, c i, i =,,, fixed, z C. () The n.th partial sum S n (z) is
More informationUses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).
1 Economics 620, Lecture 8a: Asymptotics II Uses of Asymptotic Distributions: Suppose X n! 0 in probability. (What can be said about the distribution of X n?) In order to get distribution theory, we need
More informationPolynomial Approximations and Power Series
Polynomial Approximations and Power Series June 24, 206 Tangent Lines One of the first uses of the derivatives is the determination of the tangent as a linear approximation of a differentiable function
More informationStochastic Convergence, Delta Method & Moment Estimators
Stochastic Convergence, Delta Method & Moment Estimators Seminar on Asymptotic Statistics Daniel Hoffmann University of Kaiserslautern Department of Mathematics February 13, 2015 Daniel Hoffmann (TU KL)
More informationa. Define a function called an inner product on pairs of points x = (x 1, x 2,..., x n ) and y = (y 1, y 2,..., y n ) in R n by
Real Analysis Homework 1 Solutions 1. Show that R n with the usual euclidean distance is a metric space. Items a-c will guide you through the proof. a. Define a function called an inner product on pairs
More informationProving the central limit theorem
SOR3012: Stochastic Processes Proving the central limit theorem Gareth Tribello March 3, 2019 1 Purpose In the lectures and exercises we have learnt about the law of large numbers and the central limit
More informationEconomics 620, Lecture 8: Asymptotics I
Economics 620, Lecture 8: Asymptotics I Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 8: Asymptotics I 1 / 17 We are interested in the properties of estimators
More informationNotes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770
Notes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770 Jonathan B. Hill Dept. of Economics University of North Carolina - Chapel Hill November
More informationSUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)
SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems
More informationA Primer on Asymptotics
A Primer on Asymptotics Eric Zivot Department of Economics University of Washington September 30, 2003 Revised: October 7, 2009 Introduction The two main concepts in asymptotic theory covered in these
More informationFinite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product
Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,
More informationCentral limit theorem. Paninski, Intro. Math. Stats., October 5, probability, Z N P Z, if
Paninski, Intro. Math. Stats., October 5, 2005 35 probability, Z P Z, if P ( Z Z > ɛ) 0 as. (The weak LL is called weak because it asserts convergence in probability, which turns out to be a somewhat weak
More informationConvergence in Distribution
Convergence in Distribution Undergraduate version of central limit theorem: if X 1,..., X n are iid from a population with mean µ and standard deviation σ then n 1/2 ( X µ)/σ has approximately a normal
More informationHomework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples
Homework #3 36-754, Spring 27 Due 14 May 27 1 Convergence of the empirical CDF, uniform samples In this problem and the next, X i are IID samples on the real line, with cumulative distribution function
More informationNotes on uniform convergence
Notes on uniform convergence Erik Wahlén erik.wahlen@math.lu.se January 17, 2012 1 Numerical sequences We begin by recalling some properties of numerical sequences. By a numerical sequence we simply mean
More informationLecture 11. Multivariate Normal theory
10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances
More informationEcon 508B: Lecture 5
Econ 508B: Lecture 5 Expectation, MGF and CGF Hongyi Liu Washington University in St. Louis July 31, 2017 Hongyi Liu (Washington University in St. Louis) Math Camp 2017 Stats July 31, 2017 1 / 23 Outline
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationSome Background Material
Chapter 1 Some Background Material In the first chapter, we present a quick review of elementary - but important - material as a way of dipping our toes in the water. This chapter also introduces important
More informationProbability inequalities 11
Paninski, Intro. Math. Stats., October 5, 2005 29 Probability inequalities 11 There is an adage in probability that says that behind every limit theorem lies a probability inequality (i.e., a bound on
More informationOn probabilities of large and moderate deviations for L-statistics: a survey of some recent developments
UDC 519.2 On probabilities of large and moderate deviations for L-statistics: a survey of some recent developments N. V. Gribkova Department of Probability Theory and Mathematical Statistics, St.-Petersburg
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4.1 (*. Let independent variables X 1,..., X n have U(0, 1 distribution. Show that for every x (0, 1, we have P ( X (1 < x 1 and P ( X (n > x 1 as n. Ex. 4.2 (**. By using induction
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More informationRandom Bernstein-Markov factors
Random Bernstein-Markov factors Igor Pritsker and Koushik Ramachandran October 20, 208 Abstract For a polynomial P n of degree n, Bernstein s inequality states that P n n P n for all L p norms on the unit
More informationLecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t )
LECURE NOES 21 Lecture 4 7. Sufficient statistics Consider the usual statistical setup: the data is X and the paramter is. o gain information about the parameter we study various functions of the data
More information7 Convergence in R d and in Metric Spaces
STA 711: Probability & Measure Theory Robert L. Wolpert 7 Convergence in R d and in Metric Spaces A sequence of elements a n of R d converges to a limit a if and only if, for each ǫ > 0, the sequence a
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More information) ) = γ. and P ( X. B(a, b) = Γ(a)Γ(b) Γ(a + b) ; (x + y, ) I J}. Then, (rx) a 1 (ry) b 1 e (x+y)r r 2 dxdy Γ(a)Γ(b) D
3 Independent Random Variables II: Examples 3.1 Some functions of independent r.v. s. Let X 1, X 2,... be independent r.v. s with the known distributions. Then, one can compute the distribution of a r.v.
More informationSTATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN
Massimo Guidolin Massimo.Guidolin@unibocconi.it Dept. of Finance STATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN SECOND PART, LECTURE 2: MODES OF CONVERGENCE AND POINT ESTIMATION Lecture 2:
More informationR. Koenker Spring 2017 Economics 574 Problem Set 1
R. Koenker Spring 207 Economics 574 Problem Set.: Suppose X, Y are random variables with joint density f(x, y) = x 2 + xy/3 x [0, ], y [0, 2]. Find: (a) the joint df, (b) the marginal density of X, (c)
More informationthe convolution of f and g) given by
09:53 /5/2000 TOPIC Characteristic functions, cont d This lecture develops an inversion formula for recovering the density of a smooth random variable X from its characteristic function, and uses that
More informationThe Delta Method and Applications
Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationQualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf
Part : Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section
More informationLecture 4 : Random variable and expectation
Lecture 4 : Random variable and expectation Study Objectives: to learn the concept of 1. Random variable (rv), including discrete rv and continuous rv; and the distribution functions (pmf, pdf and cdf).
More informationCHAPTER 6. Limits of Functions. 1. Basic Definitions
CHAPTER 6 Limits of Functions 1. Basic Definitions DEFINITION 6.1. Let D Ω R, x 0 be a limit point of D and f : D! R. The limit of f (x) at x 0 is L, if for each " > 0 there is a ± > 0 such that when x
More informationWeak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij
Weak convergence and Brownian Motion (telegram style notes) P.J.C. Spreij this version: December 8, 2006 1 The space C[0, ) In this section we summarize some facts concerning the space C[0, ) of real
More information1 of 7 7/16/2009 6:12 AM Virtual Laboratories > 7. Point Estimation > 1 2 3 4 5 6 1. Estimators The Basic Statistical Model As usual, our starting point is a random experiment with an underlying sample
More informationStat 5101 Lecture Slides: Deck 7 Asymptotics, also called Large Sample Theory. Charles J. Geyer School of Statistics University of Minnesota
Stat 5101 Lecture Slides: Deck 7 Asymptotics, also called Large Sample Theory Charles J. Geyer School of Statistics University of Minnesota 1 Asymptotic Approximation The last big subject in probability
More informationST5215: Advanced Statistical Theory
Department of Statistics & Applied Probability Monday, September 26, 2011 Lecture 10: Exponential families and Sufficient statistics Exponential Families Exponential families are important parametric families
More informationSelected Exercises on Expectations and Some Probability Inequalities
Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ
More informationLecture 4: September Reminder: convergence of sequences
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 4: September 6 In this lecture we discuss the convergence of random variables. At a high-level, our first few lectures focused
More informationSupplementary Notes for W. Rudin: Principles of Mathematical Analysis
Supplementary Notes for W. Rudin: Principles of Mathematical Analysis SIGURDUR HELGASON In 8.00B it is customary to cover Chapters 7 in Rudin s book. Experience shows that this requires careful planning
More informationElements of Probability Theory
Elements of Probability Theory CHUNG-MING KUAN Department of Finance National Taiwan University December 5, 2009 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, 2009 1 / 58
More informationInfinite Series. Copyright Cengage Learning. All rights reserved.
Infinite Series Copyright Cengage Learning. All rights reserved. Sequences Copyright Cengage Learning. All rights reserved. Objectives List the terms of a sequence. Determine whether a sequence converges
More informationMultivariate Analysis and Likelihood Inference
Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini April 27, 2018 1 / 80 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationProbability Lecture III (August, 2006)
robability Lecture III (August, 2006) 1 Some roperties of Random Vectors and Matrices We generalize univariate notions in this section. Definition 1 Let U = U ij k l, a matrix of random variables. Suppose
More informationThe Theory of Statistics and Its Applications
The Theory of Statistics and Its Applications 1 By Dennis D. Cox Rice University c Copyright 2000, 2004 by Dennis D. Cox. May be reproduced for personal use by students in STAT 532 at Rice University.
More informationB553 Lecture 1: Calculus Review
B553 Lecture 1: Calculus Review Kris Hauser January 10, 2012 This course requires a familiarity with basic calculus, some multivariate calculus, linear algebra, and some basic notions of metric topology.
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationIntroduction and Review of Power Series
Introduction and Review of Power Series Definition: A power series in powers of x a is an infinite series of the form c n (x a) n = c 0 + c 1 (x a) + c 2 (x a) 2 +...+c n (x a) n +... If a = 0, this is
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationp. 6-1 Continuous Random Variables p. 6-2
Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability (>). Often, there is interest in random variables
More informationMathematics for Business and Economics - I. Chapter 5. Functions (Lecture 9)
Mathematics for Business and Economics - I Chapter 5. Functions (Lecture 9) Functions The idea of a function is this: a correspondence between two sets D and R such that to each element of the first set,
More informationChapter Generating Functions
Chapter 8.1.1-8.1.2. Generating Functions Prof. Tesler Math 184A Fall 2017 Prof. Tesler Ch. 8. Generating Functions Math 184A / Fall 2017 1 / 63 Ordinary Generating Functions (OGF) Let a n (n = 0, 1,...)
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationChapter 7: Exponents
Chapter : Exponents Algebra Chapter Notes Name: Notes #: Sections.. Section.: Review Simplify; leave all answers in positive exponents:.) m -.) y -.) m 0.) -.) -.) - -.) (m ) 0.) 0 x y Evaluate if a =
More informationChapter 11 - Sequences and Series
Calculus and Analytic Geometry II Chapter - Sequences and Series. Sequences Definition. A sequence is a list of numbers written in a definite order, We call a n the general term of the sequence. {a, a
More informationStatistics for Economists. Lectures 3 & 4
Statistics for Economists Lectures 3 & 4 Asrat Temesgen Stockholm University 1 CHAPTER 2- Discrete Distributions 2.1. Random variables of the Discrete Type Definition 2.1.1: Given a random experiment with
More information2. Variance and Higher Moments
1 of 16 7/16/2009 5:45 AM Virtual Laboratories > 4. Expected Value > 1 2 3 4 5 6 2. Variance and Higher Moments Recall that by taking the expected value of various transformations of a random variable,
More informationThe uniform metric on product spaces
The uniform metric on product spaces Jordan Bell jordan.bell@gmail.com Department of Mathematics, University of Toronto April 3, 2014 1 Metric topology If (X, d) is a metric space, a X, and r > 0, then
More informationMaximum Likelihood Asymptotic Theory. Eduardo Rossi University of Pavia
Maximum Likelihood Asymtotic Theory Eduardo Rossi University of Pavia Slutsky s Theorem, Cramer s Theorem Slutsky s Theorem Let {X N } be a random sequence converging in robability to a constant a, and
More informationTheorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1
Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be
More information