Economics 241B Review of Limit Theorems for Sequences of Random Variables
|
|
- Bryan Casey
- 5 years ago
- Views:
Transcription
1 Economics 241B Review of Limit Theorems for Sequences of Random Variables Convergence in Distribution The previous de nitions of convergence focus on the outcome sequences of a random variable. Convergence in distribution refers to the probability distribution of each element of the sequence directly. De nition. A sequence of random variables Y1 ; Y 2 ; : : : in distribution to a random variable Y if at all c such that F Y lim P Y n < c = P (Y < c) is continuous. We express this as Y n D! Y. An equivalent de nition is De nition. A sequence of random variables Y1 ; Y 2 ; : : : in distribution to F Y if lim F Y n (c) = F Y (c) at all c such that F Y is continuous. is said to converge is said to converge We express this as Y D n! F Y. The distribution F Y is the asymptotic (or limiting) distribution of Y n. Convergence in distribution is simply the pointwise convergence of F Yn to F Y. ( The requirement that F Y be continuous at all c will hold for all applications in this course, with the exception of binary dependent variables.) In most cases F Y is a Gaussian distribution, for which we write Y n D! N ; 2 : Convergence in distribution is closely related to convergence in probability, if the limit quantity is changed. In the discussion above, a sequence of random variables was shown to converge (in probability) to a constant. One can also establish that a sequence of random variables converges in probability to a random variable. Recall, for convergence (i.p.) to a constant, the probability that Y n is in an neighborhood of must be high, P ( Yn < ) > 1. That is P! : Yn (!) < > 1 : For convergence to a random variable, Y, we need P! : Yn (!) Y (!) < > 1 ; (1)
2 that is, for large n we want the probability that the histogram for Y n is close to the histogram for Y. There is no natural measure of distance here, although we think of as de ning the histogram bin width. If the two histograms are arbitrarily close as n increases, then we write Y n P! Y: Because the two histograms are arbitrarily close, it should not be surprising that Y n P! Y ) Y n D! F Y (sometimes written Y n D! Y ). Why doesn t the reverse hold? Because of the basic workings of probability spaces, about which we rarely concern ourselves. By the fact that both Y n and Y are indexed by the same! in (1), both quantities are de ned on the same probability space. No such requirement is made in establishing convergence of distribution, we simply discuss a sequence of distribution functions. Not only could Y n and Y be de ned on di erent probability spaces, for each n, Y n could be de ned on a di erent probability space. Yet the two de nitions are almost i. The Skorhorod construction is a proof that establishes that if Y D n! Y, then it is always possible to construct sequences Y n 0 and Y 0 that are de ned on the same probability space and are similar to Y n and Y so that Y n 0 P! Y 0. The extension to a sequence of random vectors is immediate: Yn! D Y if the joint c.d.f. F n converges to the joint c.d.f. F of Y at every continuity point of F. Note, however, that unlike other concepts of convergence, for convergence in distribution, element by element convergence does not necessarily mean convergence for the vector sequence. That is "each element Y D n!corresponding element of Y " does not imply " Y D n! Y ". A common way to connect scalar convergence and vector convergence in distribution is Multivariate Convergence in Distribution Theorem: Let Yn be a sequence of K-dimensional random vectors. Then Y n D! Y, 0 Yn D! 0 Y for any K-dimensional vector of real numbers. Convergence in Distribution vs. Convergence in Moments The moments of the limit distribution of Y n are not necessarily equal to the limit of the moments of Y n. Thus Y n D! Y does not imply that lim E Y n = E (Y ). However
3 Lemma (convergence in distribution and moments): Let sn be the s-th moment of Yn and lim sn = s where s is nite (i.e. a real number). Then Y n D! Y ) s is the s-th moment of Y: Thus, for example, if the variance of a sequence of random variables (converging in distribution) converges to some nite number, then that number is the variance of the limiting distribution. Relation among Modes of Convergence As noted above, some modes of convergence are weaker than others. Lemma (relation among the four modes of convergence): (a) Yn m:s:! ) Y p n!. So Yn m:s:! Y ) Y p n! Y. (b) Yn a:s:! ) Y p n!. So Yn a:s:! Y ) Y p n! Y. (c) Yn p!, Yn d!. Part (c) states that if the limiting random variable is a constant (a trivial random variable) then convergence in distribution is the same as convergence in probability. Three Useful Results The great use of asymptotic analysis is that we can easily de ne the convergence properties of ratios, while we cannot easily de ne the expectation of ratios (and so cannot easily do nite sample analysis). These results, which highlight such features, are essential for large-sample theory. Lemma (continuous mapping theorem): Suppose that g () is a continuous function that does not depend on n. Then: Y n P! ) g Y n P! g () Y n D! Y ) g Y n D! g (Y ) : The basic idea is that because g () is continuous, g Y n will be close to g () provided that Y n is close to. Note g Y 1 n = n 2 Yn depends on n and is not covered by this proposition. An immediate implication is that the usual arithmetic
4 operations preserve convergence in probability. For example Y n P! ; Xn P! ) Y n + X n P! + Y n P! ; Xn P! ) Y n Xn P! Y n P! ; Xn P! ) Y n = X n P! = provided that 6= 0 Y n P! ) Y 1 n P! 1 provided that is invertible. The next result combines convergence in probability and distribution, and is used repeatedly to derive the limit theory for our estimators. Lemma (parts a and c are called Slutzky s Theorem): (a) Y n P! and X n D! X ) Y n + X n D! X + (b) Y n P! 0 and X n D! X ) Y 0 nx n P! 0 (c) A n P! A and X n D! X ) A n X n D! AX (d) A n P! A and X n D! X ) X 0 na 1 n X n D! X 0 A 1 X Parts (c)-(d) require that X and A be conformable and (for part (d)) that A be invertible. In particular, if X N (0; V ), then Y n +X n D! N (; V ) and A n X n D! N (0; AV A 0 ). Also, if we set = 0, then part (a) implies Y n P! 0 and X n D! X ) Y n + X n D! X: Let Z n = Y n + X n, so Y P p n! 0 (i.e. Z n X n! 0) implies that the asymptotic p distribution of Z n is the same as that of X n. When Z n X n! 0 we sometimes say that the two sequences are asymptotically equivalent (Z n X n ) and write a Z n = X n + o P where o P is some suitable random variable ( Y n here) that converges to zero in probability. We employ the lemma in a standard trick to derive the limit distribution of a sequence of random variables. Suppose we wish to nd the limit distribution of Y nx 0 n, where we know X D n! X and Y P n!. Then we replace Y n with, and determine the limit distribution of 0 X n. Why does this work? Because Y nx 0 n a 0 X n, which is more easily seen in the equivalent statement Y 0 nx n = 0 X n + o P
5 where o P here is Yn 0 Xn. To test nonlinear hypotheses, given the asymptotic distribution of the estimator, we need Lemma (delta method): Let fb n g n1 be a sequence of K-dimensional random variables such that B n P! n 1 2 (Bn ) D! N (0; V ) : Let g () : R K! R Q have continuous rst derivatives, with G () denoting the Q K matrix of rst derivatives evaluated at : G () 0 : Then n 1 2 (g (Bn ) g ()) D! N 0; G () V G () 0 : Proof: A mean-value expansion yields n 1 2 g (B n ) = n 1 2 g ()+n 1 2 G (B ) (B n ), where Bn is the mean value between B n and. Because B P n!, Bn P! and, by continuity, G (Bn)! P G (). Thus n 1 2 [g (Bn ) g ()] = n 1 2 G (B n ) (B n ) D! G () N (0; V ) : Viewing Estimators as Sequences of Random Variables Let B n be an estimator of. If B P n!, then B n is consistent for. The asymptotic bias of an estimator is not an agreed upon concept. common de nition is lim E (B n) ; The most so consistency and asymptotic unbiasedness are distinct concepts. Hayashi uses p lim B n, so consistency and asymptotic unbiasedness are identical for him. If, in addition, n 1 2 (B n )! D N (0; V ), then B n is asymptotically Gaussian (and n 1 2 consistent, hence CAN). The asymptotic variance is V.
6 Laws of Large Numbers and Central Limit Theorems To verify convergence in probability from basic principles, we would need to know the joint distribution of the rst n elements in the sequence. To verify convergence almost surely, we would need to know the joint distribution of the entire (in nite dimensional) sequence. To verify convergence in quadratic mean, we would need to know the rst two moments of the nite sample distribution of the estimator. In general, we do not know any of these distributions. Instead, we turn to laws of large numbers. Laws of large numbers are one method of establishing convergence in probability (or almost surely) to a constant. They are applied to sequences of random variables, for which each member of the sequence is an (increasing) sum of random variables, such as Yn. An LLN is strong if the convergence is almost sure and weak if the convergence is in probability. Chebychev s Weak LLN: " lim E Y n = ; lim V ar Y n = 0" ) Yn p! : The following strong LLN assumes that fy t g is i.i.d. but the variance does not need to be nite Kolmogorov s Second Strong LLN: Let fy t g be i.i.d. with EY t =. a:s: Y n!. Then Central limit theorems are a principal method of establishing convergence in distribution, as they govern the limit behavior of the di erence between Y n and EY n (which equals EY t if fy t g is i.i.d.) blown up by p n. For i.i.d. sequences, the only CLT we need is Lindberg-Levy CLT: Let fy t g be i.i.d. with EY t = and V ar (Y t ) =. Then p n Yn = p 1 nx (Y t )! d N (0; ) : n t=1 We read the above CLT as: a sequence of random vectors p n Y n converges in distribution to a random vector whose distribution is N (0; ). To understand how to relate the above CLT for random vectors, with the CLT for random scalars, let be any vector of real numbers with the same dimension as Y t. Now f 0 Y t g is a sequence of random variables with E ( 0 Y t ) = 0 and V ar ( 0 Y t ) = 0. The scalar version of Lindberg-Levy implies p n 0 Yn 0 = 0p n Y n d! N (0; 0 ) :
7 Yet this limit distribution is the distribution of 0 Z, where Z N (0; ), hence the multivariate convergence in distribution theorem establishes the multivariate Lindberg-Levy CLT.
Uses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).
1 Economics 620, Lecture 8a: Asymptotics II Uses of Asymptotic Distributions: Suppose X n! 0 in probability. (What can be said about the distribution of X n?) In order to get distribution theory, we need
More informationElements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley
Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the
More informationRings of Residues. S. F. Ellermeyer. September 18, ; [1] m
Rings of Residues S F Ellermeyer September 18, 2006 If m is a positive integer, then we obtain the partition C = f[0] m ; [1] m ; : : : ; [m 1] m g of Z into m congruence classes (This is discussed in
More informationNotes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770
Notes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770 Jonathan B. Hill Dept. of Economics University of North Carolina - Chapel Hill November
More informationParametric Inference on Strong Dependence
Parametric Inference on Strong Dependence Peter M. Robinson London School of Economics Based on joint work with Javier Hualde: Javier Hualde and Peter M. Robinson: Gaussian Pseudo-Maximum Likelihood Estimation
More informationChapter 1. GMM: Basic Concepts
Chapter 1. GMM: Basic Concepts Contents 1 Motivating Examples 1 1.1 Instrumental variable estimator....................... 1 1.2 Estimating parameters in monetary policy rules.............. 2 1.3 Estimating
More informationlim F n(x) = F(x) will not use either of these. In particular, I m keeping reserved for implies. ) Note:
APPM/MATH 4/5520, Fall 2013 Notes 9: Convergence in Distribution and the Central Limit Theorem Definition: Let {X n } be a sequence of random variables with cdfs F n (x) = P(X n x). Let X be a random variable
More informationEconomics 620, Lecture 8: Asymptotics I
Economics 620, Lecture 8: Asymptotics I Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 8: Asymptotics I 1 / 17 We are interested in the properties of estimators
More informationGraduate Econometrics I: Asymptotic Theory
Graduate Econometrics I: Asymptotic Theory Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Asymptotic Theory
More informationEconomics 583: Econometric Theory I A Primer on Asymptotics
Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:
More informationLimiting Distributions
We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results
More informationTime Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley
Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More information2014 Preliminary Examination
014 reliminary Examination 1) Standard error consistency and test statistic asymptotic normality in linear models Consider the model for the observable data y t ; x T t n Y = X + U; (1) where is a k 1
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationSome asymptotic expansions and distribution approximations outside a CLT context 1
6th St.Petersburg Workshop on Simulation (2009) 444-448 Some asymptotic expansions and distribution approximations outside a CLT context 1 Manuel L. Esquível 2, João Tiago Mexia 3, João Lita da Silva 4,
More informationLecture Notes 3 Convergence (Chapter 5)
Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let
More informationGMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails
GMM-based inference in the AR() panel data model for parameter values where local identi cation fails Edith Madsen entre for Applied Microeconometrics (AM) Department of Economics, University of openhagen,
More information3 Random Samples from Normal Distributions
3 Random Samples from Normal Distributions Statistical theory for random samples drawn from normal distributions is very important, partly because a great deal is known about its various associated distributions
More informationLecture 6: Contraction mapping, inverse and implicit function theorems
Lecture 6: Contraction mapping, inverse and implicit function theorems 1 The contraction mapping theorem De nition 11 Let X be a metric space, with metric d If f : X! X and if there is a number 2 (0; 1)
More informationLECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT
MARCH 29, 26 LECTURE 2 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT (Davidson (2), Chapter 4; Phillips Lectures on Unit Roots, Cointegration and Nonstationarity; White (999), Chapter 7) Unit root processes
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationSIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011
SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER By Donald W. K. Andrews August 2011 COWLES FOUNDATION DISCUSSION PAPER NO. 1815 COWLES FOUNDATION FOR RESEARCH IN ECONOMICS
More informationECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Winter 2014 Instructor: Victor Aguirregabiria
ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Winter 2014 Instructor: Victor guirregabiria SOLUTION TO FINL EXM Monday, pril 14, 2014. From 9:00am-12:00pm (3 hours) INSTRUCTIONS:
More informationEstimation, Inference, and Hypothesis Testing
Chapter 2 Estimation, Inference, and Hypothesis Testing Note: The primary reference for these notes is Ch. 7 and 8 of Casella & Berger 2. This text may be challenging if new to this topic and Ch. 7 of
More informationLecture 11. Probability Theory: an Overveiw
Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationMath Camp II. Calculus. Yiqing Xu. August 27, 2014 MIT
Math Camp II Calculus Yiqing Xu MIT August 27, 2014 1 Sequence and Limit 2 Derivatives 3 OLS Asymptotics 4 Integrals Sequence Definition A sequence {y n } = {y 1, y 2, y 3,..., y n } is an ordered set
More informationEconomics 620, Lecture 9: Asymptotics III: Maximum Likelihood Estimation
Economics 620, Lecture 9: Asymptotics III: Maximum Likelihood Estimation Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 9: Asymptotics III(MLE) 1 / 20 Jensen
More informationElements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley
Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the
More informationTerminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the
More informationStatistical Data Analysis
DS-GA 0 Lecture notes 8 Fall 016 1 Descriptive statistics Statistical Data Analysis In this section we consider the problem of analyzing a set of data. We describe several techniques for visualizing the
More informationStochastic Convergence, Delta Method & Moment Estimators
Stochastic Convergence, Delta Method & Moment Estimators Seminar on Asymptotic Statistics Daniel Hoffmann University of Kaiserslautern Department of Mathematics February 13, 2015 Daniel Hoffmann (TU KL)
More informationChapter 6: Large Random Samples Sections
Chapter 6: Large Random Samples Sections 6.1: Introduction 6.2: The Law of Large Numbers Skip p. 356-358 Skip p. 366-368 Skip 6.4: The correction for continuity Remember: The Midterm is October 25th in
More informationStat 5101 Lecture Slides: Deck 7 Asymptotics, also called Large Sample Theory. Charles J. Geyer School of Statistics University of Minnesota
Stat 5101 Lecture Slides: Deck 7 Asymptotics, also called Large Sample Theory Charles J. Geyer School of Statistics University of Minnesota 1 Asymptotic Approximation The last big subject in probability
More informationLarge Sample Theory. Convergence. Central Limit Theorems Asymptotic Distribution Delta Method. Convergence in Probability Convergence in Distribution
Large Sample Theory Covergece Covergece i Probability Covergece i Distributio Cetral Limit Theorems Asymptotic Distributio Delta Method Covergece i Probability A sequece of radom scalars {z } = (z 1,z,
More informationLecture Tricks with Random Variables: The Law of Large Numbers & The Central Limit Theorem
Math 408 - Mathematical Statistics Lecture 9-10. Tricks with Random Variables: The Law of Large Numbers & The Central Limit Theorem February 6-8, 2013 Konstantin Zuev (USC) Math 408, Lecture 9-10 February
More informationWooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics
Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics A short review of the principles of mathematical statistics (or, what you should have learned in EC 151).
More informationSampling Distributions and Asymptotics Part II
Sampling Distributions and Asymptotics Part II Insan TUNALI Econ 511 - Econometrics I Koç University 18 October 2018 I. Tunal (Koç University) Week 5 18 October 2018 1 / 29 Lecture Outline v Sampling Distributions
More informationSDS : Theoretical Statistics
SDS 384 11: Theoretical Statistics Lecture 1: Introduction Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin https://psarkar.github.io/teaching Manegerial Stuff
More informationLecture 8 Inequality Testing and Moment Inequality Models
Lecture 8 Inequality Testing and Moment Inequality Models Inequality Testing In the previous lecture, we discussed how to test the nonlinear hypothesis H 0 : h(θ 0 ) 0 when the sample information comes
More informationNotes on Time Series Modeling
Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g
More information8 Periodic Linear Di erential Equations - Floquet Theory
8 Periodic Linear Di erential Equations - Floquet Theory The general theory of time varying linear di erential equations _x(t) = A(t)x(t) is still amazingly incomplete. Only for certain classes of functions
More information2 Interval-valued Probability Measures
Interval-Valued Probability Measures K. David Jamison 1, Weldon A. Lodwick 2 1. Watson Wyatt & Company, 950 17th Street,Suite 1400, Denver, CO 80202, U.S.A 2. Department of Mathematics, Campus Box 170,
More informationEstimators as Random Variables
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until
More informationECON0702: Mathematical Methods in Economics
ECON0702: Mathematical Methods in Economics Yulei Luo SEF of HKU January 12, 2009 Luo, Y. (SEF of HKU) MME January 12, 2009 1 / 35 Course Outline Economics: The study of the choices people (consumers,
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationCentral limit theorem. Paninski, Intro. Math. Stats., October 5, probability, Z N P Z, if
Paninski, Intro. Math. Stats., October 5, 2005 35 probability, Z P Z, if P ( Z Z > ɛ) 0 as. (The weak LL is called weak because it asserts convergence in probability, which turns out to be a somewhat weak
More informationGaussian, Markov and stationary processes
Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November
More informationIn manycomputationaleconomicapplications, one must compute thede nite n
Chapter 6 Numerical Integration In manycomputationaleconomicapplications, one must compute thede nite n integral of a real-valued function f de ned on some interval I of
More informationEconomics 241B Estimation with Instruments
Economics 241B Estimation with Instruments Measurement Error Measurement error is de ned as the error resulting from the measurement of a variable. At some level, every variable is measured with error.
More informationSYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions
SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu
More informationLarge Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n
Large Sample Theory In statistics, we are interested in the properties of particular random variables (or estimators ), which are functions of our data. In ymptotic analysis, we focus on describing the
More informationECONOMET RICS P RELIM EXAM August 19, 2014 Department of Economics, Michigan State University
ECONOMET RICS P RELIM EXAM August 19, 2014 Department of Economics, Michigan State University Instructions: Answer all ve (5) questions. Be sure to show your work or provide su cient justi cation for your
More information12. Perturbed Matrices
MAT334 : Applied Linear Algebra Mike Newman, winter 208 2. Perturbed Matrices motivation We want to solve a system Ax = b in a context where A and b are not known exactly. There might be experimental errors,
More informationSIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011 Revised March 2012
SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER By Donald W. K. Andrews August 2011 Revised March 2012 COWLES FOUNDATION DISCUSSION PAPER NO. 1815R COWLES FOUNDATION FOR
More informationNonparametric Estimation of Wages and Labor Force Participation
Nonparametric Estimation of Wages Labor Force Participation John Pepper University of Virginia Steven Stern University of Virginia May 5, 000 Preliminary Draft - Comments Welcome Abstract. Model Let y
More information1 The Well Ordering Principle, Induction, and Equivalence Relations
1 The Well Ordering Principle, Induction, and Equivalence Relations The set of natural numbers is the set N = f1; 2; 3; : : :g. (Some authors also include the number 0 in the natural numbers, but number
More information11. Further Issues in Using OLS with TS Data
11. Further Issues in Using OLS with TS Data With TS, including lags of the dependent variable often allow us to fit much better the variation in y Exact distribution theory is rarely available in TS applications,
More informationChapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics
Chapter 6 Order Statistics and Quantiles 61 Extreme Order Statistics Suppose we have a finite sample X 1,, X n Conditional on this sample, we define the values X 1),, X n) to be a permutation of X 1,,
More informationProbabilities & Statistics Revision
Probabilities & Statistics Revision Christopher Ting Christopher Ting http://www.mysmu.edu/faculty/christophert/ : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 January 6, 2017 Christopher Ting QF
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationConvergence in Distribution
Convergence in Distribution Undergraduate version of central limit theorem: if X 1,..., X n are iid from a population with mean µ and standard deviation σ then n 1/2 ( X µ)/σ has approximately a normal
More informationMeasuring robustness
Measuring robustness 1 Introduction While in the classical approach to statistics one aims at estimates which have desirable properties at an exactly speci ed model, the aim of robust methods is loosely
More informationIntroduction to Bases in Banach Spaces
Introduction to Bases in Banach Spaces Matt Daws June 5, 2005 Abstract We introduce the notion of Schauder bases in Banach spaces, aiming to be able to give a statement of, and make sense of, the Gowers
More informationCoordinate Systems. S. F. Ellermeyer. July 10, 2009
Coordinate Systems S F Ellermeyer July 10, 009 These notes closely follow the presentation of the material given in David C Lay s textbook Linear Algebra and its Applications (rd edition) These notes are
More informationUniversity of Regina. Lecture Notes. Michael Kozdron
University of Regina Statistics 252 Mathematical Statistics Lecture Notes Winter 2005 Michael Kozdron kozdron@math.uregina.ca www.math.uregina.ca/ kozdron Contents 1 The Basic Idea of Statistics: Estimating
More informationProbability and Measure
Chapter 4 Probability and Measure 4.1 Introduction In this chapter we will examine probability theory from the measure theoretic perspective. The realisation that measure theory is the foundation of probability
More informationLecture 2: Convergence of Random Variables
Lecture 2: Convergence of Random Variables Hyang-Won Lee Dept. of Internet & Multimedia Eng. Konkuk University Lecture 2 Introduction to Stochastic Processes, Fall 2013 1 / 9 Convergence of Random Variables
More informationComplex Analysis Slide 9: Power Series
Complex Analysis Slide 9: Power Series MA201 Mathematics III Department of Mathematics IIT Guwahati August 2015 Complex Analysis Slide 9: Power Series 1 / 37 Learning Outcome of this Lecture We learn Sequence
More informationLinear regression COMS 4771
Linear regression COMS 4771 1. Old Faithful and prediction functions Prediction problem: Old Faithful geyser (Yellowstone) Task: Predict time of next eruption. 1 / 40 Statistical model for time between
More informationa11 a A = : a 21 a 22
Matrices The study of linear systems is facilitated by introducing matrices. Matrix theory provides a convenient language and notation to express many of the ideas concisely, and complicated formulas are
More informationStatistical inference
Statistical inference Contents 1. Main definitions 2. Estimation 3. Testing L. Trapani MSc Induction - Statistical inference 1 1 Introduction: definition and preliminary theory In this chapter, we shall
More informationEIGENVALUES AND EIGENVECTORS 3
EIGENVALUES AND EIGENVECTORS 3 1. Motivation 1.1. Diagonal matrices. Perhaps the simplest type of linear transformations are those whose matrix is diagonal (in some basis). Consider for example the matrices
More informationLecture 4: September Reminder: convergence of sequences
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 4: September 6 In this lecture we discuss the convergence of random variables. At a high-level, our first few lectures focused
More informationAsymptotic Statistics-VI. Changliang Zou
Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous
More informationnotes 5d. Details of the Construction of the Smale horseshoe map.
by 3940-07 notes 5d Details of the Construction of the Smale horseshoe map. Let f (x; y) = ( 4 x + 8 ; 5y ). Then f maps D onto the rectangle R de ned 8 x 3 8 y 9 : In particular, it squeezes D horizontally
More information8 Laws of large numbers
8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable
More informationParameter Estimation
Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y
More informationThe main purpose of this chapter is to prove the rst and second fundamental theorem of asset pricing in a so called nite market model.
1 2. Option pricing in a nite market model (February 14, 2012) 1 Introduction The main purpose of this chapter is to prove the rst and second fundamental theorem of asset pricing in a so called nite market
More informationStochastic Models (Lecture #4)
Stochastic Models (Lecture #4) Thomas Verdebout Université libre de Bruxelles (ULB) Today Today, our goal will be to discuss limits of sequences of rv, and to study famous limiting results. Convergence
More informationQualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama
Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Instructions This exam has 7 pages in total, numbered 1 to 7. Make sure your exam has all the pages. This exam will be 2 hours
More informationGaussian vectors and central limit theorem
Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables
More informationNonlinear Programming (NLP)
Natalia Lazzati Mathematics for Economics (Part I) Note 6: Nonlinear Programming - Unconstrained Optimization Note 6 is based on de la Fuente (2000, Ch. 7), Madden (1986, Ch. 3 and 5) and Simon and Blume
More informationADJUSTED POWER ESTIMATES IN. Ji Zhang. Biostatistics and Research Data Systems. Merck Research Laboratories. Rahway, NJ
ADJUSTED POWER ESTIMATES IN MONTE CARLO EXPERIMENTS Ji Zhang Biostatistics and Research Data Systems Merck Research Laboratories Rahway, NJ 07065-0914 and Dennis D. Boos Department of Statistics, North
More informationStochastic Processes
Introduction and Techniques Lecture 4 in Financial Mathematics UiO-STK4510 Autumn 2015 Teacher: S. Ortiz-Latorre Stochastic Processes 1 Stochastic Processes De nition 1 Let (E; E) be a measurable space
More informationEconomics 620, Lecture 18: Nonlinear Models
Economics 620, Lecture 18: Nonlinear Models Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 18: Nonlinear Models 1 / 18 The basic point is that smooth nonlinear
More information1. The Multivariate Classical Linear Regression Model
Business School, Brunel University MSc. EC550/5509 Modelling Financial Decisions and Markets/Introduction to Quantitative Methods Prof. Menelaos Karanasos (Room SS69, Tel. 08956584) Lecture Notes 5. The
More information1 Which sets have volume 0?
Math 540 Spring 0 Notes #0 More on integration Which sets have volume 0? The theorem at the end of the last section makes this an important question. (Measure theory would supersede it, however.) Theorem
More informationRegression #4: Properties of OLS Estimator (Part 2)
Regression #4: Properties of OLS Estimator (Part 2) Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #4 1 / 24 Introduction In this lecture, we continue investigating properties associated
More informationCS 540: Machine Learning Lecture 2: Review of Probability & Statistics
CS 540: Machine Learning Lecture 2: Review of Probability & Statistics AD January 2008 AD () January 2008 1 / 35 Outline Probability theory (PRML, Section 1.2) Statistics (PRML, Sections 2.1-2.4) AD ()
More informationIntroduction to Econometrics
Introduction to Econometrics Michael Bar October 3, 08 San Francisco State University, department of economics. ii Contents Preliminaries. Probability Spaces................................. Random Variables.................................
More informationA6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2011
A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2011 Reading Chapter 5 (continued) Lecture 8 Key points in probability CLT CLT examples Prior vs Likelihood Box & Tiao
More informationNotes on Mathematical Expectations and Classes of Distributions Introduction to Econometric Theory Econ. 770
Notes on Mathematical Expectations and Classes of Distributions Introduction to Econometric Theory Econ. 77 Jonathan B. Hill Dept. of Economics University of North Carolina - Chapel Hill October 4, 2 MATHEMATICAL
More information5 Birkhoff s Ergodic Theorem
5 Birkhoff s Ergodic Theorem Birkhoff s Ergodic Theorem extends the validity of Kolmogorov s strong law to the class of stationary sequences of random variables. Stationary sequences occur naturally even
More informationWidely applicable periodicity results for higher order di erence equations
Widely applicable periodicity results for higher order di erence equations István Gy½ori, László Horváth Department of Mathematics University of Pannonia 800 Veszprém, Egyetem u. 10., Hungary E-mail: gyori@almos.uni-pannon.hu
More information14.30 Introduction to Statistical Methods in Economics Spring 2009
MIT OpenCourseWare http://ocw.mit.edu 4.0 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationLecture Notes Part 2: Matrix Algebra
17.874 Lecture Notes Part 2: Matrix Algebra 2. Matrix Algebra 2.1. Introduction: Design Matrices and Data Matrices Matrices are arrays of numbers. We encounter them in statistics in at least three di erent
More information