V. Properties of estimators {Parts C, D & E in this file}

Size: px
Start display at page:

Download "V. Properties of estimators {Parts C, D & E in this file}"

Transcription

1 A. Definitions & Desiderata. model. estimator V. Properties of estimators {Parts C, D & E in this file}. sampling errors and sampling distribution 4. unbiasedness 5. low sampling variance 6. low mean squared error (MSE) i. definition of MSE ii. implied loss function on errors iii. relationship to variance and bias 7. evocative simple examples B. Finite sample (exact) properties (also called small sample properties, altho equally valid. relative efficiency for large samples). best linear unbiasedness (BLUness). efficiency a. definition and relationship to BLUness b. [SKIP] efficiency using Cramer-Rao lower bounds (Part ) - what the bounds mean - what must be assumed in order to calculate the lower bounds - how the bounds can show (or fail to show) efficiency c. [SKIP] likelihood function of random sample - motivation - gaussian case d. [SKIP] showing efficiency using Cramer-Rao lower bounds (Part ) - Cramer-Rao result (how to calculate the lower bounds in and parameter cases) - example: efficiency of y when ~ IID[µ, ] & known (one-parameter case)

2 - example: efficiency of y when ~ IID[µ, ] & unknown (two-parameter case) C. Large Sample or Asymptotic properties. iting distribution vs. it of sampling distribution. asymptotic variance. asymptotic mean & asymptotic unbiasedness 4. consistency i. Consistency through vanishing it of MSE (Squared Error Consistency) ii. probabilitt (p) 5. asymptotic efficiency i. definition - definition & relation to consistency - theorems for manipulating/evaluating probabilitts - illustrative examples ii. [SKIP] producing asymptotically efficient estimators (maximum likelihood estimation) iii. [SKIP] an astronomically terrible asymptotically efficient estimator D. Minimum MSE linear estimator E. The Zen of Econometrics

3 C. Large Sample or Asymptotic Properties. Limiting distribution vs. Limit of distribution Economic data is often not IID (nor even IID) because it is usually not the result of a controlled experiment. Consequently, it is often difficult or impossible to obtain estimators which we can show are either efficient or BLU and we must resort to evaluating the "goodness" of our estimators by means of what are called "large sample" or "asymptotic" properties. These properties are meaningful only when the actual sample is sufficiently large. Unfortunately, in practice it is not always clear just how large is "large enough." evertheless, since large sample properties are often all we can get, they are much prized. Suppose that y ~ IID(µ, i ) for i =.... Then we know that ȳ µ, using all observations and ỹ µ, using only the first observations ~ s obviously preferable to y for any finite sample size, but notice that the its of the sampling distributions of these two estimators as the sample size becomes arbitrarily large ~ are identical for an infinitely large sample, both y and y are distributed (µ, )! (I.e., the density function is a spike centered on µ with zero width and unit area.) In fact, because var practically any estimator constant terms that become negligible compared to as Since the area under a density function must equal one, this function must have infinite height.

4 the it of the sampling distribution of practically every estimator is a spike with zero width and unit area. Consequently, the it of the sampling distribution is not a useful largesample concept. C.. Asymptotic variance However, re-writing the last equation more explicitly, var a terms that become negligible compared to as it then follows from var(cx) = c var(x) that var a terms that become negligible compared to as so if we take the it of this as and then divide by we get an expression for how fast the sampling variance of is going to zero when the sample size is large. This is called "the asymptotic variance of " or Avar( ) or Asyvar( ): Avar var a terms that become negligible compared to as a a leading term in var() ~ otice that, even though the it of var(y) = it of var(y) =, these estimators have asymptotic variances of / and /, respectively.

5 Therefore we define the "asymptotic distribution of "or "iting distribution of " as the distribution that "tends toward" as the sample size (while finite) becomes arbitrarily large. By definition, this iting distribution will have variance equal to the Avar( ). The mean of the iting distribution of is called the asymptotic mean of or AE( ) or AsyE( ). Its definition is straightforward: AE and hence Abias ote #: Asymptotic unbiasedness is a substantially weaker property than "ordinary" or "finite-sample" unbiasedness. For example, Abias Abias 6 ote #: Since virtually all of the estimators ordinarily considered are (at least asymptotically) unbiased, there is usually no need to distinguish between the following three definitions of Avar( ), all of which appear in the literature: Avar E E or Avar E AE or Avar E E even though the last of these three definitions would more properly be called the asymptotic MSE( ). In any case, the distinction between these three alternatives is certainly not important for what we are doing; consequently, we will use the last of the three, since it is a bit simpler.

6 C.4. Consistency The "probabilitt" or "p" of an estimator is a formalization of the intuitively appealing notion that the sampling density function of (practically) all estimators "spikes out" (becomes a spike with unit area and zero width) as the sample size goes to infinity. The p of the estimator is just the value over which the spike sits; we say that the estimator is "consistent" if its probabilitt equals the population value that the estimator is trying to estimate. It turns out to be extraordinarily useful to formalize this notion of a probabilitt because the formal definition leads to a number of results which (while not easy to prove) are both intuitively appealing and extremely easy to apply. Consequently, it is usually possible to compute the probabilitt of one's estimator (and to thereby determine whether or not it is consistent) even in the messy circumstances we often face with (typically) non-experimental economic data, such as data which is not only non-gaussian, but also not iid. (In other words, people place exceptional value on consistencn their estimators because it is often the only property they can get!) C.4.i. Consistency through vanishing it of MSE (Squared Error Consistency) First, a way to prove consistency that does not require dealing with probabilitts: is (squared error) consistent if and onlf a. MSE( ) = E{[ - ] } as or (equivalently) b. Avar( ) as and Abias( ) = This kind of consistencs called "squared error consistency" to distinguish it from the definition of consistency based on probabilitts. The two forms of consistency are equivalent except in the case of an estimator whose sampling density function has tails which are so thick that the estimator does not have a finite variance or MSE; such an estimator can be consistent (based on its probabilitt being correct) even though it cannot be squared error consistent because its MSE is not well-defined for finite samples.

7 C.4.ii. Probability Limits (p) Definition of p: * p( ) = if and onlf * a. (verbal) The sampling density function of becomes a spike centered on value as the sample size goes to infinity. b. formal probability that lies in, no matter how small > is Definition of Consistency: is consistent for if and onlf p( ) = Theorems for manipulating ps:. (Slutsky's Theorem) If g( ) is a continuous function, then p{ g(x) } = g( p{ X } ).. If f() is not random, then p{ f() } = f() where "" denotes an ordinart as. I.e. the it of g(x) as x approaches z is g(z) for all values of z.

8 . So long as p{ X } and p{ Y } exist, p{ X + Y } = p{ X } + p{ Y }. 4. So long as p{ X } and p{ Y } exist, p{ X * Y } = p{ X } * p{ Y }. 5. So long as p{ X } and p{ Y } exist and p{ Y } is not zero, p{ X / Y } = p{ X } / p{ Y }. 6. (Khintchine's Theorem) p{ sample moment } = corresponding population moment ote #: Suppose, for example, that you have an unbiased estimate of the variance of some random variable, but what you need is an estimate of the variable's standard deviation. The square root of the unbiased variance estimator is not unbiased for the square root of the population variance. But slutsky's theorem guarantees that the square root of a consistent estimator of the variance is a consistent estimator of the square root of the population variance. ote # Strategy for evaluating probabilitts: a. write your estimator as a function sample moments b. use theorems # - 5 to write the p of the estimator as a function of ps of sample moments c. use Khintchine's theorem to evaluate the ps of the sample moments.

9 Illustrative Examples: ~ a. p(y) = p(y) = µ {This follows directly from Khintchine's theorem since these are sample moments, using all of the sample data and the first half of the sample data, respectively.} Since in this case we know the sampling distributions of the estimators, it is also easy to show squared error consistency. For example: ỹ µ, using only the first observations therefore MSE ỹ var ỹ bias ỹ b. Assuming that is a fixed constant and that p( ) exists, p p p p p

10 c. Assuming that p( ) exists, p p p p p 4 Therefore, if is consistent for, then so is + ( /)!! C.5 Asymptotic efficiency i. Definition: is asymptotically efficient if and onlf and. is squared error consistent (which implies that Abias( ) = ) ~ ~. Avar( ) Avar( ) for all squared error consistent. 4. And so is + ( / )

11 C.5.ii. [SKIP] Producing asymptotically efficient estimators: Maximum Likelihood Estimation Asymptotic efficiencs the most desirable large-sample property an estimator can have. Indeed, if the sample is large enough for it to be credibly meaningful, asymptotic efficiencs a wonderful property. Unfortunately, just like finite-sample efficiency, it looks like asymptotic efficiency will be difficult to prove. Remarkably enough, that turns out not to be the case because it turns out that maximum likelihood estimators are all asymptotically efficient. First I will motivate our interest in maximum likelihood estimation (MLE) by discussing the nice (asymptotic) properties that MLE estimators have; then I will show you how to obtain MLE estimators. If the model includes enough distributional information that we can express the likelihood of the observed sample as a function of the sample data and a finite number of unknown parameters (such as µ, or,,..., k ), then the unknown parameters can be estimated using the maximum likelihood method and the resulting estimators have the following nice asymptotic properties: 5 these Cramer- a. They are asymptotically unbiased. b. They are consistent. c. They are asymptotically efficient. d. They are asymptotically normal. ML i e. g( ) is the maximum likelihood estimator of g( ) {and hence has all nice asymptotic properties as an estimator of g( ) } for any continuous i function g(x). ML ML f. In fact is asymptotically efficient because Avar( ) equals the i Rao lower bound. Therefore, calculation of the k x k matrix of i i 5 The proof of these results also requires that the densits a continuous function of the observations and that the log-likelihood function be a sufficiently smooth function of the unknown parameters that its third partial derivatives with respect to these parameters are all finite. These conditions are all satisfied for gaussian data.

12 Cramer-Rao lower bounds for,,..., k provides a straightforward way to obtain V, the k x k (asymptotic) variance- ML ML ML covariance matrix for,,...,. k Thus, ML i i ii asymptotically ~ [, ] where of i ii for k=): is the Cramer-Rao lower bound on the sampling variance of any unbiased estimator, obtained as the ith diagonal element of the inverse of the k x k matrix {expressed below E L,, E L,, E L,, E L,, E L,, E L,, E L,, E L,, E L,, where L (,, ) is a contraction for the log-likelihood function, log{ likelihood(y, y,..., y ;,, ) } The calculation of this matrix (and its inverse) can be tedious, but at least it is straightforward.

13 So maximum likelihood estimation is rather wonderful if one is willing to settle for ML ML ML asymptotic properties and if one can actually obtain,,...,. How do you obtain these estimators? k First of all, recall that the likelihood function, likelihood(y, y,..., y ;,,... ) k takes,,... (the parameters specifying the distribution of y ) as given and expresses k i the likelihood of having observed y, y,..., y (the sample observations) as a function of y, y,..., y. Suppose that we turned this around in our mind and instead viewed the likelihood function as a function of our estimates (,,..., ) of the (unknown) k parameters (,,..., ) given that we have observed the sample (y, y,..., y ): k likelihood(,,..., k ; y, y,..., y ) ow this likelihood function expresses how likelt is that the underlying parameters (,,... k ) equal the particular values (,,..., k ) given that we have observed the sample (y, y,..., y ). The maximum likelihood estimates of (,,... k ) are thus the values of (,,..., ) that maximize k likelihood(,,..., k ; y, y,..., y ) for the given observation values, (y, y,..., y ). Recall our earlier example in which we had three observations ( y, y, y ) and we assumed that ~ IID(, ) for i =... This led to the likelihood function Likelihood y, y, y ;, e y e y e y

14 At that point we assumed that var(y) = = and that E{y} = = 8 and went on to evaluate likelihood(y, y, y ; 8, ) for various possible samples, such as (8,8,8) and (8,,8). To estimate and using the maximum likelihood method, we instead take y, y, and y as given, consider the likelihood function Likelihood, ;y, y, y e y e y e y and ask questions like "given that we have observed (8., 7.7, 8.5), how likels it that was and was in the distribution from which these observations were picked?" The values of and maximize the likelihood of having observed the sample we did observe are the maximum likelihood estimates of and. Since the slope of the logarithm function is always positive, the pair of values ( and ) that maximizes likelihood(, ; y, y, y ) also maximizes log{ likelihood(, ; y, y, y ) } which, in this case { y ~ IID(, ) for i =... } equals i

15 L, ; y, y, y i log f ;, i log log i log log i ML ML Therefore the maximum likelihood estimatators of, of = µ and = must satisfy the two first order conditions for maximizing log{ likelihood(, ; y, y, y ) } with respect to,, namely: L, ; y, y, y log log i I L, ; y, y, y log log i II Equation I yields:

16 L, ; y, y, y log log i i i i i i ȳ so that, since the variance estimate is surely positive, ML ȳ Substituting this result into equation II and letting

17 i yields L, ; y, y, y log log i log log oting that d log(x)/dx = /x and that d(/x)/dx = -/x, this yields L, ; y, y, y log Multiplying both sides of this equation by -( )/ yields ML i i ȳ Thus, we can conclude that,

18 ML ȳ the sample mean ML the sample variance othing was special about a sample size of three, so these results are valid for any sample size. Therefore, if y ~ IID(, ) for i =..., then y and are both i and. consistent. asymptotically unbiased. asymptotically normal 4. aymptotically efficient. This is no news for y, since it is (for ~ IID(, ) for i =...) efficient even in finite samples, but it is a useful result for. Moreover, since maximum likelihood estimators asymptotically attain the Cramer-Rao lower bounds, we can use these to complete the specification of the joint (asymptotic) density function for this pair of estimators. Using the matrix of Cramer-Rao lower bounds for this ML ML problem calculated in section.c.v above, the estimators (, ) = (y, ) are thus asymptotically distributed ȳ µ asy ~, 4 We see then (for y ~ IID(, ) for i =... ) that y and are asymptoticallndependent (since i they are asymptotically normal and asymptotically uncorrelated), but we knew this already from the fact that y and are in this case independently distributed for any sample size. We can also conclude that

19 ~ asy, 4 On your homework assignment you will use the known result that var(s ) = 4 /(-) to show that the var( ) < 4 / for any sample size. But, since we just saw that 4 / is the Cramer-Rao lower bound, how can have a finite sample variance smaller than that? (Answer: as we showed earlier, is a biased estimator for.) Recall that C.5.iii [SKIP] An astronomically terrible (but asymptotically efficient) estimator Avar E 8 and suppose that is known to be asymptotically efficient. What about + ( /)? First note that, since is to be asymptotically efficient, it is asymptotically unbiased; consequently, 8 + ( /) is asymptotically unbiased also: Abias 8 E 8 E 8 E 8 E 8 Abias

20 8 ext note that both and + ( /) in fact have the same asymptotic variance: Avar 8 E 8 E 8 E 8 6 E E E E E 6 8 Bias 6 E 8 Bias 6 Avar 8 Abias Avar 8 So, if is asymptotically efficient (and hence asymptotically unbiased) then so is + ( /)!! (.5 + ) In fact, a close look at this derivation shows that and [ + / ] both have the same asymptotic

21 (.5 + ) variance (for > ) no matter how large the constant is, as do and [ + / ]. Thus, since s s, we see that and s have the same asymptotic variance. 6 D. Minimum MSE estimator If one's loss function on estimation errors is not proportional to the square of the error, then one might prefer some other estimator over the efficient estimator. But suppose that MSE realls what one cares about, is it possible to find an estimator with a smaller MSE than that of the efficient estimator? It turns out that the answer to this question is both "yes" and "no": Let be the efficient estimator for and let k k & t var ~ ~ where (k) is an alternative estimator of and k is some positive number. What value of k yields the (k) with the smallest MSE? bias k E k E k k E k k var k var k k var 6 Indeed, they thus have the same asymptotic distribution.

22 Thus, MSE k var k bias k k var k k var k var var k var k var var k var k var t var k k t

23 ~ Thus k*, the value of k which minimizes the MSE{ (k)}, must satisfy dmse k dk k k d dk var k k t var d dk k k t var k k t var k k t var t k t which implies that k t t <. Therefore, it is always optimal (i.e. it lowers the MSE) to bias the estimator a bit toward zero. However, it ~ * * must be noted that (k ) is not in fact a feasible estimator since the expression for k involves t which depends on.

24

25 E. The Zen of Econometrics We have now gone to considerable trouble to see that s a good (in some senses, optimal) estimator of µ when y ~ IID(µ, i ). But what if our actual sample observations are clearly not gaussian? For example, consider the problem of estimating the mean size of a U.S. firm, as quantified bts annual sales revenues. The empirically observed distribution of firm sizes looks more like the graph of a density function than it does like the graph of a gaussian density function: However, since the size of a firm can be viewed as the product of a large number of (more or less) independent factors, the empirically observed distribution of the logarithm of the firm sizes looks reasonably gaussian. firm: or Therefore there are two different ways to approach the problem of estimating the mean size of a U.S.. Try to figure out the most efficient estimator for the (population) mean of a random sample drawn from a log-normally distributed population. Then one must still face the problem of calculating its sampling distribution so that you can do inference. These problems can be dealt with to some degree, but even their approximate solution is difficult and complicated.. Since it is reasonable to suppose that the observations on log(firm sales revenue) are IID, use the sample mean of the logarithms of the observed sales revenue figures to efficiently estimate the population mean of the logarithm of firm sales revenue. Similarly, use the methods we have covered to obtain a 95% confidence interval for the population mean of the logarithm of firm sales revenue. (This interval will be centered around the sample mean of the logarithms of the observed sales revenue data.) Letting "c95upper" and "c95lower" denote the upper and lower endpoints of this 95% confidence interval for the population mean of the logarithm of firm sales revenue, one can c95lower c95upper immediatelnfer that the interval [ e, e ] is a 95% confidence interval for the population

26 mean of the of sales revenue itself. (This interval will not be centered around the exponential of the sample mean of the logarithms of the observed sales revenue data, however.) Or, more reasonably, one can and, I would argue, typically should decide that the data itself is telling us that the size of the firm ought to be quantified by the logarithm of its annual sales revenues, in which case the sample mean of the logarithms of the observed sales revenue figures is precisely the estimator one wants. This example suggests a completely different way of looking at the theoretical results we have obtained. Instead of seeing them as arbitrary special cases where everything worked out nicely and resolving to always assume that our data is IID (whether it is or not) so that we can "use" these results, this example suggests that we could more gracefully observe that our statistical machinery for using data to obtain knowledge about the world "works" best (i.e., most simply and effectively) when we have framed the problem in such a way that the sample data is IID and act accordingly.

Statistical inference

Statistical inference Statistical inference Contents 1. Main definitions 2. Estimation 3. Testing L. Trapani MSc Induction - Statistical inference 1 1 Introduction: definition and preliminary theory In this chapter, we shall

More information

A General Overview of Parametric Estimation and Inference Techniques.

A General Overview of Parametric Estimation and Inference Techniques. A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying

More information

STATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN

STATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN Massimo Guidolin Massimo.Guidolin@unibocconi.it Dept. of Finance STATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN SECOND PART, LECTURE 2: MODES OF CONVERGENCE AND POINT ESTIMATION Lecture 2:

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Principles of Statistical Inference Recap of statistical models Statistical inference (frequentist) Parametric vs. semiparametric

More information

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics A short review of the principles of mathematical statistics (or, what you should have learned in EC 151).

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

ECE531 Lecture 10b: Maximum Likelihood Estimation

ECE531 Lecture 10b: Maximum Likelihood Estimation ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So

More information

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn!

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Questions?! C. Porciani! Estimation & forecasting! 2! Cosmological parameters! A branch of modern cosmological research focuses

More information

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics)

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Probability quantifies randomness and uncertainty How do I estimate the normalization and logarithmic slope of a X ray continuum, assuming

More information

Notes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770

Notes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770 Notes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770 Jonathan B. Hill Dept. of Economics University of North Carolina - Chapel Hill November

More information

Parameter Estimation

Parameter Estimation Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y

More information

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind

More information

ECON The Simple Regression Model

ECON The Simple Regression Model ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In

More information

Advanced Signal Processing Introduction to Estimation Theory

Advanced Signal Processing Introduction to Estimation Theory Advanced Signal Processing Introduction to Estimation Theory Danilo Mandic, room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 6: Bias and variance (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 49 Our plan today We saw in last lecture that model scoring methods seem to be trading off two different

More information

Large Sample Properties of Estimators in the Classical Linear Regression Model

Large Sample Properties of Estimators in the Classical Linear Regression Model Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in

More information

ECE 275A Homework 7 Solutions

ECE 275A Homework 7 Solutions ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Slope Fields: Graphing Solutions Without the Solutions

Slope Fields: Graphing Solutions Without the Solutions 8 Slope Fields: Graphing Solutions Without the Solutions Up to now, our efforts have been directed mainly towards finding formulas or equations describing solutions to given differential equations. Then,

More information

Inference with Simple Regression

Inference with Simple Regression 1 Introduction Inference with Simple Regression Alan B. Gelder 06E:071, The University of Iowa 1 Moving to infinite means: In this course we have seen one-mean problems, twomean problems, and problems

More information

Fitting a Straight Line to Data

Fitting a Straight Line to Data Fitting a Straight Line to Data Thanks for your patience. Finally we ll take a shot at real data! The data set in question is baryonic Tully-Fisher data from http://astroweb.cwru.edu/sparc/btfr Lelli2016a.mrt,

More information

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM Subject Business Economics Paper No and Title Module No and Title Module Tag 8, Fundamentals of Econometrics 3, The gauss Markov theorem BSE_P8_M3 1 TABLE OF CONTENTS 1. INTRODUCTION 2. ASSUMPTIONS OF

More information

Parametric Models: from data to models

Parametric Models: from data to models Parametric Models: from data to models Pradeep Ravikumar Co-instructor: Manuela Veloso Machine Learning 10-701 Jan 22, 2018 Recall: Model-based ML DATA MODEL LEARNING MODEL MODEL INFERENCE KNOWLEDGE Learning:

More information

Part 4: Multi-parameter and normal models

Part 4: Multi-parameter and normal models Part 4: Multi-parameter and normal models 1 The normal model Perhaps the most useful (or utilized) probability model for data analysis is the normal distribution There are several reasons for this, e.g.,

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

A Primer on Statistical Inference using Maximum Likelihood

A Primer on Statistical Inference using Maximum Likelihood A Primer on Statistical Inference using Maximum Likelihood November 3, 2017 1 Inference via Maximum Likelihood Statistical inference is the process of using observed data to estimate features of the population.

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Markus Haas LMU München Summer term 2011 15. Mai 2011 The Simple Linear Regression Model Considering variables x and y in a specific population (e.g., years of education and wage

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Cosmological parameters A branch of modern cosmological research focuses on measuring

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 12: Frequentist properties of estimators (v4) Ramesh Johari ramesh.johari@stanford.edu 1 / 39 Frequentist inference 2 / 39 Thinking like a frequentist Suppose that for some

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)

More information

Estimation Theory Fredrik Rusek. Chapters

Estimation Theory Fredrik Rusek. Chapters Estimation Theory Fredrik Rusek Chapters 3.5-3.10 Recap We deal with unbiased estimators of deterministic parameters Performance of an estimator is measured by the variance of the estimate (due to the

More information

Lecture 3: Introduction to Complexity Regularization

Lecture 3: Introduction to Complexity Regularization ECE90 Spring 2007 Statistical Learning Theory Instructor: R. Nowak Lecture 3: Introduction to Complexity Regularization We ended the previous lecture with a brief discussion of overfitting. Recall that,

More information

Masters Comprehensive Examination Department of Statistics, University of Florida

Masters Comprehensive Examination Department of Statistics, University of Florida Masters Comprehensive Examination Department of Statistics, University of Florida May 6, 003, 8:00 am - :00 noon Instructions: You have four hours to answer questions in this examination You must show

More information

Estimating the Mean of a Normally Distributed Random Variable

Estimating the Mean of a Normally Distributed Random Variable CH03 11/19/2011 15:40:41 Page 46 3 Estimating the Mean of a ormally Distributed Random Variable 3.1 ITRODUCTIO As noted in Chapter 2, the two most important attributes of a random variable are its population

More information

Bias Variance Trade-off

Bias Variance Trade-off Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]

More information

Machine Learning Basics: Maximum Likelihood Estimation

Machine Learning Basics: Maximum Likelihood Estimation Machine Learning Basics: Maximum Likelihood Estimation Sargur N. srihari@cedar.buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics 1. Learning

More information

Ridge regression. Patrick Breheny. February 8. Penalized regression Ridge regression Bayesian interpretation

Ridge regression. Patrick Breheny. February 8. Penalized regression Ridge regression Bayesian interpretation Patrick Breheny February 8 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/27 Introduction Basic idea Standardization Large-scale testing is, of course, a big area and we could keep talking

More information

Introduction to Maximum Likelihood Estimation

Introduction to Maximum Likelihood Estimation Introduction to Maximum Likelihood Estimation Eric Zivot July 26, 2012 The Likelihood Function Let 1 be an iid sample with pdf ( ; ) where is a ( 1) vector of parameters that characterize ( ; ) Example:

More information

ARIMA Modelling and Forecasting

ARIMA Modelling and Forecasting ARIMA Modelling and Forecasting Economic time series often appear nonstationary, because of trends, seasonal patterns, cycles, etc. However, the differences may appear stationary. Δx t x t x t 1 (first

More information

Math 5a Reading Assignments for Sections

Math 5a Reading Assignments for Sections Math 5a Reading Assignments for Sections 4.1 4.5 Due Dates for Reading Assignments Note: There will be a very short online reading quiz (WebWork) on each reading assignment due one hour before class on

More information

September Math Course: First Order Derivative

September Math Course: First Order Derivative September Math Course: First Order Derivative Arina Nikandrova Functions Function y = f (x), where x is either be a scalar or a vector of several variables (x,..., x n ), can be thought of as a rule which

More information

ECON 4160, Autumn term Lecture 1

ECON 4160, Autumn term Lecture 1 ECON 4160, Autumn term 2017. Lecture 1 a) Maximum Likelihood based inference. b) The bivariate normal model Ragnar Nymoen University of Oslo 24 August 2017 1 / 54 Principles of inference I Ordinary least

More information

Asymptotic Theory. L. Magee revised January 21, 2013

Asymptotic Theory. L. Magee revised January 21, 2013 Asymptotic Theory L. Magee revised January 21, 2013 1 Convergence 1.1 Definitions Let a n to refer to a random variable that is a function of n random variables. Convergence in Probability The scalar a

More information

Inference in Regression Analysis

Inference in Regression Analysis Inference in Regression Analysis Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 4, Slide 1 Today: Normal Error Regression Model Y i = β 0 + β 1 X i + ǫ i Y i value

More information

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let

More information

Basic Concepts of Inference

Basic Concepts of Inference Basic Concepts of Inference Corresponds to Chapter 6 of Tamhane and Dunlop Slides prepared by Elizabeth Newton (MIT) with some slides by Jacqueline Telford (Johns Hopkins University) and Roy Welsch (MIT).

More information

II. An Application of Derivatives: Optimization

II. An Application of Derivatives: Optimization Anne Sibert Autumn 2013 II. An Application of Derivatives: Optimization In this section we consider an important application of derivatives: finding the minimum and maximum points. This has important applications

More information

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments /4/008 Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University C. A Sample of Data C. An Econometric Model C.3 Estimating the Mean of a Population C.4 Estimating the Population

More information

3.3 Estimator quality, confidence sets and bootstrapping

3.3 Estimator quality, confidence sets and bootstrapping Estimator quality, confidence sets and bootstrapping 109 3.3 Estimator quality, confidence sets and bootstrapping A comparison of two estimators is always a matter of comparing their respective distributions.

More information

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and

More information

557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE

557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE 557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE An estimator, T (X), of θ can be evaluated via its statistical properties. Typically, two aspects are considered: Expectation Variance either in terms

More information

Economics 573 Problem Set 5 Fall 2002 Due: 4 October b. The sample mean converges in probability to the population mean.

Economics 573 Problem Set 5 Fall 2002 Due: 4 October b. The sample mean converges in probability to the population mean. Economics 573 Problem Set 5 Fall 00 Due: 4 October 00 1. In random sampling from any population with E(X) = and Var(X) =, show (using Chebyshev's inequality) that sample mean converges in probability to..

More information

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012 Problem Set #6: OLS Economics 835: Econometrics Fall 202 A preliminary result Suppose we have a random sample of size n on the scalar random variables (x, y) with finite means, variances, and covariance.

More information

A Few Notes on Fisher Information (WIP)

A Few Notes on Fisher Information (WIP) A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties

More information

MS&E 226: Small Data. Lecture 6: Bias and variance (v2) Ramesh Johari

MS&E 226: Small Data. Lecture 6: Bias and variance (v2) Ramesh Johari MS&E 226: Small Data Lecture 6: Bias and variance (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 47 Our plan today We saw in last lecture that model scoring methods seem to be trading o two di erent

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

Parametric Techniques

Parametric Techniques Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure

More information

Estimators as Random Variables

Estimators as Random Variables Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until

More information

Problem set 1: answers. April 6, 2018

Problem set 1: answers. April 6, 2018 Problem set 1: answers April 6, 2018 1 1 Introduction to answers This document provides the answers to problem set 1. If any further clarification is required I may produce some videos where I go through

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

A Course in Applied Econometrics Lecture 18: Missing Data. Jeff Wooldridge IRP Lectures, UW Madison, August Linear model with IVs: y i x i u i,

A Course in Applied Econometrics Lecture 18: Missing Data. Jeff Wooldridge IRP Lectures, UW Madison, August Linear model with IVs: y i x i u i, A Course in Applied Econometrics Lecture 18: Missing Data Jeff Wooldridge IRP Lectures, UW Madison, August 2008 1. When Can Missing Data be Ignored? 2. Inverse Probability Weighting 3. Imputation 4. Heckman-Type

More information

High-dimensional regression

High-dimensional regression High-dimensional regression Advanced Methods for Data Analysis 36-402/36-608) Spring 2014 1 Back to linear regression 1.1 Shortcomings Suppose that we are given outcome measurements y 1,... y n R, and

More information

ter. on Can we get a still better result? Yes, by making the rectangles still smaller. As we make the rectangles smaller and smaller, the

ter. on Can we get a still better result? Yes, by making the rectangles still smaller. As we make the rectangles smaller and smaller, the Area and Tangent Problem Calculus is motivated by two main problems. The first is the area problem. It is a well known result that the area of a rectangle with length l and width w is given by A = wl.

More information

2 = = 0 Thus, the number which is largest in magnitude is equal to the number which is smallest in magnitude.

2 = = 0 Thus, the number which is largest in magnitude is equal to the number which is smallest in magnitude. Limits at Infinity Two additional topics of interest with its are its as x ± and its where f(x) ±. Before we can properly discuss the notion of infinite its, we will need to begin with a discussion on

More information

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg Statistics for Data Analysis PSI Practical Course 2014 Niklaus Berger Physics Institute, University of Heidelberg Overview You are going to perform a data analysis: Compare measured distributions to theoretical

More information

Introduction to the Mathematical and Statistical Foundations of Econometrics Herman J. Bierens Pennsylvania State University

Introduction to the Mathematical and Statistical Foundations of Econometrics Herman J. Bierens Pennsylvania State University Introduction to the Mathematical and Statistical Foundations of Econometrics 1 Herman J. Bierens Pennsylvania State University November 13, 2003 Revised: March 15, 2004 2 Contents Preface Chapter 1: Probability

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26 Estimation: Regression and Least Squares This note explains how to use observations to estimate unobserved random variables.

More information

Regression Models - Introduction

Regression Models - Introduction Regression Models - Introduction In regression models, two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent variable,

More information

Review. December 4 th, Review

Review. December 4 th, Review December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter

More information

Analogy Principle. Asymptotic Theory Part II. James J. Heckman University of Chicago. Econ 312 This draft, April 5, 2006

Analogy Principle. Asymptotic Theory Part II. James J. Heckman University of Chicago. Econ 312 This draft, April 5, 2006 Analogy Principle Asymptotic Theory Part II James J. Heckman University of Chicago Econ 312 This draft, April 5, 2006 Consider four methods: 1. Maximum Likelihood Estimation (MLE) 2. (Nonlinear) Least

More information

Today. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use?

Today. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use? Today Statistical Learning Parameter Estimation: Maximum Likelihood (ML) Maximum A Posteriori (MAP) Bayesian Continuous case Learning Parameters for a Bayesian Network Naive Bayes Maximum Likelihood estimates

More information

Discrete Dependent Variable Models

Discrete Dependent Variable Models Discrete Dependent Variable Models James J. Heckman University of Chicago This draft, April 10, 2006 Here s the general approach of this lecture: Economic model Decision rule (e.g. utility maximization)

More information

Physics 509: Bootstrap and Robust Parameter Estimation

Physics 509: Bootstrap and Robust Parameter Estimation Physics 509: Bootstrap and Robust Parameter Estimation Scott Oser Lecture #20 Physics 509 1 Nonparametric parameter estimation Question: what error estimate should you assign to the slope and intercept

More information

Multicollinearity and A Ridge Parameter Estimation Approach

Multicollinearity and A Ridge Parameter Estimation Approach Journal of Modern Applied Statistical Methods Volume 15 Issue Article 5 11-1-016 Multicollinearity and A Ridge Parameter Estimation Approach Ghadban Khalaf King Khalid University, albadran50@yahoo.com

More information

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter

More information

An overview of applied econometrics

An overview of applied econometrics An overview of applied econometrics Jo Thori Lind September 4, 2011 1 Introduction This note is intended as a brief overview of what is necessary to read and understand journal articles with empirical

More information

MATH 320, WEEK 6: Linear Systems, Gaussian Elimination, Coefficient Matrices

MATH 320, WEEK 6: Linear Systems, Gaussian Elimination, Coefficient Matrices MATH 320, WEEK 6: Linear Systems, Gaussian Elimination, Coefficient Matrices We will now switch gears and focus on a branch of mathematics known as linear algebra. There are a few notes worth making before

More information

Regression Models - Introduction

Regression Models - Introduction Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent

More information

i=1 h n (ˆθ n ) = 0. (2)

i=1 h n (ˆθ n ) = 0. (2) Stat 8112 Lecture Notes Unbiased Estimating Equations Charles J. Geyer April 29, 2012 1 Introduction In this handout we generalize the notion of maximum likelihood estimation to solution of unbiased estimating

More information

Properties of the least squares estimates

Properties of the least squares estimates Properties of the least squares estimates 2019-01-18 Warmup Let a and b be scalar constants, and X be a scalar random variable. Fill in the blanks E ax + b) = Var ax + b) = Goal Recall that the least squares

More information

Practice Problems Section Problems

Practice Problems Section Problems Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,

More information

Stat260: Bayesian Modeling and Inference Lecture Date: February 10th, Jeffreys priors. exp 1 ) p 2

Stat260: Bayesian Modeling and Inference Lecture Date: February 10th, Jeffreys priors. exp 1 ) p 2 Stat260: Bayesian Modeling and Inference Lecture Date: February 10th, 2010 Jeffreys priors Lecturer: Michael I. Jordan Scribe: Timothy Hunter 1 Priors for the multivariate Gaussian Consider a multivariate

More information

Parametric Inference

Parametric Inference Parametric Inference Moulinath Banerjee University of Michigan April 14, 2004 1 General Discussion The object of statistical inference is to glean information about an underlying population based on a

More information

Economics 583: Econometric Theory I A Primer on Asymptotics

Economics 583: Econometric Theory I A Primer on Asymptotics Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:

More information

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y

More information

Introduction to Econometrics. Heteroskedasticity

Introduction to Econometrics. Heteroskedasticity Introduction to Econometrics Introduction Heteroskedasticity When the variance of the errors changes across segments of the population, where the segments are determined by different values for the explanatory

More information

Regression of Time Series

Regression of Time Series Mahlerʼs Guide to Regression of Time Series CAS Exam S prepared by Howard C. Mahler, FCAS Copyright 2016 by Howard C. Mahler. Study Aid 2016F-S-9Supplement Howard Mahler hmahler@mac.com www.howardmahler.com/teaching

More information

Journal of Statistical Research 2007, Vol. 41, No. 1, pp Bangladesh

Journal of Statistical Research 2007, Vol. 41, No. 1, pp Bangladesh Journal of Statistical Research 007, Vol. 4, No., pp. 5 Bangladesh ISSN 056-4 X ESTIMATION OF AUTOREGRESSIVE COEFFICIENT IN AN ARMA(, ) MODEL WITH VAGUE INFORMATION ON THE MA COMPONENT M. Ould Haye School

More information

Stability and the elastic net

Stability and the elastic net Stability and the elastic net Patrick Breheny March 28 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/32 Introduction Elastic Net Our last several lectures have concentrated on methods for

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

Homoskedasticity. Var (u X) = σ 2. (23)

Homoskedasticity. Var (u X) = σ 2. (23) Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This

More information

Lecture 4: September Reminder: convergence of sequences

Lecture 4: September Reminder: convergence of sequences 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 4: September 6 In this lecture we discuss the convergence of random variables. At a high-level, our first few lectures focused

More information

Lecture 4: Types of errors. Bayesian regression models. Logistic regression

Lecture 4: Types of errors. Bayesian regression models. Logistic regression Lecture 4: Types of errors. Bayesian regression models. Logistic regression A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting more generally COMP-652 and ECSE-68, Lecture

More information

Investigation of Possible Biases in Tau Neutrino Mass Limits

Investigation of Possible Biases in Tau Neutrino Mass Limits Investigation of Possible Biases in Tau Neutrino Mass Limits Kyle Armour Departments of Physics and Mathematics, University of California, San Diego, La Jolla, CA 92093 (Dated: August 8, 2003) We study

More information

Applied Econometrics (QEM)

Applied Econometrics (QEM) Applied Econometrics (QEM) The Simple Linear Regression Model based on Prinicples of Econometrics Jakub Mućk Department of Quantitative Economics Jakub Mućk Applied Econometrics (QEM) Meeting #2 The Simple

More information