Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator

Size: px
Start display at page:

Download "Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator"

Transcription

1 Slide Set 13 Liear Model with Edogeous Regressors ad the GMM estimator Pietro Coretto Ecoometrics Master i Ecoomics ad Fiace (MEF) Uiversità degli Studi di Napoli Federico II Versio: Friday 1 st March, 2019 (h17:43) P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 1 / 25 Summary Liear model with edogeous regressors The GMM estimator ad its special case (IV estimator) Asymptotic distributio of the GMM estimator Cosistet estimatio of error variace Cosistet estimatio of stadard errors Efficiecy ad the 2-steps efficiet GMM P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 2 / 25

2 Itro ad otatios Here we preset a framework where the model is specified i terms of momet coditios (MC). IV are aturally embodied i this framework. Now we cosider the case where we observe three types of variables y i : the usual respose/outcome/idepedet variable z i : a set of L regressors for y i. The primary goal is to kow their partial effect o the expected y (that is E[y i z i ]/ z i ). However, some of these variables may be edogeous. x i : a set of K exogeous variables. x i ad z i ca share some of the variables. x i usually cotais IVs ad exogeous regressors Remark: the iferece is based o asymptotic approximatios, hece max {K, L} < P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 3 / 25 Assumptios (C1): liearity Let z i R L be a vector of regressors, δ R L a vector of L coefficiets, {ε i } is a uobservable error term, ad y i =δ 1 + δ 2 z i2 + δ 3 z i3 +,..., +δ L z il + ε i for i = 1, 2,..., =z iδ + ε i z i1 = 1 for all i = 1, 2,..., if we have a costat term i the model. (C2): ergodic statioarity Let x i be a K-dimesioal vector of istrumets, let w i be the uique ad ocostat term of (y i, z i, x i). Assume {w i } is joitly statioary ad ergodic. P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 4 / 25

3 Deote: g i = g(w i, δ) = x i ε i = x i (y i z iδ) (C3): momet equatios, orthogoality coditios All the K elemets of x i are predetermied i the sese that E[x ik ε i ] = 0 for all i, k; or E[g i ] = 0. Exogeous variables that are also regressors are called predetermied regressors. The rest of regressors are called edogeous regressors (C3) gives a key tool: a system of K momet equatios with L ukows: (eq. 1) E[x i1 (y i z iδ)] = 0 (eq. 2) E[x i2 (y i z iδ)] = 0. (eq. K) E[x ik (y i z iδ)] = 0 Vector form: E[g(w i, δ)] = E[x i (y i z iδ)] = 0 P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 5 / 25 (C4): rak coditio The (K L) matrix E[x i z i] = Σ xz has rak equal to L (full colum rak). Propositio (idetificatio) Uder (C2), (C3) ad (C4), the parameter vector δ is uiquely idetified. Proof. Based o (C3) we kow that there exists a δ such that That is E[x i (y i z iδ)] = E[x i y i ] E[x i z i]δ = 0 E[x i z i]δ = E[x i y i ] (13.1) Based o (C2) we kow that these momets exists ad do ot deped o i. From liear algebra we kow that the there exists a uique δ satisfyig (13.1) if ad oly if rak(e[x i z i]) = L. Based o (C4) the ukow parameter δ is uiquely determied based o momets obtaied from the joit distributio of the observables. Therefore, δ is idetified. P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 6 / 25

4 We use the otatio σ xy = E[x i y i ], ad eq. (13.1) is rewritte as Σ xz δ = σ xy. Note that rak(σ xz ) mi {K, L}, therefore a ecessary (but ot sufficiet) coditio for idetifiability is that K L (order coditio). The Model is: uderidetified: order coditio is ot satisfied because K < L. It s hopeless to estimate δ because it s ot idetified. just idetified: rak coditio is satisfied with K = L. We ca estimate δ overidetified: rak coditio is satisfied with K > L We ca estimate δ P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 7 / 25 (C5): martigale depedece {g i } is a MDS. The (K K) matrix of cross momets, S = E[g i g i] is o sigular. (C5) allows to derive CLT-types approximatios. It also allows to ivert S A sufficiet coditio for (C5) is E[ε i ε i 1, ε i 2,..., ε 1, x i, x i 1,..., x 1 ] = 0 which implies that the error term is orthogoal to both past ad preset values of the exogeous variables x i. P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 8 / 25

5 Empirical aalog of (C3) Deote s xy = 1 x i y i ad S xz = 1 x i z i these are sample couterparts of σ xy, Σ xz Rewrite the sample versio of E[g i ] g (δ) = 1 g(w i ; δ) = 1 x i (y i z iδ) = 1 ( ) 1 x i y i x i z i δ =s xy S xz δ P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 9 / 25 Now the empirical versio of the system of momet equatios g (δ) = 1 x i (y i z iδ) = 0 is equivalet to (eq. 1) 1 x i1 (y i z iδ) = 0 (eq. 2) 1 x i2 (y i z iδ) = 0.. (eq. K) 1 x ik (y i z iδ) = 0 P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 10 / 25

6 K = L (model just idetified): IV estimator The typical case is that some of the regressors are edogeous, ad we have a IV for each edogeous variable. Recall: if a regressor is exogeous, it ca be istrumeted by itself. Suppose we have L regressors icludig the itercept. Assume that the last q regressors are edogeous z i = (1, z i,2, z i,3,..., z i,(l q), z i,(l q+1),..., z il ) A typical choice for the exogeous vector is: x i = (1, z i,2, z i,3,..., z i,(l q), x i1, x i2,..., x iq ) K = q istrumets + (L q) exogeous regressors, the x i R L. Data ca be arraged i two desig matrices P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 11 / 25 Z is the desig matrix with the L regressors where the first L K colums are samples of the exogeous regressors 1 z 1,2... z 1,L 1 z 2,2... z 2,L Z = z,2... z,l X is the desig matrix of the K = L exogeous variables. The last q colums are the sample values of the IVs 1 z 1,2 z 1,(L q) x 1,1 x 1,2 x 1,q 1 z 2,2 z X = 2,(L q) x 2,1 x 2,2 x 2,q z,2 z,(l q) x,1 x,2 x,q P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 12 / 25

7 If K = L by (C4) Σ xz is square ivertible. Moreover, by (C2) as S xz Σ xz, this meas that S xz is ivertible with probability 1 for sufficietly large. The for large we ca estimate δ by solvig the sample versio of the momet equatios. That is, we look for ˆδ IV such that g (ˆδ IV ) = 0: g (ˆδ IV ) = 0 = S xz ˆδIV = s xy = S 1 xz S xz ˆδIV = S 1 xz s xy ˆδ IV = S 1 xz s xy = ( 1 x i z i ) 1 ( 1 ˆδ IV is called istrumetal variable estimator ˆδ IV is a method of momets estimator (MM) If z i = x i for all i, the we go back to OLS ) x i y i P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 13 / 25 K > L (model overidetified): GMM Sice K > L we have more equatios tha ukow. There may be multiple solutios to g (δ) = 0. We may look for a δ that makes g (δ) as close as possible to 0. Maybe a δ that miimizes the orm g (δ) 2 However if the elemets of g ( ) are ot orthogoal, g ( ) 2 is ot really good at measurig the size of g (δ). I these cases we kow that we ca measure the size of g (δ) based o a quadratic form (of Mahalaobis type) with some appropriate symmetric ad PD matrix W : J(δ, W ) := g (δ) W g (δ) P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 14 / 25

8 GMM estimator Defiitio (GMM: geeralized method of momets estimator) Suppose W is a symmetric ad PD (K K) matrix. Assume that there exists Ŵ such that Ŵ p W. ˆδ(Ŵ ) is the GMM estimator if ˆδ(Ŵ ) := arg mi δ g (δ) Ŵ g (δ) Ŵ ca be costat or radom. The latter covers the case whe W is data-depedet Ŵ weights the equatios P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 15 / 25 g (δ) is the sample versio of the K momet equatios g (δ) = g (1) g (2) g (K) (δ) (δ). (δ) where g (k) (δ) = 1 x ik (y i z iδ) Here Ŵ = (ŵ k,j). It is easily see that Ŵ weights the samplig momet equatios ito the objective fuctio J(δ, Ŵ ) =g (δ) Ŵ g (δ) K K = ŵ k,j g (k) (δ) g (j) (δ) k=1 j=1 P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 16 / 25

9 I vector form J(δ, Ŵ ) = (s xy S xz δ) Ŵ (s xy S xz δ) J( ) is quadratic wrt ψ = (s xy S xz δ) ψ is liear i δ. FOC for the miimum of J(δ, Ŵ ) are S xzŵ s xy S xzŵ S xzδ = 0 Uder (C2) ad (C4), S xz is full colum rak with probability 1, for large Ŵ is PD with probability 1. Hece, S xzŵ S xz is osigular with probability 1. Pre-multiply both sides by (S xzŵ S xz) 1 ad we get the GMM estimator δ(ŵ ) = (S xzŵ S xz) 1 S xzŵ s xy P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 17 / 25 Properties of the GMM estimator For ay suitable choice of Ŵ Cosistecy: uder (C1)-(C4), δ(ŵ ) Normality: uder (C1)-(C5) where ( δ( Ŵ ) δ) p δ d N (0, AVar( δ(ŵ ))) AVar( δ(ŵ )) = (Σ xzw Σ xz ) 1 (Σ xzw SW Σ xz ) (Σ xzw Σ xz ) 1 Recall: Σ xz = E[x i z i], S = E[g i g i] = E[ε 2 i x i x i], Ŵ p W P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 18 / 25

10 By (C2) S xz = 1 x i z i as Σ xz Assume you ca fid ay estimator Ŝ p S, the by cotiuous mappig theorem we get a cosistet estimator for the asymptotic variace AVar( δ(ŵ )) = (S xzŵ S xz) 1 (S xzŵ ŜŴ S xz) (S xzŵ S xz) 1 How to fid Ŝ? Aswer: aalog priciple applied to the followig populatio momet E[g i g i] = E[ε 2 i x i x i] Problem: as usual oe caot observe ε i directly, we have to rely o its recostructio P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 19 / 25 Cosistet estimatio of the error variace Residual e i = y i z i δ(ŵ ) (C6): existece of momets of {z i } E[ z i z i ] < Uder (C1), (C2), ad (C6), it ca be show that 1 ei 2 p E[ε 2 i ] = Var[ε i ] The proof of this statemet is almost idetical to the oe for the case of large sample OLS. P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 20 / 25

11 (C7): existece of higher order momets E[(x ik z il ) 2 ] <, for all k = 1, 2,..., K ad l = 1, 2,..., L. (C7) implies that the joit distributio of regressors ad istrumets has a ice tail behavior Uder (C1), (C2), ad (C7): Ŝ = 1 ei 2 x i x i p S We ow have a cosistet estimator for the asymptotic variace. So we have everythig we eed to make iferece o δ Wait a miute... what about Ŵ? P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 21 / 25 Efficiet GMM Efficiecy of δ depeds crucially o Ŵ. We would like to have a Ŵ p W such that AVar( δ(ŵ )) is as small as possible i a large sample. Efficiecy lower boud: it ca be proved that (Σ xzs 1 Σ xz ) 1 AVar( δ(w )) for ay appropriate choices of Ŵ we achieve the lower boud with Ŵ = S 1 therefore the best choice is Ŵ = Ŝ 1 p S 1 P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 22 / 25

12 δ (Ŝ 1 ) is called efficiet GMM estimator. It s asymptotic variace is ow give by It is estimated by AVar ( δ( Ŝ 1 ) ) 1 ) = (Σ xzs 1 Σ xz AVar ( δ( Ŝ 1 ) ) 1 ) = (S xzŝ 1 S xz Ŝ depeds o e i which depeds o the estimated δ, which i turs depeds o Ŵ. P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 23 / 25 2-steps efficiet GMM Step 1: Step 2: take ay Ŵ 1 that coverges to a (K K) positive symmetric matrix. Ŵ 1 = I K, or Ŵ 1 = S 1 xx are commo choices. compute the cosistet δ(ŵ 1) compute e i = y i z i δ(ŵ 1), ad get the correspodig cosistet Ŝ1 set Ŵ 2 = Ŝ 1 1 compute the cosistet δ(ŵ 2) δ(ŵ 2) is the efficiet GMM. Hece, δ(ŵ 2) has the lower boud variace-covariace matrix. P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 24 / 25

13 Cocludig remarks Whe K = L the GMM is equivalet to the IV estimator. IV estimator is a particular case of the GMM estimator. Whe errors are homoscedastic, the efficiet GMM is equivalet to a classical estimator called 2SLS= 2-stages least square. 2SLS played a importat role i the begiig of IV literature. P. Coretto MEF Liear Model with Edogeous Regressors ad the GMM estimator 25 / 25

CLRM estimation Pietro Coretto Econometrics

CLRM estimation Pietro Coretto Econometrics Slide Set 4 CLRM estimatio Pietro Coretto pcoretto@uisa.it Ecoometrics Master i Ecoomics ad Fiace (MEF) Uiversità degli Studi di Napoli Federico II Versio: Thursday 24 th Jauary, 2019 (h08:41) P. Coretto

More information

Efficient GMM LECTURE 12 GMM II

Efficient GMM LECTURE 12 GMM II DECEMBER 1 010 LECTURE 1 II Efficiet The estimator depeds o the choice of the weight matrix A. The efficiet estimator is the oe that has the smallest asymptotic variace amog all estimators defied by differet

More information

Solution to Chapter 2 Analytical Exercises

Solution to Chapter 2 Analytical Exercises Nov. 25, 23, Revised Dec. 27, 23 Hayashi Ecoometrics Solutio to Chapter 2 Aalytical Exercises. For ay ε >, So, plim z =. O the other had, which meas that lim E(z =. 2. As show i the hit, Prob( z > ε =

More information

11 THE GMM ESTIMATION

11 THE GMM ESTIMATION Cotets THE GMM ESTIMATION 2. Cosistecy ad Asymptotic Normality..................... 3.2 Regularity Coditios ad Idetificatio..................... 4.3 The GMM Iterpretatio of the OLS Estimatio.................

More information

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator Ecoomics 24B Relatio to Method of Momets ad Maximum Likelihood OLSE as a Maximum Likelihood Estimator Uder Assumptio 5 we have speci ed the distributio of the error, so we ca estimate the model parameters

More information

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d Liear regressio Daiel Hsu (COMS 477) Maximum likelihood estimatio Oe of the simplest liear regressio models is the followig: (X, Y ),..., (X, Y ), (X, Y ) are iid radom pairs takig values i R d R, ad Y

More information

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors ECONOMETRIC THEORY MODULE XIII Lecture - 34 Asymptotic Theory ad Stochastic Regressors Dr. Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Asymptotic theory The asymptotic

More information

Asymptotic Results for the Linear Regression Model

Asymptotic Results for the Linear Regression Model Asymptotic Results for the Liear Regressio Model C. Fli November 29, 2000 1. Asymptotic Results uder Classical Assumptios The followig results apply to the liear regressio model y = Xβ + ε, where X is

More information

Statistical Inference Based on Extremum Estimators

Statistical Inference Based on Extremum Estimators T. Rotheberg Fall, 2007 Statistical Iferece Based o Extremum Estimators Itroductio Suppose 0, the true value of a p-dimesioal parameter, is kow to lie i some subset S R p : Ofte we choose to estimate 0

More information

POLS, GLS, FGLS, GMM. Outline of Linear Systems of Equations. Common Coefficients, Panel Data Model. Preliminaries

POLS, GLS, FGLS, GMM. Outline of Linear Systems of Equations. Common Coefficients, Panel Data Model. Preliminaries Outlie of Liear Systems of Equatios POLS, GLS, FGLS, GMM Commo Coefficiets, Pael Data Model Prelimiaries he liear pael data model is a static model because all explaatory variables are dated cotemporaeously

More information

Regression with an Evaporating Logarithmic Trend

Regression with an Evaporating Logarithmic Trend Regressio with a Evaporatig Logarithmic Tred Peter C. B. Phillips Cowles Foudatio, Yale Uiversity, Uiversity of Aucklad & Uiversity of York ad Yixiao Su Departmet of Ecoomics Yale Uiversity October 5,

More information

Single-Equation GMM: Estimation

Single-Equation GMM: Estimation Sigle-Equatio GMM: Estimatio Lecture for Ecoomics 241B Douglas G. Steigerwald UC Sata Barbara Jauary 2012 Iitial Questio Iitial Questio How valuable is ivestmet i college educatio? ecoomics - measure value

More information

Chapter 6 Principles of Data Reduction

Chapter 6 Principles of Data Reduction Chapter 6 for BST 695: Special Topics i Statistical Theory. Kui Zhag, 0 Chapter 6 Priciples of Data Reductio Sectio 6. Itroductio Goal: To summarize or reduce the data X, X,, X to get iformatio about a

More information

Matrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Pearson Education, Inc.

Matrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Pearson Education, Inc. 2 Matrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES 2012 Pearso Educatio, Ic. Theorem 8: Let A be a square matrix. The the followig statemets are equivalet. That is, for a give A, the statemets

More information

Solution to Chapter 3 Analytical Exercises

Solution to Chapter 3 Analytical Exercises December 7, 003 Solutio to Chapter 3 Aalytical Exercises Hayashi Ecoometrics. If A is symmetric a iempotet, the A = A a AA = A. So x Ax = x AAx = x A Ax = z z 0 where z Ax.. (a) By assumptio, {x i, ε i

More information

1 General linear Model Continued..

1 General linear Model Continued.. Geeral liear Model Cotiued.. We have We kow y = X + u X o radom u v N(0; I ) b = (X 0 X) X 0 y E( b ) = V ar( b ) = (X 0 X) We saw that b = (X 0 X) X 0 u so b is a liear fuctio of a ormally distributed

More information

Distribution of Random Samples & Limit theorems

Distribution of Random Samples & Limit theorems STAT/MATH 395 A - PROBABILITY II UW Witer Quarter 2017 Néhémy Lim Distributio of Radom Samples & Limit theorems 1 Distributio of i.i.d. Samples Motivatig example. Assume that the goal of a study is to

More information

Chapter 1 Simple Linear Regression (part 6: matrix version)

Chapter 1 Simple Linear Regression (part 6: matrix version) Chapter Simple Liear Regressio (part 6: matrix versio) Overview Simple liear regressio model: respose variable Y, a sigle idepedet variable X Y β 0 + β X + ε Multiple liear regressio model: respose Y,

More information

Matrix Algebra 2.2 THE INVERSE OF A MATRIX Pearson Education, Inc.

Matrix Algebra 2.2 THE INVERSE OF A MATRIX Pearson Education, Inc. 2 Matrix Algebra 2.2 THE INVERSE OF A MATRIX MATRIX OPERATIONS A matrix A is said to be ivertible if there is a matrix C such that CA = I ad AC = I where, the idetity matrix. I = I I this case, C is a

More information

Random Variables, Sampling and Estimation

Random Variables, Sampling and Estimation Chapter 1 Radom Variables, Samplig ad Estimatio 1.1 Itroductio This chapter will cover the most importat basic statistical theory you eed i order to uderstad the ecoometric material that will be comig

More information

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample. Statistical Iferece (Chapter 10) Statistical iferece = lear about a populatio based o the iformatio provided by a sample. Populatio: The set of all values of a radom variable X of iterest. Characterized

More information

1 Covariance Estimation

1 Covariance Estimation Eco 75 Lecture 5 Covariace Estimatio ad Optimal Weightig Matrices I this lecture, we cosider estimatio of the asymptotic covariace matrix B B of the extremum estimator b : Covariace Estimatio Lemma 4.

More information

Convergence of random variables. (telegram style notes) P.J.C. Spreij

Convergence of random variables. (telegram style notes) P.J.C. Spreij Covergece of radom variables (telegram style otes).j.c. Spreij this versio: September 6, 2005 Itroductio As we kow, radom variables are by defiitio measurable fuctios o some uderlyig measurable space

More information

An Introduction to Asymptotic Theory

An Introduction to Asymptotic Theory A Itroductio to Asymptotic Theory Pig Yu School of Ecoomics ad Fiace The Uiversity of Hog Kog Pig Yu (HKU) Asymptotic Theory 1 / 20 Five Weapos i Asymptotic Theory Five Weapos i Asymptotic Theory Pig Yu

More information

x iu i E(x u) 0. In order to obtain a consistent estimator of β, we find the instrumental variable z which satisfies E(z u) = 0. z iu i E(z u) = 0.

x iu i E(x u) 0. In order to obtain a consistent estimator of β, we find the instrumental variable z which satisfies E(z u) = 0. z iu i E(z u) = 0. 27 However, β MM is icosistet whe E(x u) 0, i.e., β MM = (X X) X y = β + (X X) X u = β + ( X X ) ( X u ) \ β. Note as follows: X u = x iu i E(x u) 0. I order to obtai a cosistet estimator of β, we fid

More information

LECTURE 8: ASYMPTOTICS I

LECTURE 8: ASYMPTOTICS I LECTURE 8: ASYMPTOTICS I We are iterested i the properties of estimators as. Cosider a sequece of radom variables {, X 1}. N. M. Kiefer, Corell Uiversity, Ecoomics 60 1 Defiitio: (Weak covergece) A sequece

More information

Algebra of Least Squares

Algebra of Least Squares October 19, 2018 Algebra of Least Squares Geometry of Least Squares Recall that out data is like a table [Y X] where Y collects observatios o the depedet variable Y ad X collects observatios o the k-dimesioal

More information

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2. SAMPLE STATISTICS A radom sample x 1,x,,x from a distributio f(x) is a set of idepedetly ad idetically variables with x i f(x) for all i Their joit pdf is f(x 1,x,,x )=f(x 1 )f(x ) f(x )= f(x i ) The sample

More information

The Method of Least Squares. To understand least squares fitting of data.

The Method of Least Squares. To understand least squares fitting of data. The Method of Least Squares KEY WORDS Curve fittig, least square GOAL To uderstad least squares fittig of data To uderstad the least squares solutio of icosistet systems of liear equatios 1 Motivatio Curve

More information

1.010 Uncertainty in Engineering Fall 2008

1.010 Uncertainty in Engineering Fall 2008 MIT OpeCourseWare http://ocw.mit.edu.00 Ucertaity i Egieerig Fall 2008 For iformatio about citig these materials or our Terms of Use, visit: http://ocw.mit.edu.terms. .00 - Brief Notes # 9 Poit ad Iterval

More information

Properties and Hypothesis Testing

Properties and Hypothesis Testing Chapter 3 Properties ad Hypothesis Testig 3.1 Types of data The regressio techiques developed i previous chapters ca be applied to three differet kids of data. 1. Cross-sectioal data. 2. Time series data.

More information

Lecture 33: Bootstrap

Lecture 33: Bootstrap Lecture 33: ootstrap Motivatio To evaluate ad compare differet estimators, we eed cosistet estimators of variaces or asymptotic variaces of estimators. This is also importat for hypothesis testig ad cofidece

More information

[412] A TEST FOR HOMOGENEITY OF THE MARGINAL DISTRIBUTIONS IN A TWO-WAY CLASSIFICATION

[412] A TEST FOR HOMOGENEITY OF THE MARGINAL DISTRIBUTIONS IN A TWO-WAY CLASSIFICATION [412] A TEST FOR HOMOGENEITY OF THE MARGINAL DISTRIBUTIONS IN A TWO-WAY CLASSIFICATION BY ALAN STUART Divisio of Research Techiques, Lodo School of Ecoomics 1. INTRODUCTION There are several circumstaces

More information

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1 EECS564 Estimatio, Filterig, ad Detectio Hwk 2 Sols. Witer 25 4. Let Z be a sigle observatio havig desity fuctio where. p (z) = (2z + ), z (a) Assumig that is a oradom parameter, fid ad plot the maximum

More information

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise First Year Quatitative Comp Exam Sprig, 2012 Istructio: There are three parts. Aswer every questio i every part. Questio I-1 Part I - 203A A radom variable X is distributed with the margial desity: >

More information

Linear Regression Models, OLS, Assumptions and Properties

Linear Regression Models, OLS, Assumptions and Properties Chapter 2 Liear Regressio Models, OLS, Assumptios ad Properties 2.1 The Liear Regressio Model The liear regressio model is the sigle most useful tool i the ecoometricia s kit. The multiple regressio model

More information

Machine Learning Brett Bernstein

Machine Learning Brett Bernstein Machie Learig Brett Berstei Week Lecture: Cocept Check Exercises Starred problems are optioal. Statistical Learig Theory. Suppose A = Y = R ad X is some other set. Furthermore, assume P X Y is a discrete

More information

MATHEMATICAL SCIENCES PAPER-II

MATHEMATICAL SCIENCES PAPER-II MATHEMATICAL SCIENCES PAPER-II. Let {x } ad {y } be two sequeces of real umbers. Prove or disprove each of the statemets :. If {x y } coverges, ad if {y } is coverget, the {x } is coverget.. {x + y } coverges

More information

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4 MATH 30: Probability ad Statistics 9. Estimatio ad Testig of Parameters Estimatio ad Testig of Parameters We have bee dealig situatios i which we have full kowledge of the distributio of a radom variable.

More information

Large Sample Theory. Convergence. Central Limit Theorems Asymptotic Distribution Delta Method. Convergence in Probability Convergence in Distribution

Large Sample Theory. Convergence. Central Limit Theorems Asymptotic Distribution Delta Method. Convergence in Probability Convergence in Distribution Large Sample Theory Covergece Covergece i Probability Covergece i Distributio Cetral Limit Theorems Asymptotic Distributio Delta Method Covergece i Probability A sequece of radom scalars {z } = (z 1,z,

More information

CHAPTER 5. Theory and Solution Using Matrix Techniques

CHAPTER 5. Theory and Solution Using Matrix Techniques A SERIES OF CLASS NOTES FOR 2005-2006 TO INTRODUCE LINEAR AND NONLINEAR PROBLEMS TO ENGINEERS, SCIENTISTS, AND APPLIED MATHEMATICIANS DE CLASS NOTES 3 A COLLECTION OF HANDOUTS ON SYSTEMS OF ORDINARY DIFFERENTIAL

More information

ECON 3150/4150, Spring term Lecture 3

ECON 3150/4150, Spring term Lecture 3 Itroductio Fidig the best fit by regressio Residuals ad R-sq Regressio ad causality Summary ad ext step ECON 3150/4150, Sprig term 2014. Lecture 3 Ragar Nymoe Uiversity of Oslo 21 Jauary 2014 1 / 30 Itroductio

More information

Matrix Representation of Data in Experiment

Matrix Representation of Data in Experiment Matrix Represetatio of Data i Experimet Cosider a very simple model for resposes y ij : y ij i ij, i 1,; j 1,,..., (ote that for simplicity we are assumig the two () groups are of equal sample size ) Y

More information

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss ECE 90 Lecture : Complexity Regularizatio ad the Squared Loss R. Nowak 5/7/009 I the previous lectures we made use of the Cheroff/Hoeffdig bouds for our aalysis of classifier errors. Hoeffdig s iequality

More information

Math 220B Final Exam Solutions March 18, 2002

Math 220B Final Exam Solutions March 18, 2002 Math 0B Fial Exam Solutios March 18, 00 1. (1 poits) (a) (6 poits) Fid the Gree s fuctio for the tilted half-plae {(x 1, x ) R : x 1 + x > 0}. For x (x 1, x ), y (y 1, y ), express your Gree s fuctio G(x,

More information

Lecture 19: Convergence

Lecture 19: Convergence Lecture 19: Covergece Asymptotic approach I statistical aalysis or iferece, a key to the success of fidig a good procedure is beig able to fid some momets ad/or distributios of various statistics. I may

More information

Probability 2 - Notes 10. Lemma. If X is a random variable and g(x) 0 for all x in the support of f X, then P(g(X) 1) E[g(X)].

Probability 2 - Notes 10. Lemma. If X is a random variable and g(x) 0 for all x in the support of f X, then P(g(X) 1) E[g(X)]. Probability 2 - Notes 0 Some Useful Iequalities. Lemma. If X is a radom variable ad g(x 0 for all x i the support of f X, the P(g(X E[g(X]. Proof. (cotiuous case P(g(X Corollaries x:g(x f X (xdx x:g(x

More information

Technical Proofs for Homogeneity Pursuit

Technical Proofs for Homogeneity Pursuit Techical Proofs for Homogeeity Pursuit bstract This is the supplemetal material for the article Homogeeity Pursuit, submitted for publicatio i Joural of the merica Statistical ssociatio. B Proofs B. Proof

More information

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

Discrete Mathematics for CS Spring 2008 David Wagner Note 22 CS 70 Discrete Mathematics for CS Sprig 2008 David Wager Note 22 I.I.D. Radom Variables Estimatig the bias of a coi Questio: We wat to estimate the proportio p of Democrats i the US populatio, by takig

More information

STAT331. Example of Martingale CLT with Cox s Model

STAT331. Example of Martingale CLT with Cox s Model STAT33 Example of Martigale CLT with Cox s Model I this uit we illustrate the Martigale Cetral Limit Theorem by applyig it to the partial likelihood score fuctio from Cox s model. For simplicity of presetatio

More information

Estimation for Complete Data

Estimation for Complete Data Estimatio for Complete Data complete data: there is o loss of iformatio durig study. complete idividual complete data= grouped data A complete idividual data is the oe i which the complete iformatio of

More information

Exponential Families and Bayesian Inference

Exponential Families and Bayesian Inference Computer Visio Expoetial Families ad Bayesia Iferece Lecture Expoetial Families A expoetial family of distributios is a d-parameter family f(x; havig the followig form: f(x; = h(xe g(t T (x B(, (. where

More information

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence Chapter 3 Strog covergece As poited out i the Chapter 2, there are multiple ways to defie the otio of covergece of a sequece of radom variables. That chapter defied covergece i probability, covergece i

More information

Lecture 8: Convergence of transformations and law of large numbers

Lecture 8: Convergence of transformations and law of large numbers Lecture 8: Covergece of trasformatios ad law of large umbers Trasformatio ad covergece Trasformatio is a importat tool i statistics. If X coverges to X i some sese, we ofte eed to check whether g(x ) coverges

More information

Inverse Matrix. A meaning that matrix B is an inverse of matrix A.

Inverse Matrix. A meaning that matrix B is an inverse of matrix A. Iverse Matrix Two square matrices A ad B of dimesios are called iverses to oe aother if the followig holds, AB BA I (11) The otio is dual but we ofte write 1 B A meaig that matrix B is a iverse of matrix

More information

Economics 326 Methods of Empirical Research in Economics. Lecture 18: The asymptotic variance of OLS and heteroskedasticity

Economics 326 Methods of Empirical Research in Economics. Lecture 18: The asymptotic variance of OLS and heteroskedasticity Ecoomics 326 Methods of Empirical Research i Ecoomics Lecture 8: The asymptotic variace of OLS ad heteroskedasticity Hiro Kasahara Uiversity of British Columbia December 24, 204 Asymptotic ormality I I

More information

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett Lecture Note 8 Poit Estimators ad Poit Estimatio Methods MIT 14.30 Sprig 2006 Herma Beett Give a parameter with ukow value, the goal of poit estimatio is to use a sample to compute a umber that represets

More information

CEU Department of Economics Econometrics 1, Problem Set 1 - Solutions

CEU Department of Economics Econometrics 1, Problem Set 1 - Solutions CEU Departmet of Ecoomics Ecoometrics, Problem Set - Solutios Part A. Exogeeity - edogeeity The liear coditioal expectatio (CE) model has the followig form: We would like to estimate the effect of some

More information

Lecture 7: Properties of Random Samples

Lecture 7: Properties of Random Samples Lecture 7: Properties of Radom Samples 1 Cotiued From Last Class Theorem 1.1. Let X 1, X,...X be a radom sample from a populatio with mea µ ad variace σ

More information

1 Inferential Methods for Correlation and Regression Analysis

1 Inferential Methods for Correlation and Regression Analysis 1 Iferetial Methods for Correlatio ad Regressio Aalysis I the chapter o Correlatio ad Regressio Aalysis tools for describig bivariate cotiuous data were itroduced. The sample Pearso Correlatio Coefficiet

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

6. Kalman filter implementation for linear algebraic equations. Karhunen-Loeve decomposition

6. Kalman filter implementation for linear algebraic equations. Karhunen-Loeve decomposition 6. Kalma filter implemetatio for liear algebraic equatios. Karhue-Loeve decompositio 6.1. Solvable liear algebraic systems. Probabilistic iterpretatio. Let A be a quadratic matrix (ot obligatory osigular.

More information

Lecture 11 and 12: Basic estimation theory

Lecture 11 and 12: Basic estimation theory Lecture ad 2: Basic estimatio theory Sprig 202 - EE 94 Networked estimatio ad cotrol Prof. Kha March 2 202 I. MAXIMUM-LIKELIHOOD ESTIMATORS The maximum likelihood priciple is deceptively simple. Louis

More information

Review Questions, Chapters 8, 9. f(y) = 0, elsewhere. F (y) = f Y(1) = n ( e y/θ) n 1 1 θ e y/θ = n θ e yn

Review Questions, Chapters 8, 9. f(y) = 0, elsewhere. F (y) = f Y(1) = n ( e y/θ) n 1 1 θ e y/θ = n θ e yn Stat 366 Lab 2 Solutios (September 2, 2006) page TA: Yury Petracheko, CAB 484, yuryp@ualberta.ca, http://www.ualberta.ca/ yuryp/ Review Questios, Chapters 8, 9 8.5 Suppose that Y, Y 2,..., Y deote a radom

More information

Asymptotics. Hypothesis Testing UMP. Asymptotic Tests and p-values

Asymptotics. Hypothesis Testing UMP. Asymptotic Tests and p-values of the secod half Biostatistics 6 - Statistical Iferece Lecture 6 Fial Exam & Practice Problems for the Fial Hyu Mi Kag Apil 3rd, 3 Hyu Mi Kag Biostatistics 6 - Lecture 6 Apil 3rd, 3 / 3 Rao-Blackwell

More information

Lecture 20: Multivariate convergence and the Central Limit Theorem

Lecture 20: Multivariate convergence and the Central Limit Theorem Lecture 20: Multivariate covergece ad the Cetral Limit Theorem Covergece i distributio for radom vectors Let Z,Z 1,Z 2,... be radom vectors o R k. If the cdf of Z is cotiuous, the we ca defie covergece

More information

ARIMA Models. Dan Saunders. y t = φy t 1 + ɛ t

ARIMA Models. Dan Saunders. y t = φy t 1 + ɛ t ARIMA Models Da Sauders I will discuss models with a depedet variable y t, a potetially edogeous error term ɛ t, ad a exogeous error term η t, each with a subscript t deotig time. With just these three

More information

Simple Regression Model

Simple Regression Model Simple Regressio Model 1. The Model y i 0 1 x i u i where y i depedet variable x i idepedet variable u i disturbace/error term i 1,..., Eg: y wage (measured i 1976 dollars per hr) x educatio (measured

More information

2.2. Central limit theorem.

2.2. Central limit theorem. 36.. Cetral limit theorem. The most ideal case of the CLT is that the radom variables are iid with fiite variace. Although it is a special case of the more geeral Lideberg-Feller CLT, it is most stadard

More information

In this section we derive some finite-sample properties of the OLS estimator. b is an estimator of β. It is a function of the random sample data.

In this section we derive some finite-sample properties of the OLS estimator. b is an estimator of β. It is a function of the random sample data. 17 3. OLS Part III I this sectio we derive some fiite-sample properties of the OLS estimator. 3.1 The Samplig Distributio of the OLS Estimator y = Xβ + ε ; ε ~ N[0, σ 2 I ] b = (X X) 1 X y = f(y) ε is

More information

Estimation of the Mean and the ACVF

Estimation of the Mean and the ACVF Chapter 5 Estimatio of the Mea ad the ACVF A statioary process {X t } is characterized by its mea ad its autocovariace fuctio γ ), ad so by the autocorrelatio fuctio ρ ) I this chapter we preset the estimators

More information

Optimally Sparse SVMs

Optimally Sparse SVMs A. Proof of Lemma 3. We here prove a lower boud o the umber of support vectors to achieve geeralizatio bouds of the form which we cosider. Importatly, this result holds ot oly for liear classifiers, but

More information

Lecture 23: Minimal sufficiency

Lecture 23: Minimal sufficiency Lecture 23: Miimal sufficiecy Maximal reductio without loss of iformatio There are may sufficiet statistics for a give problem. I fact, X (the whole data set) is sufficiet. If T is a sufficiet statistic

More information

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara Poit Estimator Eco 325 Notes o Poit Estimator ad Cofidece Iterval 1 By Hiro Kasahara Parameter, Estimator, ad Estimate The ormal probability desity fuctio is fully characterized by two costats: populatio

More information

, then cv V. Differential Equations Elements of Lineaer Algebra Name: Consider the differential equation. and y2 cos( kx)

, then cv V. Differential Equations Elements of Lineaer Algebra Name: Consider the differential equation. and y2 cos( kx) Cosider the differetial equatio y '' k y 0 has particular solutios y1 si( kx) ad y cos( kx) I geeral, ay liear combiatio of y1 ad y, cy 1 1 cy where c1, c is also a solutio to the equatio above The reaso

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 9 Multicolliearity Dr Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Multicolliearity diagostics A importat questio that

More information

4. Partial Sums and the Central Limit Theorem

4. Partial Sums and the Central Limit Theorem 1 of 10 7/16/2009 6:05 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 4. Partial Sums ad the Cetral Limit Theorem The cetral limit theorem ad the law of large umbers are the two fudametal theorems

More information

Introduction to Optimization Techniques. How to Solve Equations

Introduction to Optimization Techniques. How to Solve Equations Itroductio to Optimizatio Techiques How to Solve Equatios Iterative Methods of Optimizatio Iterative methods of optimizatio Solutio of the oliear equatios resultig form a optimizatio problem is usually

More information

Definition 4.2. (a) A sequence {x n } in a Banach space X is a basis for X if. unique scalars a n (x) such that x = n. a n (x) x n. (4.

Definition 4.2. (a) A sequence {x n } in a Banach space X is a basis for X if. unique scalars a n (x) such that x = n. a n (x) x n. (4. 4. BASES I BAACH SPACES 39 4. BASES I BAACH SPACES Sice a Baach space X is a vector space, it must possess a Hamel, or vector space, basis, i.e., a subset {x γ } γ Γ whose fiite liear spa is all of X ad

More information

Applications in Linear Algebra and Uses of Technology

Applications in Linear Algebra and Uses of Technology 1 TI-89: Let A 1 4 5 6 7 8 10 Applicatios i Liear Algebra ad Uses of Techology,adB 4 1 1 4 type i: [1,,;4,5,6;7,8,10] press: STO type i: A type i: [4,-1;-1,4] press: STO (1) Row Echelo Form: MATH/matrix

More information

MA Advanced Econometrics: Properties of Least Squares Estimators

MA Advanced Econometrics: Properties of Least Squares Estimators MA Advaced Ecoometrics: Properties of Least Squares Estimators Karl Whela School of Ecoomics, UCD February 5, 20 Karl Whela UCD Least Squares Estimators February 5, 20 / 5 Part I Least Squares: Some Fiite-Sample

More information

Math Solutions to homework 6

Math Solutions to homework 6 Math 175 - Solutios to homework 6 Cédric De Groote November 16, 2017 Problem 1 (8.11 i the book): Let K be a compact Hermitia operator o a Hilbert space H ad let the kerel of K be {0}. Show that there

More information

ECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization

ECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization ECE 90 Lecture 4: Maximum Likelihood Estimatio ad Complexity Regularizatio R Nowak 5/7/009 Review : Maximum Likelihood Estimatio We have iid observatios draw from a ukow distributio Y i iid p θ, i,, where

More information

Mathematical Statistics - MS

Mathematical Statistics - MS Paper Specific Istructios. The examiatio is of hours duratio. There are a total of 60 questios carryig 00 marks. The etire paper is divided ito three sectios, A, B ad C. All sectios are compulsory. Questios

More information

Lecture 24: Variable selection in linear models

Lecture 24: Variable selection in linear models Lecture 24: Variable selectio i liear models Cosider liear model X = Z β + ε, β R p ad Varε = σ 2 I. Like the LSE, the ridge regressio estimator does ot give 0 estimate to a compoet of β eve if that compoet

More information

LECTURE 8: ORTHOGONALITY (CHAPTER 5 IN THE BOOK)

LECTURE 8: ORTHOGONALITY (CHAPTER 5 IN THE BOOK) LECTURE 8: ORTHOGONALITY (CHAPTER 5 IN THE BOOK) Everythig marked by is ot required by the course syllabus I this lecture, all vector spaces is over the real umber R. All vectors i R is viewed as a colum

More information

Apply change-of-basis formula to rewrite x as a linear combination of eigenvectors v j.

Apply change-of-basis formula to rewrite x as a linear combination of eigenvectors v j. Eigevalue-Eigevector Istructor: Nam Su Wag eigemcd Ay vector i real Euclidea space of dimesio ca be uiquely epressed as a liear combiatio of liearly idepedet vectors (ie, basis) g j, j,,, α g α g α g α

More information

5 Birkhoff s Ergodic Theorem

5 Birkhoff s Ergodic Theorem 5 Birkhoff s Ergodic Theorem Amog the most useful of the various geeralizatios of KolmogorovâĂŹs strog law of large umbers are the ergodic theorems of Birkhoff ad Kigma, which exted the validity of the

More information

First, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So,

First, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So, 0 2. OLS Part II The OLS residuals are orthogoal to the regressors. If the model icludes a itercept, the orthogoality of the residuals ad regressors gives rise to three results, which have limited practical

More information

Unbiased Estimation. February 7-12, 2008

Unbiased Estimation. February 7-12, 2008 Ubiased Estimatio February 7-2, 2008 We begi with a sample X = (X,..., X ) of radom variables chose accordig to oe of a family of probabilities P θ where θ is elemet from the parameter space Θ. For radom

More information

32 estimating the cumulative distribution function

32 estimating the cumulative distribution function 32 estimatig the cumulative distributio fuctio 4.6 types of cofidece itervals/bads Let F be a class of distributio fuctios F ad let θ be some quatity of iterest, such as the mea of F or the whole fuctio

More information

Chapter 4. Fourier Series

Chapter 4. Fourier Series Chapter 4. Fourier Series At this poit we are ready to ow cosider the caoical equatios. Cosider, for eample the heat equatio u t = u, < (4.) subject to u(, ) = si, u(, t) = u(, t) =. (4.) Here,

More information

Statistical and Mathematical Methods DS-GA 1002 December 8, Sample Final Problems Solutions

Statistical and Mathematical Methods DS-GA 1002 December 8, Sample Final Problems Solutions Statistical ad Mathematical Methods DS-GA 00 December 8, 05. Short questios Sample Fial Problems Solutios a. Ax b has a solutio if b is i the rage of A. The dimesio of the rage of A is because A has liearly-idepedet

More information

Geometry of LS. LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT

Geometry of LS. LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT OCTOBER 7, 2016 LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT Geometry of LS We ca thik of y ad the colums of X as members of the -dimesioal Euclidea space R Oe ca

More information

62. Power series Definition 16. (Power series) Given a sequence {c n }, the series. c n x n = c 0 + c 1 x + c 2 x 2 + c 3 x 3 +

62. Power series Definition 16. (Power series) Given a sequence {c n }, the series. c n x n = c 0 + c 1 x + c 2 x 2 + c 3 x 3 + 62. Power series Defiitio 16. (Power series) Give a sequece {c }, the series c x = c 0 + c 1 x + c 2 x 2 + c 3 x 3 + is called a power series i the variable x. The umbers c are called the coefficiets of

More information

Statistics 511 Additional Materials

Statistics 511 Additional Materials Cofidece Itervals o mu Statistics 511 Additioal Materials This topic officially moves us from probability to statistics. We begi to discuss makig ifereces about the populatio. Oe way to differetiate probability

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5 CS434a/54a: Patter Recogitio Prof. Olga Veksler Lecture 5 Today Itroductio to parameter estimatio Two methods for parameter estimatio Maimum Likelihood Estimatio Bayesia Estimatio Itroducto Bayesia Decisio

More information

10-701/ Machine Learning Mid-term Exam Solution

10-701/ Machine Learning Mid-term Exam Solution 0-70/5-78 Machie Learig Mid-term Exam Solutio Your Name: Your Adrew ID: True or False (Give oe setece explaatio) (20%). (F) For a cotiuous radom variable x ad its probability distributio fuctio p(x), it

More information

Lecture 3. Properties of Summary Statistics: Sampling Distribution

Lecture 3. Properties of Summary Statistics: Sampling Distribution Lecture 3 Properties of Summary Statistics: Samplig Distributio Mai Theme How ca we use math to justify that our umerical summaries from the sample are good summaries of the populatio? Lecture Summary

More information

The standard deviation of the mean

The standard deviation of the mean Physics 6C Fall 20 The stadard deviatio of the mea These otes provide some clarificatio o the distictio betwee the stadard deviatio ad the stadard deviatio of the mea.. The sample mea ad variace Cosider

More information