Regression and Statistical Inference
|
|
- Alyson Cook
- 5 years ago
- Views:
Transcription
1 Regression and Statistical Inference Walid Mnif Department of Applied Mathematics The University of Western Ontario, London, Canada 1
2 Elements of Probability 2
3 Elements of Probability CDF&PDF Recall for a given random variable (r.v.) X, the cumulative distribution function (CDF) is dened: F (x) = P {X x}, x R N (1) The probability density function (PDF): Discrete r.v. p(x) = P {X = x} (2) Continuous r.v. F (a) = a f(x)dx (3) W. Mnif 3
4 Elements of Probability CDF&PDF Example: Normal distribution with mean µ and variance σ 2 : N (µ, σ) f(x) = 1 σ 2π exp( (x µ)2 /(2σ 2 )) (4) Figure 1: PDF of the standard normal r.v. (µ = 0, σ = 1) W. Mnif 4
5 Elements of Probability CDF&PDF Example 2: Chi-square distribution with k degrees of freedom The distribution of a sum of the squares of k independent standard normal random variables Pdf: f(x; k) = 1 2 k/2 Γ(k/2) xk/2 1 e x/2, if x > 0 and 0 elsewhere, (5) where Γ is the Gamma function dened as: Γ(x) = 0 t x 1 e t dt. (6) Exercise: Find out the characteristics of the Gamma function. W. Mnif 5
6 Example 3: t-distribution Dened as the distribution of Elements of Probability CDF&PDF Z V/υ, (7) where Z N (0, 1), V χ 2 υ, and Z and V are independent. Pdf f(x) = Γ(υ+1 2 ) πυγ( υ 2 )(1 + x2 υ ) (υ+1)/2 (8) W. Mnif 6
7 k th moment about the origin Elements of Probability Moments m k = E[X k ] (9) k th central moment µ k = E[(X E[X]) k ] = E[(X m 1 ) k ] (10) Example: X N (µ, σ 2 ) m 1 = µ, µ 1 = 0, µ 2 = σ 2 W. Mnif 7
8 Elements of Probability Moments Theorem. If E[ X k ] < for some positive integer k, then E[ X j ] <, j = 1, 2,..., k 1 In other words: the existence of the k th moment implies the existence of all moments of lower order W. Mnif 8
9 Elements of Probability Moments Generating Functions Denition. Let X be a r.v. with pdf f. The moment generating function (MGF) of X, denoted by M X (t), is M X (t) = E[e tx ] provided E[e tx ] for all values t in an interval ( ɛ, ɛ), ɛ > 0. Explicitly, Continuous case: M X (t) = etx f(x)dx Discrete case: M X (t) = x etx f(x) W. Mnif 9
10 Elements of Probability Moments Generating Functions Theorem. If X has MGF M X (t), then E[X n ] = M (n) X (0) = dn dt nm X(t) t=0 (11) Example: Suppose X N (µ, σ 2 ), then M X (t) = e tµ+t2 σ 2 /2 (12) Proof. Exercise. W. Mnif 10
11 Elements of Probability Moments Generating Functions Theorem. Let X and Y be r.v.s that have the same MGF. Then, X and Y have the same distribution. Theorem. Assume X is a r.v. with MGF M X (t). Then, the MGF of Y = ax + b is for any value t such that M X (at) exists. M Y (t) = e bt M X (at) (13) W. Mnif 11
12 Elements of Probability Markov's Inequality Proposition. If X is a nonnegative r.v., then ɛ > 0 P {X ɛ} E[X] ɛ (14) Proof. Exercise! (Hint: Dene the r.v. Y s.t. Y = ɛ, if X ɛ and 0, elsewhere.) Corollary. (Chebyshev's Inequality) if X is a r.v. with mean µ and variance σ 2, then k > 0, P { X µ kσ} 1 k2. (15) Proof. Exercise! (Hint: Use Markov's Inequality ) W. Mnif 12
13 Elements of Probability Weak Law of Large Numbers Theorem. Let X 1,X 2,... be a sequence of independent and identically distributed r.v. with E[X i ] = µ and V ar(x i ) = σ 2 <. Then, ɛ > 0, P ( X X n n µ > ɛ) 0 as n (16) Proof. Exercise! (Hint: Use Chebyshev's Inequality) Remark: The Weak Law of Large Numbers holds for V ar(x i ) = W. Mnif 13
14 Elements of Probability Strong Law of Large Numbers Theorem 1. Let X 1,X 2,..., be a sequence of independent and identically distributed r.v. with E[X i ] = µ and V ar(x i ) = σ 2 <. Then, ɛ > 0, P ( lim n X X n n µ < ɛ) = 1 (17) W. Mnif 14
15 Elements of Probability Strong Law of Large Numbers Figure 2: Testing Law Large number for X i N (4, 2)) W. Mnif 15
16 Elements of Probability Central Limit Theorem Theorem. Let {X i } be an i.i.d vector in R d, with E[X i ] = µ and V ar[x i ] = Σ <, then 1 n (X i µ) n i=1 D n N (0, Σ) (18) Comment: The CLT is applicable to any arbitrary probability distribution. Proof. See Kallenberg(1997). W. Mnif 16
17 Elements of Probability Continuous Mapping Theorem Theorem. Continuous mapping 1: If X n P a and a function f continuous at a, then g(x n ) P n g(a) n Theorem. Continuous mapping 2: If X n D X and f(.) is a continuous function, then g(x n ) D n g(x) n W. Mnif 17
18 Elements of Probability Slutsky's Theorem Theorem. Slutsky's Theorem Suppose that X n D Y, where Y is a constant. Then, n X and Y n P n 1. X n + Y n D n X + Y 2. X n Y n D n XY 3. Y 1 D n X n n Y 1 X, provided Y is invertible. W. Mnif 18
19 Linear regression W. Mnif 19
20 Linear regression Least Squares tting Construct a straight line with equation y = b 0 + b 1 x, "best ts" the data (x 1, y 1 ),...,(x n, y n ) Denote the i th tted (or predicted) value by ŷ i = b 0 + b 1 x i, i = 1,..., n and the i th residual by e i = y i ŷ i = y i (b 0 + b 1 x i ) W. Mnif 20
21 Linear regression Least Squares tting Dene SSE = 1 i n e 2 i SSE is a measure of good tness to the data. When SSE = 0, all points y i lie on the line b 0 + b 1 x i. The method of least squares: ( β 0, β 1 ) = arg min (b 0,b 1 ) R 2 SSE (19) W. Mnif 21
22 Linear regression Least Squares tting Proposition. Dene n i=1 x = x n i i=1, ȳ = y i, S xx = n n n S yy = (y i ȳ) 2, S xy = i=1 i=1 n (x i x) 2, i=1 n (x i x)(y i ȳ) The line that best ts the data using the least squares criterion is: y = β 0 + β 1 x, W. Mnif 22
23 where β 1 = S xy S xx, and β 0 = ȳ β 1 x Proof. Exercise Hint: Dene r = S xy Sxx S yy as the sample covariance coecient and write the SSE as a sum of three squares. Proof that SSE = S yy (1 r 2 ) + (b 1 Sxx r S yy ) 2 + n(ȳ b 0 b 1 x) 2 Remarks: Under the least squares criterion, SSE = S yy (1 r 2 ) SSE is a measure of unexplained variability (error could not be explained by the model) W. Mnif 23
24 r 2 is called the coecient of determination, and is denoted by R 2 β 1 = r S yy Sxx The sample correlation and the estimated slope have the same sign Exercice Show that n i=1 (y i ȳ) 2 = n i=1 (ŷ i ȳ) 2 + n i=1 (y i ŷ) 2 and interpret the meaning of each term. W. Mnif 24
25 Linear regression Least Squares tting Dene: Total sum of squares SST = n i=1 (y i ȳ) 2 measure the total variability of the original observations. Regression Sum of squares SSR = n i=1 (ŷ i ȳ) 2 measure the total variability explained by the regression model. From previous slide: SST = SSR + SSE 1 = SSR SST + SSE SST SSR SST SSE SST : portion of the total variability that is explained by the model : the portion of unexplained variability. W. Mnif 25
26 Linear regression Least Squares tting Recall SSE = SST (1 R 2 ) R 2 = SSR SST, so if R 2 1 Most of the variability is explained if R 2 0 Inecient regression model Some properties of the residuals n i=1 e i = 0, n i=1 e i x i = 0, 1 n n i=1 ŷ i = ȳ. Exercise: Check the previous properties. W. Mnif 26
27 Linear regression Linear regression model Linear relationship between the response and explanatory variables + an error term The model: y = β 0 + β 1 x + ɛ where ɛ, called the error term or disturbance, includes all the other factors. Let denote by y i the response to the x i, i = 1,..., n. Assume that {(y i, x i )} n i=1 is i.i.d from a distribution y i = β 0 + β 1 x i + ɛ i where ɛ i are i.i.d and ɛ 1 x i N (0, σ 2 ). β 0 (resp. β 1 ) is called the intercept (resp. slope). W. Mnif 27
28 Linear regression Linear regression model Afterwards, we suppose that E(y 2 i ) <, E(x2 i ) < and E(x2 i ) 0 Theorem. follows: Under linear model assumptions, the coecients are given as β 1 = cov(x i, y i ) V ar(x i ) β 0 = E(y i ) β 1 E(x i ) Proof. By the LIE, E(x i ɛ i ) = 0 E(x i (y i β 0 β 1 x i )) = 0 On the other hand, E(y i ) = β 0 + β 1 E(x i ). After solving the two equations with two unknowns β 0 and β 1, the result is shown. W. Mnif 28
29 Linear regression Linear regression model Lemma. Assume that {(y i, x i )} n i=1 then is i.i.d. sample from population, ˆβ 1 = S xy S xx ˆβ 0 = y ˆβ 1 x are consistent estimators respectively to β 1 and β 0. Furthermore, the conditional distribution of ˆβ1 (resp. ˆβ0 ) is normal with mean β 1 (resp. β 0 ) and variance σ2 S xx (resp. σ 2 ( 1 n + x2 S xx ) ) W. Mnif 29
30 Linear regression Linear regression model Proof. Since {(y i, x i )} n i=1 is i.i.d, then {(x iy i )} n i=1 and {(x2 i )}n i=1 are also i.i.d. From the WLLN, S xy n P n cov(x i, y i ). The same for S xx P n V ar(x i). Since x 1 n x is continuous for any neighborhood of x 0 0, we apply the Continuous Mapping Theorem to get n S xx P n 1 V ar(x i ). Therefore, S xy S xx P n cov(x i,y i ) V ar(x i ) = β 1. As y and x are consistent respectively to E(y i ) and E(x i ), then we obtain the consistency for ˆβ 0 to β 0. W. Mnif 30
31 It can be shown that ˆβ 1 = ˆβ 0 = n i=1 n i=1 x i x y i S xx ( 1 n x(x i x S xx ))y i As y i x i N (β 0 + β 1 x i, σ 2 ) and the sum of a linear combination of i.i.d normally distributed random variables is normally distributed, we obtain the proof of the rest of the lemma. Remark: Since E( ˆβ 0 ) = β 0 and E( ˆβ 1 ) = β 1, so the estimators ˆβ 0 and ˆβ 1 are unbiased. An unknown variable should be estimated too. What is it????? W. Mnif 31
32 Linear regression Linear regression model The answer is σ 2. Recall that σ 2 is the variance of the dispersion the model. Small σ 2 (x i, y i ) lie close to the true regression line Large σ 2 the model is weak to explain the observed values (x i, y i ) Lemma. The statistic s 2 = n i=1 (Y i Ŷi) 2 n 2 (20) is an unbiased estimator of σ 2. Furthermore, (n 2)s 2 σ 2 χ 2 n 2, (21) W. Mnif 32
33 where χ 2 n 2 is the chi-square distribution with n 2 degrees of freedom. Moreover, s 2 is independent of both ˆβ 0 and ˆβ 1. Proof. The proof is basically straightforward if we show β 0 + β 1 x j, j, is normally distributed. β 0 + β 1 x j = n i=1 ( 1 n + (x j x) (x i x) S xx )y i Which is a linear combination of mutually independent, normally distributed random variables y i. Given E(y i x i ) = β 0 + β 1 x i and V (y i ) = σ 2, we get β 0 + β 1 x j is normally distributed with mean, variance and W. Mnif 33
34 estimated standard error resp.: E( β 0 + β 1 x j x) = β 0 + β 1 x j V ( β 0 + β 1 x j x) = σ 2 [ 1 n + (x j x) 2 ] S xx s( β 0 + β 1 1 x j x) = s n + (x j x) 2 S xx On the other hand, recall that the residuals satisfy: n i=1 n i=1 n e i = (Y i Ŷi) = 0 e i x i = i=1 n (Y i Ŷi)x i = 0 i=1 Therefore we can eliminate two summands from the sum of squares W. Mnif 34
35 n i=1 (y i ŷ i ) 2. Thus there are only n 2 independent y i ŷ i. We need to show that s 2 is unbiased estimator, which is equivalent to show that E( n i=1 (y i ŷ i ) 2 ) = (n 2)σ 2 We can easily proof that SSE = S yy β 2 1S xx. By taking expectations of both sides, we get E(SSE x) = E(S yy x) E( β 2 1 x)s xx or E( β 2 1 x) = σ2 S xx + β 2 1 ( Use V (X) = E(X2 ) (E(X)) 2 ) E(y 2 i ) = σ 2 + (β 0 + β 1 x i ) 2, and E(y 2 ) = σ2 n + (β 0 + β 1 x) 2 W. Mnif 35
36 So, E(S yy ) = (n 1)σ 2 + β 2 1S xx E(SSE) = (n 2)σ 2. Conclusion: s 2 is unbiased estimator of σ 2. W. Mnif 36
37 Linear regression Condence intervals for the Regression Coecients Denition. (Two sided condence intervals) An interval of the form [a, b], where a b,is said to be a 100(1 α)% condence interval for the parameter θ if P (a θ b) 1 α Exercise: Suppose that θ N (µ, σ 2 ). Find 95% condence interval for θ. W. Mnif 37
38 Linear regression Condence intervals for the Regression Coecients Denition. (One sided condence intervals) An interval of the form [a, ) is a 100(1 α)% lower one sided condence interval for the parameter θ if P (a θ) 1 α Similarly, an interval of the form (, b] is upper one sided condence interval for the parameter θ if P (θ b) 1 α Exercise: Suppose that θ N (µ, σ 2 ). Find 95% lower and upper one sided condence interval for θ. W. Mnif 38
39 Linear regression Condence intervals for the Regression Coecients For β 1 : Recall: β 1 β 1 σ/ S xx N (0, 1), (n 2)s 2 σ 2 χ 2 n 2 So, β 1 β 1 s/ S xx t n 2 s( β 1 ) = s S xx : the estimated standard error of the estimate. 100(1 α)% condence interval for β 1 β1 ± tn 2(α/2)s( β1) W. Mnif 39
40 Linear regression Condence intervals for the Regression Coecients For β 0 : The estimated standard error of β 0 is : s( β 0 ) = s 1 n + x2 S xx As we did for β 1, the sampling distribution of β 0 is: β 0 β 0 s( β 0 ) t n 2 100(1 α)% condence interval for β 0 β 0 ± t n 2 (α/2)s( β 0 ) W. Mnif 40
41 Linear regression Test of Hypotheses Suppose we want to test the hypothesis H 0 : β 1 = 0 against H 1 : β 1 0. Comparable to answer to the question: Does a variation of x have an eect on the response variable y? Under hypothesis H 0, β 1 = 0 β 1 s/ S xx t n 2 Then, a two sided level α test of H 0 is reject H 0 if β 1 s/ S xx t n 2 (α/2) W. Mnif 41
42 Linear regression CI for the mean response E(y i x i ) Derive point and interval estimates for E(y i x i ) we need to study the sampling distribution of the estimator β 0 + β 1 x i Theorem. The point estimator β 0 + β 1 x i is unbiased estimator of the mean response E(y i x i ). Furthermore, It is normally distributed with mean, variance, and estimated standard error given by: E( β 0 + β 1 x i ) = β 0 + β 1 x i, V ( β 0 + β 1 x i ) = σ 2 ( 1 n s( β 0 + β 1 1 x i ) = s n + (x x)2 S xx ), + (x x)2 S xx. W. Mnif 42
43 Proof. Exercise. Consequently, β 0 + β 1 x i (β 0 + β 1 x i ) s( β 0 + β 1 x i ) t n 2 100(1 α)% CI for the mean response: β 0 + β 1 x i ± t n 2 (α/2)s( β 0 + β 1 x i ) W. Mnif 43
44 Linear regression Prediction Interval Theorem. The estimated standard error s(y i ŷ i ) is given by: s(y i ŷ i ) = s n + (x x)2 S xx and y i ŷ i s(y i ŷ i ) t n 2 Proof. Use y i = β 0 + β 1 x i + ɛ where ɛ x i N (0, σ 2 ) W. Mnif 44
45 ŷ i = β 0 + β 1 x i ŷ i is independent of ɛ V (y i ŷ i ) = V (y i ) + V (ŷ i ) 100(1 α)% prediction interval β 0 + β 1 x i ± t n 2 (α/2)s(y i ŷ i ) W. Mnif 45
46 Linear regression Least squares linear predictor Dene y i R: explained variable or regressand; x i R d : explanatory variables or regressors The idea is to nd a "good" predictor of y i by a linear combination of x i, i = 1,..., n. "good"=pick up a linear combination of x i which minimizes the expected squared error. In other words, β = arg min b R d E[(y i X ib) 2 ] (22) W. Mnif 46
47 Linear regression Least squares linear predictor Theorem. Under the assumptions 1. E[y 2 i ] < 2. E[x i x i ] is a non-singular d d matrix 3. E[x i x i] < Then, β = (E[x i x i ]) 1 E(x i y i ) Proof. We have, b, E[(y i X i b)2 ] = E(y 2 i ) 2E(y ix i )b + b E(x i x i )b W. Mnif 47
48 First order condition, E[(y i X i b)2 ] b = 0 Recall if Σ is a d d matrix and b and X R d, then b X X = C X ΣX X = 2ΣX 2E(x i y i ) + 2E(x i x i)b = 0 Therefore, β = (E[x i x i ]) 1 E(x i y i ) Note that if we choose for example x i1 = 1, we work within an ane W. Mnif 48
49 framework. Moreover, if we choose d = 2, then we obtain the same results as in slide (28). W. Mnif 49
50 Linear regression Least squares linear predictor Theorem. Assume {(y i, x i )} n i=1 Under model assumptions, is i.i.d sample from a given population. β = [ n x i x i] 1 n x i y i (23) i=1 i=1 is a consistent estimator to β. Proof. We have from Cauchy Schwartz inequality E(x i y i ) E( x i y i ) (E(x i x i)) 1/2 E(y 2 i ) 1 2 < As {(y i, x i )} n i=1 is i.i.d sample, then {(y ix i )} n i=1 and {(x ix i )}n i=1 are W. Mnif 50
51 i.i.d. From WLLN, 1 n 1 n n i=1 n i=1 x i y i x i x i P n E(x iy i ) P n E(x ix i) Since g(a) = A 1 is continuous for any invertible A. continuous mapping theorem, By using the [ 1 n n i=1 x i x i] 1 P n [E(x ix i)] 1 W. Mnif 51
52 So, [ n i=1 x i x i] 1 n i=1 x i y i P n [E(x ix i)] 1 E(x i y i ) We can proof easily that the estimator dened before is the same as the OLS estimator. Recall the OLS estimator is dened as arg min b R d e e, where e = Y Xb and y 1 Y =. R n 1, X = y n x 1. x n R n d W. Mnif 52
53 Using the rst order condition, we get β = (X X) 1 X Y Now we can proof the OLS estimator is the same as (23). It is easy to proof it, just we use the following if A = (a 1,..., a n ) R m n and b 1 B =. R n p then, AB = n b i=1 a ib i n W. Mnif 53
54 Linear regression OLS Denition 2. Dene a regression model by y i = x i β + ɛ i such as E(ɛ i x i ) = 0. If V (y i x i ) is constant, then the model is homoskedastic. If it depends on x i, it is called heteroskedastic model. In fact, V (y i x i ) = V (x iβ + ɛ i x i ) = V (ɛ i x i ) = E(ɛ 2 i x i ) (E(E(ɛ i x i ))) 2 = E(ɛ 2 i x i ) = σ 2 (x i ) W. Mnif 54
55 Linear regression Finite-Sample Properties: Homoskedastic Case The model: {y i, x i }n i=1 i.i.d sample from y i = x iβ + ɛ i E(ɛ i x i ) = 0 V (ɛ i x i ) = E(ɛ 2 i x i ) = σ 2 Let study the OLS estimator proprieties when we applied to this model. We need to study its: bias variance consistency W. Mnif 55
56 asymptotic distribution W. Mnif 56
57 Linear regression OLS Bias Recall β = (X X) 1 X Y So the bias is: β β = (X X) 1 X (Xβ + ɛ) β = β + (X X) 1 X ɛ β = (X X) 1 X ɛ So, E( β β X) = (X X) 1 X E(ɛ X) = 0 W. Mnif 57
58 The OLS is conditionally unbiased. We can deduce that E( β β) = 0. Therefore, the OLS is unconditionally unbiased. W. Mnif 58
59 Linear regression OLS Variance V ( β X) = V ( β β X) = V ((X X) 1 X ɛ X) = (X X) 1 X V (ɛ)x(x X) 1 We used the following propriety: V (AX) = AV (X)A. On the other hand, E(ɛ 2 i X) = E(ɛ 2 i x i ) = σ 2 W. Mnif 59
60 For i j, E(ɛ i ɛ j X) = E(ɛ i E(ɛ j X, ɛ i ) X) = E(ɛ i E(ɛ j X) X) = E(ɛ i X)E(ɛ j X) = 0 So, E(ɛɛ X) = σ 2 I n and therefore V ( β X) = σ 2 (X X) 1. Recall that for ane model, we are estimating 2 coecients, and we found that the statistic s 2 = e e n 2 is unbiased consistent estimator for σ 2. We can, using the same methodology, show that s 2 = e e n d W. Mnif 60
61 is also unbiased consistent estimator for σ 2 for a dimension d model. W. Mnif 61
62 Linear regression Consistency of OLS Suppose that {(y i, x i )}n i=1 is i.i.d from y i = x i β + ɛ i and E(y 2 i ) <, E(x i x i) < E(x i x i ) is non singular E(ɛ i x i ) = 0, E(ɛ 2 i x i) = σ 2 E(ɛ 4 i ) <, E(x4 ij ) <, j = 1,.., d Then β = (X X) 1 X Y W. Mnif 62
63 is a consistent estimator to β. Proof. Then β = (X X) 1 X Y = [ 1 n x i x n i] 11 n = [ 1 n i=1 n x i x i] 11 n i=1 = β + [ 1 n n x i y i i=1 n x i (x iβ + ɛ i ) i=1 n x i x i] 11 n i=1 n x i ɛ i i=1 so β P β [ 1 n n n i=1 x ix i ] 1 1 n n i=1 x P iɛ i 1 n n 0 n i=1 x iɛ i P n 0 W. Mnif 63
64 [ 1 n We have E(x i x i ) <, by CMT, n i=1 x ix i ] 1 (E(x ix i )) 1 And, P n E(x ij ɛ i ) E( x ij ɛ i ) (E(x 2 ij)) 1 2 (E(ɛ 2 i )) 1 2 (E(x ix i )) 1 2 σ 2 < So from WLLN, 1 n Using the CMT β n i=1 x iɛ i P n β P n E(x iɛ i ) = 0, because E(ɛ i x i ) = 0. W. Mnif 64
65 Linear regression Asymptotic Propriety of OLS Under the previous assumption, we have n( β β) N (0, σ 2 [E(x i x i)] 1 ) Proof. We have {y i, x i } n i=1 is i.i.d, (x iɛ i ) n i=1 E(x i ɛ i ) = 0. Also, i.i.d. Furthermore, E(ɛ 2 i x ij x im ) E( ɛ 2 i x ij x im ) (E(ɛ 4 i )) 1 2 ((E(x 4 ij )) 2 1 (E(x 4 im )) ) 2 < then V (x i ɛ i ) < and is well dened. Furthermore, V (x i ɛ i ) = E(ɛ 2 i x i x i) = σ 2 E(x i x i) W. Mnif 65
66 By the CLT, 1 n (x i ɛ i E(x i ɛ))) = 1 n x i ɛ i n n i=1 D n i=1 N (0, σ2 E(x i x i)) By applying the Slutsky's theorem, we obtain the result. Lemma. Under model assumptions, a R d, na ( β β) s 2 a ( X X n ) 1 a N (0, 1). Proof. As s 2 = e e n d σ 2 [E(x i x i )] 1 P n σ2, Then from the CMT, s 2 ( X X n ) 1 P n W. Mnif 66
67 From CMT, na ( β β) And D n N (0, σ2 a [E(x i x i )] 1 a) 1 s 2 a ( X X n ) 1 a P n 1. σ 2 a [E(x i x i )] 1 a The result is proofed by using the Slutsky's theorem. W. Mnif 67
68 Linear regression Null Hypothesis test Suppose H 0 : a β = c 0 is true, then n(a β c0 ) s 2 a ( X X n ) 1 a N (0, 1). Test with signicance level α: We reject H 0 : a β = c 0 and accept H 1 : a β c 0 if n(a β c0 ) > q, s 2 a ( X X n ) 1 a where q is dened such P ( Z > q) = α, Z N (0, 1) W. Mnif 68
69 Condence Interval set We cannot reject H 0 : β = c 0,if a β q s2 a (X X) 1 a c 0 a β + q s2 a (X X) 1 a Exercise: Consider c 0 = 0 and a = (1, 0,..., 0). Interpret the result. W. Mnif 69
70 Linear regression Heteroscedastic Model The model: {y i, x i }n i=1 i.i.d sample from y i = x i β + ɛ i E(ɛ i x i ) = 0 V (ɛɛ x) = E(ɛ 2 i x i) = diag(σ 2 (x 1 ),..., σ 2 (x i ),..., σ 2 (x n )) = Ω E(y 2 i ) <, E(x i x i) < E(x i x i ) is non singular E(ɛ 4 i ) <, E(x4 ij ) <, j = 1,.., d W. Mnif 70
71 We can proof (Exercise) V ( β X) = (X X) 1 X ΩX(X X) 1 E( β X) = E( β) = β D n( β β) N (0, [E(x ix i)] 1 E(σ 2 (x i )x i x i)[e(x i x i)] 1 ) n Exercise: Derive an estimator to [E(x i x i )] 1 E(σ 2 (x i )x i x i )[E(x ix i )] 1 and a condence interval for a coecient β j. W. Mnif 71
72 Linear regression Generalized Least Squares estimators The model: {y i, x i }n i=1 i.i.d sample from y i = x i β + ɛ i E(ɛ i x i ) = 0 V (ɛɛ x) = E(ɛ 2 i x i) = Ω > 0 E(y 2 i ) <, E(x i x i) < E(x i x i ) is non singular E(ɛ 4 i ) <, E(x4 ij ) <, j = 1,.., d W. Mnif 72
73 The GLS dened by β = (X Ω 1 X) 1 X Ω 1 Y is the best linear unbiased and consistent estimator. Proof. Decompose Ω 1 = C C Dene variables Ỹ = CY, X = CX, and ɛ = Cɛ Use the following lemma: Lemma 3. (Gauss-Markov) The OLS estimator is the Best Linear Unbiased Estimator (BLUE) under the linear model with E(ɛ i x i ) = 0 and E(ɛ i x i ) = 0 W. Mnif 73
74 References [1] Kallenberg, O., "Foundations of Modern Probability", New York, Springer-Verlag, [2] Rosenkrantz, W. A.,"Probability and Statistics for Science, Engineering, and Finance", Chapman & Hall, [3] Wooldridge, J., "Econometric Analysis of Cross Section and Panel Data", MIT Press, W. Mnif 74
Simple Linear Regression: The Model
Simple Linear Regression: The Model task: quantifying the effect of change X in X on Y, with some constant β 1 : Y = β 1 X, linear relationship between X and Y, however, relationship subject to a random
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationReview of Econometrics
Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,
More informationLinear Models and Estimation by Least Squares
Linear Models and Estimation by Least Squares Jin-Lung Lin 1 Introduction Causal relation investigation lies in the heart of economics. Effect (Dependent variable) cause (Independent variable) Example:
More informationMFin Econometrics I Session 4: t-distribution, Simple Linear Regression, OLS assumptions and properties of OLS estimators
MFin Econometrics I Session 4: t-distribution, Simple Linear Regression, OLS assumptions and properties of OLS estimators Thilo Klein University of Cambridge Judge Business School Session 4: Linear regression,
More informationAdvanced Econometrics I
Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics
More informationIntermediate Econometrics
Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the
More informationReview of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley
Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate
More informationLecture 11. Multivariate Normal theory
10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More informationLinear models and their mathematical foundations: Simple linear regression
Linear models and their mathematical foundations: Simple linear regression Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/21 Introduction
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More informationECON The Simple Regression Model
ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In
More informationLECTURE 2 LINEAR REGRESSION MODEL AND OLS
SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another
More informationMultivariate Regression
Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the
More informationLeast Squares Estimation-Finite-Sample Properties
Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions
More informationLinear Regression and Its Applications
Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationMeasuring the fit of the model - SSR
Measuring the fit of the model - SSR Once we ve determined our estimated regression line, we d like to know how well the model fits. How far/close are the observations to the fitted line? One way to do
More informationTMA4255 Applied Statistics V2016 (5)
TMA4255 Applied Statistics V2016 (5) Part 2: Regression Simple linear regression [11.1-11.4] Sum of squares [11.5] Anna Marie Holand To be lectured: January 26, 2016 wiki.math.ntnu.no/tma4255/2016v/start
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationWISE International Masters
WISE International Masters ECONOMETRICS Instructor: Brett Graham INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This examination paper contains 32 questions. You are
More informationEconometrics I Lecture 3: The Simple Linear Regression Model
Econometrics I Lecture 3: The Simple Linear Regression Model Mohammad Vesal Graduate School of Management and Economics Sharif University of Technology 44716 Fall 1397 1 / 32 Outline Introduction Estimating
More informationLecture 14 Simple Linear Regression
Lecture 4 Simple Linear Regression Ordinary Least Squares (OLS) Consider the following simple linear regression model where, for each unit i, Y i is the dependent variable (response). X i is the independent
More informationUses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).
1 Economics 620, Lecture 8a: Asymptotics II Uses of Asymptotic Distributions: Suppose X n! 0 in probability. (What can be said about the distribution of X n?) In order to get distribution theory, we need
More informationLecture 15. Hypothesis testing in the linear model
14. Lecture 15. Hypothesis testing in the linear model Lecture 15. Hypothesis testing in the linear model 1 (1 1) Preliminary lemma 15. Hypothesis testing in the linear model 15.1. Preliminary lemma Lemma
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 11, 2012 Outline Heteroskedasticity
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationQuestions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares
Questions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares L Magee Fall, 2008 1 Consider a regression model y = Xβ +ɛ, where it is assumed that E(ɛ X) = 0 and E(ɛɛ X) =
More informationLIST OF FORMULAS FOR STK1100 AND STK1110
LIST OF FORMULAS FOR STK1100 AND STK1110 (Version of 11. November 2015) 1. Probability Let A, B, A 1, A 2,..., B 1, B 2,... be events, that is, subsets of a sample space Ω. a) Axioms: A probability function
More informationRewrap ECON November 18, () Rewrap ECON 4135 November 18, / 35
Rewrap ECON 4135 November 18, 2011 () Rewrap ECON 4135 November 18, 2011 1 / 35 What should you now know? 1 What is econometrics? 2 Fundamental regression analysis 1 Bivariate regression 2 Multivariate
More informationWooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model
Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Most of this course will be concerned with use of a regression model: a structure in which one or more explanatory
More informationEconometrics Summary Algebraic and Statistical Preliminaries
Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L
More informationEconometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018
Econometrics I KS Module 1: Bivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: March 12, 2018 Alexander Ahammer (JKU) Module 1: Bivariate
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More information3. Probability and Statistics
FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationReview of Statistics
Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and
More informationIntroduction to Estimation Methods for Time Series models. Lecture 1
Introduction to Estimation Methods for Time Series models Lecture 1 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 1 SNS Pisa 1 / 19 Estimation
More informationthe error term could vary over the observations, in ways that are related
Heteroskedasticity We now consider the implications of relaxing the assumption that the conditional variance Var(u i x i ) = σ 2 is common to all observations i = 1,..., n In many applications, we may
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More informationSTA 2201/442 Assignment 2
STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution
More informationBusiness Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'
Business Statistics Tommaso Proietti DEF - Università di Roma 'Tor Vergata' Linear Regression Specication Let Y be a univariate quantitative response variable. We model Y as follows: Y = f(x) + ε where
More informationLECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit
LECTURE 6 Introduction to Econometrics Hypothesis testing & Goodness of fit October 25, 2016 1 / 23 ON TODAY S LECTURE We will explain how multiple hypotheses are tested in a regression model We will define
More informationMAT2377. Rafa l Kulik. Version 2015/November/26. Rafa l Kulik
MAT2377 Rafa l Kulik Version 2015/November/26 Rafa l Kulik Bivariate data and scatterplot Data: Hydrocarbon level (x) and Oxygen level (y): x: 0.99, 1.02, 1.15, 1.29, 1.46, 1.36, 0.87, 1.23, 1.55, 1.40,
More informationSTAT 540: Data Analysis and Regression
STAT 540: Data Analysis and Regression Wen Zhou http://www.stat.colostate.edu/~riczw/ Email: riczw@stat.colostate.edu Department of Statistics Colorado State University Fall 205 W. Zhou (Colorado State
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationPart IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015
Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationLinear Regression. Junhui Qian. October 27, 2014
Linear Regression Junhui Qian October 27, 2014 Outline The Model Estimation Ordinary Least Square Method of Moments Maximum Likelihood Estimation Properties of OLS Estimator Unbiasedness Consistency Efficiency
More informationSimple and Multiple Linear Regression
Sta. 113 Chapter 12 and 13 of Devore March 12, 2010 Table of contents 1 Simple Linear Regression 2 Model Simple Linear Regression A simple linear regression model is given by Y = β 0 + β 1 x + ɛ where
More informationEconometrics I. Introduction to Regression Analysis (Outline of the course) Marijus Radavi ius
Econometrics I Introduction to Regression Analysis (Outline of the course) Marijus Radavi ius Vilnius 2012 Content 1 Introduction 2 1.1 Motivation................................. 2 1.2 Terminology................................
More informationLecture 4: Heteroskedasticity
Lecture 4: Heteroskedasticity Econometric Methods Warsaw School of Economics (4) Heteroskedasticity 1 / 24 Outline 1 What is heteroskedasticity? 2 Testing for heteroskedasticity White Goldfeld-Quandt Breusch-Pagan
More informationCh 3: Multiple Linear Regression
Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery
More informationOrdinary Least Squares Regression
Ordinary Least Squares Regression Goals for this unit More on notation and terminology OLS scalar versus matrix derivation Some Preliminaries In this class we will be learning to analyze Cross Section
More informationAdvanced Quantitative Methods: ordinary least squares
Advanced Quantitative Methods: Ordinary Least Squares University College Dublin 31 January 2012 1 2 3 4 5 Terminology y is the dependent variable referred to also (by Greene) as a regressand X are the
More informationSimple Linear Regression
Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.
More informationLinear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,
Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,
More informationQuantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017
Summary of Part II Key Concepts & Formulas Christopher Ting November 11, 2017 christopherting@smu.edu.sg http://www.mysmu.edu/faculty/christophert/ Christopher Ting 1 of 16 Why Regression Analysis? Understand
More information1. The Multivariate Classical Linear Regression Model
Business School, Brunel University MSc. EC550/5509 Modelling Financial Decisions and Markets/Introduction to Quantitative Methods Prof. Menelaos Karanasos (Room SS69, Tel. 08956584) Lecture Notes 5. The
More informationChapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE
Chapter 6. Panel Data Joan Llull Quantitative Statistical Methods II Barcelona GSE Introduction Chapter 6. Panel Data 2 Panel data The term panel data refers to data sets with repeated observations over
More informationGARCH Models Estimation and Inference
GARCH Models Estimation and Inference Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 1 Likelihood function The procedure most often used in estimating θ 0 in
More informationApplied Econometrics (QEM)
Applied Econometrics (QEM) based on Prinicples of Econometrics Jakub Mućk Department of Quantitative Economics Jakub Mućk Applied Econometrics (QEM) Meeting #3 1 / 42 Outline 1 2 3 t-test P-value Linear
More informationSimple Linear Regression
Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)
More informationLecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is
Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is Q = (Y i β 0 β 1 X i1 β 2 X i2 β p 1 X i.p 1 ) 2, which in matrix notation is Q = (Y Xβ) (Y
More informationSimple Linear Regression. Material from Devore s book (Ed 8), and Cengagebrain.com
12 Simple Linear Regression Material from Devore s book (Ed 8), and Cengagebrain.com The Simple Linear Regression Model The simplest deterministic mathematical relationship between two variables x and
More informationSTAT5044: Regression and Anova. Inyoung Kim
STAT5044: Regression and Anova Inyoung Kim 2 / 47 Outline 1 Regression 2 Simple Linear regression 3 Basic concepts in regression 4 How to estimate unknown parameters 5 Properties of Least Squares Estimators:
More informationPeter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8
Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall
More informationThe Simple Regression Model. Part II. The Simple Regression Model
Part II The Simple Regression Model As of Sep 22, 2015 Definition 1 The Simple Regression Model Definition Estimation of the model, OLS OLS Statistics Algebraic properties Goodness-of-Fit, the R-square
More informationAsymptotic Theory. L. Magee revised January 21, 2013
Asymptotic Theory L. Magee revised January 21, 2013 1 Convergence 1.1 Definitions Let a n to refer to a random variable that is a function of n random variables. Convergence in Probability The scalar a
More informationp(z)
Chapter Statistics. Introduction This lecture is a quick review of basic statistical concepts; probabilities, mean, variance, covariance, correlation, linear regression, probability density functions and
More informationQualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf
Part : Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section
More informationChapter 2. Continued. Proofs For ANOVA Proof of ANOVA Identity. the product term in the above equation can be simplified as n
Chapter 2. Continued Proofs For ANOVA Proof of ANOVA Identity We are going to prove that Writing SST SSR + SSE. Y i Ȳ (Y i Ŷ i ) + (Ŷ i Ȳ ) Squaring both sides summing over all i 1,...n, we get (Y i Ȳ
More informationECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes
ECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes ED HERBST September 11, 2017 Background Hamilton, chapters 15-16 Trends vs Cycles A commond decomposition of macroeconomic time series
More informationRegression Models - Introduction
Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent
More informationwhere x and ȳ are the sample means of x 1,, x n
y y Animal Studies of Side Effects Simple Linear Regression Basic Ideas In simple linear regression there is an approximately linear relation between two variables say y = pressure in the pancreas x =
More informationGeneral Linear Model: Statistical Inference
Chapter 6 General Linear Model: Statistical Inference 6.1 Introduction So far we have discussed formulation of linear models (Chapter 1), estimability of parameters in a linear model (Chapter 4), least
More informationEconometrics Multiple Regression Analysis: Heteroskedasticity
Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties
More informationConfidence Intervals, Testing and ANOVA Summary
Confidence Intervals, Testing and ANOVA Summary 1 One Sample Tests 1.1 One Sample z test: Mean (σ known) Let X 1,, X n a r.s. from N(µ, σ) or n > 30. Let The test statistic is H 0 : µ = µ 0. z = x µ 0
More informationLinear Regression. Simple linear regression model determines the relationship between one dependent variable (y) and one independent variable (x).
Linear Regression Simple linear regression model determines the relationship between one dependent variable (y) and one independent variable (x). A dependent variable is a random variable whose variation
More informationLecture 11. Probability Theory: an Overveiw
Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 16, 2013 Outline Introduction Simple
More informationScatter plot of data from the study. Linear Regression
1 2 Linear Regression Scatter plot of data from the study. Consider a study to relate birthweight to the estriol level of pregnant women. The data is below. i Weight (g / 100) i Weight (g / 100) 1 7 25
More informationQuick Review on Linear Multiple Regression
Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,
More informationEconometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague
Econometrics Week 4 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 23 Recommended Reading For the today Serial correlation and heteroskedasticity in
More informationSimple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.
Statistical Computation Math 475 Jimin Ding Department of Mathematics Washington University in St. Louis www.math.wustl.edu/ jmding/math475/index.html October 10, 2013 Ridge Part IV October 10, 2013 1
More informationChapter 12 - Lecture 2 Inferences about regression coefficient
Chapter 12 - Lecture 2 Inferences about regression coefficient April 19th, 2010 Facts about slope Test Statistic Confidence interval Hypothesis testing Test using ANOVA Table Facts about slope In previous
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More informationRecent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data
Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationComprehensive Examination Quantitative Methods Spring, 2018
Comprehensive Examination Quantitative Methods Spring, 2018 Instruction: This exam consists of three parts. You are required to answer all the questions in all the parts. 1 Grading policy: 1. Each part
More informationSTA 2101/442 Assignment 3 1
STA 2101/442 Assignment 3 1 These questions are practice for the midterm and final exam, and are not to be handed in. 1. Suppose X 1,..., X n are a random sample from a distribution with mean µ and variance
More informationMath Camp II. Calculus. Yiqing Xu. August 27, 2014 MIT
Math Camp II Calculus Yiqing Xu MIT August 27, 2014 1 Sequence and Limit 2 Derivatives 3 OLS Asymptotics 4 Integrals Sequence Definition A sequence {y n } = {y 1, y 2, y 3,..., y n } is an ordered set
More informationLarge Sample Properties of Estimators in the Classical Linear Regression Model
Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in
More informationSTAT420 Midterm Exam. University of Illinois Urbana-Champaign October 19 (Friday), :00 4:15p. SOLUTIONS (Yellow)
STAT40 Midterm Exam University of Illinois Urbana-Champaign October 19 (Friday), 018 3:00 4:15p SOLUTIONS (Yellow) Question 1 (15 points) (10 points) 3 (50 points) extra ( points) Total (77 points) Points
More informationREGRESSION ANALYSIS AND INDICATOR VARIABLES
REGRESSION ANALYSIS AND INDICATOR VARIABLES Thesis Submitted in partial fulfillment of the requirements for the award of degree of Masters of Science in Mathematics and Computing Submitted by Sweety Arora
More informationChapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability
Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with
More informationReview: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.
1 Review: Let X 1, X,..., X n denote n independent random variables sampled from some distribution might not be normal!) with mean µ) and standard deviation σ). Then X µ σ n In other words, X is approximately
More informationMultivariate Analysis and Likelihood Inference
Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density
More information