Lecture 1, Jan 19. i=1 p i = 1.

Size: px
Start display at page:

Download "Lecture 1, Jan 19. i=1 p i = 1."

Transcription

1 Lecture 1, Ja 19 Review of the expected value, covariace, correlatio coefficiet, mea, ad variace. Radom variable. A variable that takes o alterative values accordig to chace. More specifically, a radom variable assumes differet values, each with probability less tha or equal to 1. A cotiuous radom variable may take o ay value o the real umber lie. A discrete radom variable may take oly a specific umber of values. Probability mass fuctio (pmf) ad probability desity fuctio (pdf). The process that geerates the values of a radom variable. It lists all possible outcomes ad the probability that each will occur. Expected values. The mea, or the expected value, of a radom variable X is a weighted average of the possible outcomes, where the probabilities of the outcomes serve as the weights. Expectatio operator deoted E, mea of X deoted µ X. I the discrete case, N µ X = E(X) = p 1 X 1 + p 2 X p N X N = p i X i, where p i = P (X = X i ) ad N p i = 1. I the cotiuous case, µ X = E(X) = Ω f(x)xdx, where Ω is the support of X, ad f(x) is the pdf of X. The sample mea of a set of outcomes o X is deoted by X. X = 1 x i, where x 1,..., x are realizatios of X. 1

2 Variace. The variace of a radom variable provides a measure of the spread, or dispersio, aroud the mea. Var(X) = σx 2 = E[X E(X)] 2. I the discrete case, N Var(X) = σx 2 = p i [X i E(X)] 2. The positive square root of the variace is called the stadard deviatio. Joit distributio. I the discrete case, joit distributios of X ad Y are described by a list of probabilities of occurrece of all possible outcomes o both X ad Y. The covariace of X ad Y is Cov(X, Y ) = E[X E(X)(Y E(Y )]. I the discrete case, N M Cov(X, Y ) = p ij [X i E(X)][Y j E(Y )], j=1 where p ij represets the joit probability of X = X i ad Y = Y j occurrig. The covariace is a measure of the liear associatio betwee X ad Y. If both variables are always above ad below their meas at the same time, the covariace will be positive. If X is above its mea whe Y is below its mea ad vice versa, the covariace will be egative. The value of covariace depeds upo the uits i which X ad Y are measured. The correlatio coefficiet ρ XY is a measure of the associatio which has bee ormalized ad is scale-free. Cov(X, Y) ρ XY =, σ X σ Y where σ X ad σ Y represet the stadard deviatio of X ad Y respectively. The correlatio coefficiet is always betwee -1 ad 1. A positive correlatio idicates that the variables move i the same directio, while a egative correlatio implies that they move i opposite directios. Properties. Suppose X ad Y are radom variables, a ad b are costats: 1. E(aX + by ) = ae(x) + be(y ); 2. Var(aX + b) = a 2 Var(X); 3. Var(X + Y ) = Var(X) + Var(Y ) + 2Cov(X, Y ); 4. If X ad Y are idepedet, the Cov(X, Y ) = 0. 2

3 Estimatio. Meas, variaces ad covariaces ca be measured with certaity oly if we kow all there is to kow about all possible outcomes. I practice, we may obtai a sample of the relevat iformatio eeded. Give x 1,..., x are radom observatios of X, we wat to estimate a populatio parameter (like the mea or the variace). The sample mea X is a ubiased estimator of the populatio mea µ X. E( X) = E( 1 x i ) = 1 E(x i ) = 1 µ X = µ X. The sample variace s 2 X = 1 (x 1 i X) 2 is a ubiased estimator of the populatio variace Var(X). The sample covariace s XY = 1 (x 1 i X)(y i Ȳ ) is a ubiased estimator of the populatio covariace Cov(X, Y ). The sample correlatio coefficiet is defied as r XY = (x i X)(y i Ȳ ) (x i X) 2 (y i Ȳ )2. Desired properties of estimators: 1. Lack of bias. The bias associated with a estimated parameter is defied to be: Bias( ˆβ) = E( ˆβ) β. ˆβ is a ubiased estimator if the mea or the expected value of ˆβ is equal to the true value, that is, E( ˆβ) = β. 2. Cosistecy. ˆβ is a cosistet estimator of β if for ay δ > 0, lim P ( β ˆβ < δ) = 1. As sample size approaches ifiity, the probability that ˆβ will differ from β will get very small. 3. Efficiecy. We say that ˆβ is a efficiet ubiased estimator if for a give sample size, the variace of ˆβ is smaller tha the variace of ay other ubiased estimators. 3

4 Lecture 2, Ja 26 Tradeoff betwee bias ad variace of estimators. Whe the goal is to maximize the precisio of the predictios, a estimator with low variace ad some bias may be more desirable tha a ubiased estimator with high variace. We may wat to miimize the mea square error, defied as Mea square error( ˆβ) = E( ˆβ β) 2. It ca be show that Mea square error( ˆβ) = [Bias( ˆβ)] 2 +Var( ˆβ). The criterio of miimizig mea square error take ito accout of both the variace ad the bias of the estimator. A alterative criterio to cosistecy is that the mea square error of the estimator approaches zero as the sample size icreases. This implies asymptotically, or whe the sample size is very large, the estimator is ubiased ad its variace goes to zero. A estimator with a mea square error that approaches zero will be cosistet estimator but that the reverse eed ot be true. 1

5 Probability distributios. The Normal distributio is a cotiuous bell-shaped probability distributio. It ca be fully described by its mea ad its variace. If X is ormally distributed, we write X N(µ X, σ 2 X), which is read X is ormally distributed with mea µ X ad variace σ 2 X. The probability that a sigle observatio of a ormally distributed variable will lie withi 1.96 stadard deviatios of its mea is approximately The probability that a sigle observatio of a ormally distributed variable will lie withi 2.57 stadard deviatios of its mea is approximately P (µ X 1.96σ X < X < µ X σ X ) 0.95, P (µ X 2.57σ X < X < µ X σ X ) Compariso of ormal distributios Desity Distributios Variace=1 Variace=4 Variace= x value The weighted sum of ormal radom variables is still ormal. 2

6 The chi square distributio with N degrees of freedom is the sum of the squares of N idepedetly distributed stadard ormal radom variables (with mea 0 ad variace 1). The chi square distributio starts at the origi, is skewed to the right, ad has a tail which exteds ifiitely to the right. The distributio becomes more ad more symmetric as the umber of degrees of freedom gets larger. It becomes close to ormal distributio whe degrees of freedom is very large. Compariso of chi square distributios Desity Distributios df=3 df=8 df= x value As a example, whe we calculate the sample variaces s 2 of observatios from a ormal distributio with variaces σ 2, ( 1)s 2 /σ 2 is chi square with 1 degrees of freedom. 3

7 The t distributio. Assume X is ormal with mea 0 ad variace 1, Z is chi square with N degrees of freedom, X ad Z are idepedet. The X/ Z/N has a t distributio with N degrees of freedom. The t distributio is symmetric, has fatter tails tha the ormal distributio, ad approximates the ormal distributio whe the degrees of freedom is large. Compariso of t distributios Desity Distributios df=1 df=3 df=8 df=30 ormal x value As a example, X N(µ X, σx), 2 ad X = 1 x i, is the sample mea based o a sample x 1,..., x of size. The ( X µ X )/(σ X / ) is ormal with mea 0 ad variace 1. For ukow σx, 2 we replace σx 2 by sample variace s 2 X, the ( X µx ) ( = X µ X )/(σ X / ) s X [( 1)s 2 X/σX]/( 2 1), which is a t distributio with 1 degrees of freedom. 4

8 The F distributio. If X ad Z are idepedet ad distributed as chi square with N 1 ad N 2 degrees of freedom, respectively, the (X/N 1 )/(Z/N 2 ) is distributed as a F distributio with N 1 ad N 2 degrees of freedom. These two parameters are called the umerator ad the deomiator degrees of freedom. F Distributios Desity Distributios F(33,10) F(5,10) F(2,7) x value 5

9 Forecastig: to calculate or predict some future evet or coditio, usually as a result of ratioal study or aalysis of pertiet data. We all make forecasts: a perso waitig for a bus; parets expectig a telephoe call from their childre; bak maager predicts cash flow for the ext quarter; compay maager predicts sales or estimates umber of ma-hours required to meet a give productio schedule. Future evets ivolve ucertaity. Forecasts are ot perfect. The objective of forecastig is to reduce the forecast error. Each forecast has its ow specificatios, ad solutios to oe are ot solutios i aother situatio. Geeral priciples for ay forecast system: Model specificatio; Model estimatio; Diagostic checkig; Forecast geeratio; Stability checkig; Forecast updatig. Choice of forecast model. (1) Degrees of accuracy required; (2) forecast horizo; (3) budget; (4) what data is available. Oe may ot costruct accurate empirical forecast models from limited ad icomplete data base. Forecast criteria. Actual observatio at time t is z t. Its forecast, which uses the iformatio up to ad icludig time t 1, is z t 1 (1). Objective is to make the future forecast error z t z t 1 (1) as small as possible. However, z t is ukow, we ca oly talk about its expected value, coditioal o the observed data up to ad icludig time t 1. We miimize the mea absolute error E z t z t 1 (1) or the mea square error E[z t z t 1 (1)] 2. We use the mea square error criterio for simpler mathematical calculatios. What we will discuss. I sigle-variable forecastig, we use past history of the series, z t, where t is the time idex, to extrapolate ito the future. I regressio forecastig, we use the relatioships betwee the variable to be forecast ad the other variables. I the most geeral form, the regressio model ca be writte as y t = f(x t1,..., x tp ; β 1,..., β m ) + ε t Describes relatioship betwee oe depedet variable Y ad p idepedet variables X 1,..., X p. Idex t meas at time t or for subject t. At idex t, we observe y t ad x t1,..., x tp. Ukow parameters β 1,..., β m. Kow mathematical fuctio form f. Radomess for this model rises from the error term ε t. 6

10 Lecture 3, Feb 2 I the most geeral form, the regressio model ca be writte as y t = f(x t1,..., x tp ; β 1,..., β m ) + ε t Describes relatioship betwee oe depedet variable Y ad p idepedet variables X 1,..., X p. Idex t meas at time t or for subject t. At idex t, we observe y t ad x t1,..., x tp. Ukow parameters β 1,..., β m. Kow mathematical fuctio form f. Radomess for this model rises from the error term ε t. Models liear i the parameter: 1. y = β 0 + β 1 x 1 + ε; 2. y = β 0 + β 1 x 1 + β 2 x 2 + ε; 3. y = β 0 + β 1 x 1 + β 2 x ε; Models liear i the idepedet variable: 1. y = β 0 + β 1 x 1 + ε; 2. y = β 0 + β 1 x 1 + β 2 x 2 + ε. Regressio through origi. y: salary of a sales perso; x: umber of products sold. No base salary! y: merit I get for icrease i salary; x: umber of papers published. Model: y = βx + ε. 1

11 Give observatios (x 1, y 1 ),..., (x, y ), we wat to miimize S(β) = (y i βx i ) 2. Take derivatives ad set to 0, the we have S (β) = ds(β) dβ = 2 (y i βx i )x i = 0 The least squares estimator of β is ˆβ = x iy i. x2 i Take secod order derivative to make sure this is a miimizer. Check S (β) = d2 S(β) dβ 2 = ds (β) dβ = 2 x 2 i 0. Simple liear regressio. y: resale price of a preowed car; x: milage! y: ruig time for 10-km road race; x: maximal aerobic capacity (oxyge uptake, milliliter per kilogram per miite, ml/(kg mi) Show MINITAB example! Scatter plot! Model: y = β 0 + β 1 x + ε. Give observatios (x 1, y 1 ),..., (x, y ), we wat to miimize S(β 0, β 1 ) = (y i β 0 β 1 x i ) 2. Take derivatives ad set to 0, the we have β 0 S(β 0, β 1 ) = 2 (y i β 0 β 1 x i ) = 0 2

12 β 1 S(β 0, β 1 ) = 2 (y i β 0 β 1 x i )x i = 0 We have two equatios ad two ukows, solve for β 0 ad β 1. The we have ˆβ 1 = s xy = (x i x)(y i ȳ) s 2 x (x, i x) 2 ˆβ0 = ȳ ˆβ 1 x, where ( s xy = 1 x i y i x ) i y i = (x i x)(y i ȳ) 1 1 [ ] s 2 x = 1 x 2 i ( x i) 2 = (x i x) We still eed to take secod order derivatives to make sure ( ˆβ 0, ˆβ 1 ) are actual miimizers. Model assumptios: 1. The relatioship betwee x ad y is liear, y i = β 0 + β 1 x i + ε i ; 2. The x are o-stochastic variable whose values x 1,..., x are fixed; 3. Normality The error term ε is ormally distributed; Homoscedasticity The error term ε has mea zero ad costat variace for all observatios, E(ε i ) = 0, Var(ε i ) = σ 2 ; Idepedecy The radom error ε i ad ε j are idepedet of each other for i j, or errors correspodig to differet observatios are idepedet. How do we check those assumptios? DO AN EXAMPLE BY HAND. Show MINITAB RESULTS! ˆβ 1, ˆβ 0 as weighted sum x Let w i = i x, the ˆβ (x i x) 2 1 = w iy i. We otice that w i = 0, ad w i x i = w i (x i x) = 1. 3

13 Thus, E( ˆβ 1 ) = E( w i y i ) = w i E(y i ) = = β 0 w i + β 1 w i x i = β 1, w i (β 0 + β 1 x i ) which meas ˆβ 1 is ubiased estimator of β 1. Furthermore wi 2 = (x i x) 2 [ (x i x) 2 ] 2 = (x i x) 2 [ (x i x) 2 ] = 1 2 (x i x), 2 ad Var( ˆβ 1 ) = Var( w i y i ) = wi 2 Var(y i ) = σ 2 wi 2 = σ 2 (x i x) 2. Similarly, ˆβ 0 = ȳ ˆβ 1 x = 1 E( ˆβ 0 ) = y i x w iy i = ( 1 xw i)y i. Thus ( 1 xw i)e(y i ) = ( 1 xw i)(β 0 + β 1 x i ) = β 0 xβ 0 w i + β 1 x β 1 x w i x i = β 0. which meas ˆβ 0 is ubiased estimator of β 0. Furthermore Var( ˆβ 0 ) = Var[ ( 1 xw i)y i ] = ( 1 xw i) 2 Var(y i ) = σ 2 ( 1 xw i) 2 = σ 2 ( x2 wi 2 2 xw i) = σ 2 ( 1 + x 2 (x i x) ) = 2 We ca also show that Cov( ˆβ 0, ˆβ xσ 2 1 ) = (x i x) 2 σ2 x2 i (x i x) 2 4

14 Gauss-Markov Theorem: If the previous model assumptios are satisfied, the amog all the liear ubiased estimators of β 0 ad β 1, the least squares estimators ˆβ 0 ad ˆβ 1 have the smallest variace. The Theorem implies that for ay ubiased estimator of β 1 with form wi y i, its variace should be Var( ˆβ σ 1 ) = 2. (x i x) 2 The fitted least squares lie: ŷ = ˆβ 0 + ˆβ 1 x is used to get the fitted value y. The i-th residual is the differece betwee the i-th observatio of depedet variable ad its fitted value: r i = y i ŷ i. I practice, Var(ε) = σ 2 is ukow ad we eed to estimate it. The sum of squares of error (SSE; also kow as sum of squares of residual) is SSE = (y i ŷ i ) 2 = (y i ˆβ 0 ˆβ 1 x i ) 2. It has 2 degrees of freedom. We have radom observatios, that s degrees of freedom. Estimatig ˆβ 0 ad ˆβ 1 uses 2 degrees of freedom. The mea squares error (MSE) is defied to be MSE = SSE 2 = (y i ˆβ 0 ˆβ 1 x i ) 2. 2 MSE is also deoted by ˆσ 2. It is a ubiased estimator of σ 2. It ca be show that E(MSE) = E(ˆσ 2 ) = σ 2. The i-th stadardized residual sr i = r i /ˆσ. I summarizatio, we have ˆβ 1 N ˆβ 0 N SSE σ 2 ( ) σ 2 β 1, (x. i x) 2 ( σ 2 ) β 0, x2 i (x. i x) 2 = ( 2)ˆσ2 σ 2 χ

15 Lecture 4, Feb 16 Simple liear regressio. Model: y = β 0 + β 1 x + ε. Model assumptios: 1. The relatioship betwee x ad y is liear, y i = β 0 + β 1 x i + ε i ; 2. The x are o-stochastic variable whose values x 1,..., x are fixed; 3. Normality The error term ε is ormally distributed; Homoscedasticity The error term ε has mea zero ad costat variace for all observatios, E(ε i ) = 0, Var(ε i ) = σ 2 ; Idepedecy The radom error ε i ad ε j are idepedet of each other for i j, or errors correspodig to differet observatios are idepedet. Give observatios (x 1, y 1 ),..., (x, y ), we wat to miimize S(β 0, β 1 ) = (y i β 0 β 1 x i ) 2. Take derivatives ad set to 0, the we have S(β 0, β 1 ) = 2 (y i β 0 β 1 x i ) = 0 β 0 β 1 S(β 0, β 1 ) = 2 (y i β 0 β 1 x i )x i = 0 We have two equatios ad two ukows, solve for β 0 ad β 1. The we have ˆβ 1 = s xy = (x i x)(y i ȳ) s 2 x (x, i x) 2 ˆβ0 = ȳ ˆβ 1 x, where 1

16 ( s xy = 1 x i y i x i y i 1 [ s 2 x = 1 1 x 2 i ( x i) 2 ) ] = = (x i x)(y i ȳ) 1 (x i x) 2 1 We still eed to take secod order derivatives to make sure ( ˆβ 0, ˆβ 1 ) are actual miimizers. The fitted least squares lie: ŷ = ˆβ 0 + ˆβ 1 x is used to get the fitted value y. The i-th residual is the differece betwee the i-th observatio of depedet variable ad its fitted value: r i = y i ŷ i. I practice, Var(ε) = σ 2 is ukow ad we eed to estimate it. The sum of squares of error (SSE; also kow as sum of squares of residual) is SSE = (y i ŷ i ) 2 = (y i ˆβ 0 ˆβ 1 x i ) 2. It has 2 degrees of freedom. We have radom observatios, that s degrees of freedom. Estimatig ˆβ 0 ad ˆβ 1 uses 2 degrees of freedom. The mea squares error (MSE) is defied to be MSE = SSE 2 = (y i ˆβ 0 ˆβ 1 x i ) 2. 2 MSE is also deoted by ˆσ 2. It is a ubiased estimator of σ 2. It ca be show that E(MSE) = E(ˆσ 2 ) = σ 2. The i-th stadardized residual sr i = r i /ˆσ. I summarizatio, we have ˆβ 1 N ˆβ 0 N SSE σ 2 ( ) σ 2 β 1, (x. i x) 2 ( σ 2 ) β 0, x2 i (x. i x) 2 = ( 2)ˆσ2 σ 2 χ

17 Cofidece iterval Deote σ 2ˆβ1 = σ 2 (x i x) 2, s 2ˆβ1 = ˆσ 2 (x i x) 2, the ˆβ 1 β 1 σ ˆβ1 N(0, 1). For kow σ 2, the 100(1 α)% cofidece iterval for β 1 is ˆβ 1 ± z α/2 σ ˆβ1. For ukow σ 2, plug i s ˆβ1 as a estimate of ukow σ ˆβ1 ad we have t = ˆβ 1 β 1 s ˆβ1 = ( ˆβ 1 β 1 )/σ ˆβ1 = s ( ˆβ1 β 1 )/σ ˆβ 1 2ˆβ1 /σ 2ˆβ1 ( 2)ˆσ 2 /( 2) σ 2 N(0, 1) χ 2 2 /( 2) = t 2 Deote the upper percetile t 2,α/2, which satisfies P (t > t 2,α/2 ) = α/2 for radom variable t t 2. Thus ( ) P rob t 2,α/2 < ˆβ 1 β 1 s ˆβ1 < t 2,α/2 = 100(1 α)% P rob( ˆβ 1 t 2,α/2 s ˆβ1 < β 1 < ˆβ 1 + t 2,α/2 s ˆβ1 ) = 100(1 α)% The 100(1 α)% cofidece iterval for β 1 is ˆβ 1 ± t 2,α/2 s ˆβ1. Hypothesis test We are iterested whether there is sigificat predictor effect. Null hypothesis: H 0 : β 1 = 0 v.s. H 1 : β 1 0. Test statistics: t = ˆβ 1 s ˆβ1. Uder H 0, t t 2. At 1 α sigificace level, reject H 0 if t > t 2,α/2. We are iterested whether there is sigificat positive predictor effect. Null hypothesis: H 0 : β 1 = 0 v.s. H 0 : β 1 > 0. Test statistics: t = ˆβ 1 s ˆβ1. Uder H 0, t t 2. At 1 α sigificace level, reject H 0 if t > t 2,α. We are iterested whether the slope is at a certai specified level c. Null hypothesis: H 0 : β 1 = c v.s. H 0 : β 1 c. Test statistics: t = ˆβ 1 c s ˆβ1. Uder H 0, t t 2. At 1 α sigificace level, reject H 0 if t > t 2,α/2. 3

18 Lecture 5, Feb 19 Please check with your classmates for missed lecture otes. The materials listed here are the key poits for ANOVA. Make sure you uderstad EV- ERYTHING listed here. We are goig to get back to ANOVA whe we deal with multiple liear regressio. This versio does ot have ay examples or demostratio i MINITAB. TUCAPTURE should be available soo. ANOVA The variatio of y has the followig decompositio (y i ȳ) 2 = (ŷ i ȳ) 2 + (y i ŷ i ) 2 Total Sum of Squares (SST)= Sum of Squares of Regressio (SSR)+ Sum of Squares of Error(SSE) R 2, or the R squared, coefficiet of determiatio, is the proportio of variatio i y explaied by the regressio equatio. R 2 = SSR SST = 1 SSE SST The sample correlatio coefficiet R = ± R 2. R > 0 whe ˆβ 1 > 0; ad R < 0 whe ˆβ 1 < 0. R measures how closely the data fit a straight lie. ANOVA table Source df Sum of Squares Mea Squares F Regressio 1 SSR = (ŷ i ȳ) 2 MSR=SSR/1 F=MSR/MSE Error -2 SSE = (y i ŷ i ) 2 MSE=SSE/(-2) Total -1 SST = (y i ȳ) 2 1

19 To test whether there is sigificat predictor effect, we do t use R 2. We use F, which takes ito cosideratio of the degrees of freedom. Null hypothesis: H 0 : β 1 = 0 v.s. H 1 : β 1 0. If test statistic F is large, which meas a large proportio of variatio i y is explaied by the regressio, the ituitio is that we reject H 0. It ca be show that uder H 0, F = MSR/MSE F 1, 2. Deote the upper percetile F 1, 2,α, which satisfies P (F > F 1, 2,α ) = α for radom variable F F 1, 2. At 1 α sigificace level, reject H 0 if F = MSR/MSE > F 1, 2,α. Notice that F = MSR MSE = SSR SSE/( 2) = ( 2)SSR SST SSR = F = ( 2)R2 1 R 2 ( 2)SSR/SST SST/SST SSR/SST Large R 2 with small may ot be sigificat; but moderate R 2 with large sample size ca be highly sigificat. It ca be show that the t test ad the F test agree with each other. 2

20 Lecture 6, Feb 23 Model assumptios: 1. The relatioship betwee x ad y is liear, y i = β 0 + β 1 x i + ε i ; 2. The x are o-stochastic variable whose values x 1,..., x are fixed; 3. Normality The error term ε is ormally distributed; Homoscedasticity The error term ε has mea zero ad costat variace for all observatios, E(ε i ) = 0, Var(ε i ) = σ 2 ; Idepedecy The radom error ε i ad ε j are idepedet of each other for i j, or errors correspodig to differet observatios are idepedet. Gauss-Markov Theorem: If the previous model assumptios are satisfied, the amog all the liear ubiased estimators of β 0 ad β 1, the least squares estimators ˆβ 0 ad ˆβ 1 have the smallest variace. The Theorem implies that for ay ubiased estimator of β 1 with form wi y i, its variace should be Var( ˆβ σ 1 ) = 2. (x i x) 2 Predictio Iterval Cosider a college tryig to predict first year GPA of a ew studet based o the studet s high school GPA. Based o preset studets, the regressio equatio is ŷ = x, where x =high school GPA ad y = first year college GPA. For a studet with high school GPA x ew = 3.5, we estimate its first year college GPA by ŷ ew = ˆβ 0 + ˆβ 1 x ew = = 3.2. The mea of the forecast error is E(y ew ŷ ew ) = E[(β 0 + β 1 x ew + ε ew ) ( ˆβ 0 + ˆβ 1 x ew )] = 0. 1

21 By the Gauss-Markov theorem, ŷ ew is the miimum mea square error forecast amog all liear ubiased forecasts. The variace of the forecast error is Var(y ew ŷ ew ) = Var[(β 0 + β 1 x ew + ε ew ) ( ˆβ 0 + ˆβ 1 x ew )] = σ 2 + Var( ˆβ 0 ) + x 2 ewvar( ˆβ 1 ) + 2x ew Cov( ˆβ 0, ˆβ 1 ) = σ 2 + σ2 x2 i (x i x) + x2 ewσ 2 2 (x i x) 2x ew xσ 2 2 (x i x) ( 2 = σ (x ) ew x) 2 (x i x) 2 For kow σ 2, the predictio iterval for y ew is ( ŷ ew ± z α/2 σ (x ew x) 2 ). (x i x) 2 is For ukow σ 2, plug i MSE = ˆσ 2 ad the predictio iterval for y ew ( ŷ ew ± t 2,α/2 ˆσ (x ew x) 2 ). (x i x) 2 Which predictio iterval is wider? Why? Cofidece Iterval for expectatio of y If we are ot iterested i just oe particular studet with high school GPA=3.5, but we are iterested i the expectatio of college GPA for all studets with high school GPA=3.5. With kow σ 2, the Cofidece Iterval for y 0 = E(y x 0 ) is ŷ 0 ± z α/2 σ 2 ( 1 + (x 0 x) 2 (x i x) 2 ). For ukow σ 2, plug i MSE = ˆσ 2 ad the predictio iterval for y 0 is ( 1 ŷ 0 ± t 2,α/2 ˆσ 2 + (x 0 x) 2 ). (x i x) 2 2

22 Here ŷ 0 = ˆβ 0 + ˆβ 1 x 0. Is the predictio iterval wider or the cofidece iterval wider? Why? Multiple liear regressio Istead of studyig the relatioship betwee respose y ad oe idepedet variable (predictor) x, we may also use liear model to study the relatioship betwee respose y ad multiple idepedet variables x 1,..., x p. y = β 0 + β 1 x β p x p + ε. We have set of observatios (y i ; x i1, x i2,..., x ip ) for i = 1,...,. We obtai least square estimators of the ukow parameters by miimizig the SSE: S(β 1,..., β p ) = [y i (β 0 + β 1 x i1 + + β p x ip )] 2. Set the first order derivatives equal to 0: S = 0, S = 0,, S = 0. β 0 β 1 β p The solutios are deoted by ˆβ 0, ˆβ 1,..., ˆβ p. The fitted least squares lie is ŷ = ˆβ 0 + ˆβ 1 x ˆβ p x p. The i-th residual is r i = y i ŷ i. ANOVA The variatio of y has the same decompositio as i the simple liear regressio (y i ȳ) 2 = (ŷ i ȳ) 2 + (y i ŷ i ) 2 SST=SSR+ SSE 3

23 ANOVA table Source df Sum of Squares Mea Squares F Regressio p SSR = (ŷ i ȳ) 2 MSR=SSR/p F=MSR/MSE Error -p-1 SSE = (y i ŷ i ) 2 MSE=SSE/(-p-1) Total -1 SST = (y i ȳ) 2 To test whether there is sigificat predictor effect, simultaeous test Null hypothesis H 0 : β 1 = β 2 = = β p = 0 v.s. Alterative hypothesis H 1 : At least oe coefficiet ot equal to zero. If test statistic F is large, which meas a large proportio of variatio i y is explaied by the regressio, the ituitio is that we reject H 0. It ca be show that uder H 0, F = MSR/MSE F p, p 1. Deote the upper percetile F p, p 1,α, which satisfies P (F > F p, p 1,α ) = α for radom variable F F p, p 1. At 1 α sigificace level, reject H 0 if F = MSR/MSE > F p, p 1,α, ad this meas the regressio is sigificat. Fail to reject H 0 if F = MSR/MSE < F p, p 1,α, ad this meas oe of the regressio coefficiets sigificatly differs from 0. Notice that F = MSR MSE = = p 1 p SSR/(p) SSE/( p 1) = p 1 SSR p SST SSR SSR/SST SST/SST SSR/SST = p 1 R 2 p 1 R. 2 The R 2, or coefficiet of determiatio, is R 2 = SSR/SST = 1 SSE/SST. R 2 always icrease as the umber of predictors p icrease, which may result i model overfittig. We eed a ew criterio. The corrected R 2, or R 2, is R 2 = 1 SSE/( p 1). It takes degrees of SST/( 1) freedom ito cosideratio ad pealizes overfittig. 4

24 Lecture 7, March 2 Please read Chapter 2 of the textbook, from page 8 to page 41. You may skip Sectio ad Sectio For the basics of liear algebra about matrix calculatio, you may fid the etry from Wikipedia useful: http : //e.wikipedia.org/wiki/matrix ( mathematics) Matrix represetatio of multiple liear regressio We have set of idepedet observatios (y i ; x i1, x i2,..., x ip ) from model Let y = β 0 + β 1 x β p x p + ε. y 1 β 0 + β 1 x β p x 1p + ε 1 y 2. = β 0 + β 1 x β p x 2p + ε 2. y β 0 + β 1 x β p x p + ε y 1 β 0 1 x 11 x 1p ε 1 y 2 Y =., β = β 1., X = 1 x 21 x 2p...., ε = ε 2.. y β p 1 x 1 x p ε The we have Y = Xβ + ε, where ε is multivariate ormal with E(ε) = 0 ad Var(ε) = σ 2 I The SSE ca be writte as S(β) = [y i (β 0 + β 1 x i1 + + β p x ip )] 2 = (Y Xβ) (Y Xβ). 1

25 Take derivatives with respect to β ad we have the followig ormal equatio: X Xβ = X Y, which leads to least squares estimator ˆβ = (X X) 1 X Y. Because E(ˆβ) = E[(X X) 1 X Y] = (X X) 1 X E(Y) = (X X) 1 X E(Xβ + ε) = (X X) 1 X [Xβ + E(ε)] = (X X) 1 X Xβ = I p+1 β = β, ˆβ is ubiased estimator of β. Because Var(AX) = AVar(X)A for radom vector X ad costat matrix A, we have Var(ˆβ) = Var[(X X) 1 X Y] = [(X X) 1 X ]Var(Y)[(X X) 1 X ] = Var[(X X) 1 X Y] = [(X X) 1 X ]Var(Xβ + ε)[(x X) 1 X ] = σ 2 (X X) 1 X I X(X X) 1 = σ 2 (X X) 1. Thus we kow ˆβ N p+1 (β, σ 2 (X X) 1 ). Thus ˆβ i N(β i, σ 2 d i ), where d i is the i-th elemet o the diagoal of matrix (X X) 1. 2

1 Inferential Methods for Correlation and Regression Analysis

1 Inferential Methods for Correlation and Regression Analysis 1 Iferetial Methods for Correlatio ad Regressio Aalysis I the chapter o Correlatio ad Regressio Aalysis tools for describig bivariate cotiuous data were itroduced. The sample Pearso Correlatio Coefficiet

More information

Random Variables, Sampling and Estimation

Random Variables, Sampling and Estimation Chapter 1 Radom Variables, Samplig ad Estimatio 1.1 Itroductio This chapter will cover the most importat basic statistical theory you eed i order to uderstad the ecoometric material that will be comig

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Properties and Hypothesis Testing

Properties and Hypothesis Testing Chapter 3 Properties ad Hypothesis Testig 3.1 Types of data The regressio techiques developed i previous chapters ca be applied to three differet kids of data. 1. Cross-sectioal data. 2. Time series data.

More information

Simple Linear Regression

Simple Linear Regression Simple Liear Regressio 1. Model ad Parameter Estimatio (a) Suppose our data cosist of a collectio of pairs (x i, y i ), where x i is a observed value of variable X ad y i is the correspodig observatio

More information

3/3/2014. CDS M Phil Econometrics. Types of Relationships. Types of Relationships. Types of Relationships. Vijayamohanan Pillai N.

3/3/2014. CDS M Phil Econometrics. Types of Relationships. Types of Relationships. Types of Relationships. Vijayamohanan Pillai N. 3/3/04 CDS M Phil Old Least Squares (OLS) Vijayamohaa Pillai N CDS M Phil Vijayamoha CDS M Phil Vijayamoha Types of Relatioships Oly oe idepedet variable, Relatioship betwee ad is Liear relatioships Curviliear

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Linear Regression Models

Linear Regression Models Liear Regressio Models Dr. Joh Mellor-Crummey Departmet of Computer Sciece Rice Uiversity johmc@cs.rice.edu COMP 528 Lecture 9 15 February 2005 Goals for Today Uderstad how to Use scatter diagrams to ispect

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara Poit Estimator Eco 325 Notes o Poit Estimator ad Cofidece Iterval 1 By Hiro Kasahara Parameter, Estimator, ad Estimate The ormal probability desity fuctio is fully characterized by two costats: populatio

More information

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2. SAMPLE STATISTICS A radom sample x 1,x,,x from a distributio f(x) is a set of idepedetly ad idetically variables with x i f(x) for all i Their joit pdf is f(x 1,x,,x )=f(x 1 )f(x ) f(x )= f(x i ) The sample

More information

Statistical Properties of OLS estimators

Statistical Properties of OLS estimators 1 Statistical Properties of OLS estimators Liear Model: Y i = β 0 + β 1 X i + u i OLS estimators: β 0 = Y β 1X β 1 = Best Liear Ubiased Estimator (BLUE) Liear Estimator: β 0 ad β 1 are liear fuctio of

More information

Regression, Inference, and Model Building

Regression, Inference, and Model Building Regressio, Iferece, ad Model Buildig Scatter Plots ad Correlatio Correlatio coefficiet, r -1 r 1 If r is positive, the the scatter plot has a positive slope ad variables are said to have a positive relatioship

More information

Response Variable denoted by y it is the variable that is to be predicted measure of the outcome of an experiment also called the dependent variable

Response Variable denoted by y it is the variable that is to be predicted measure of the outcome of an experiment also called the dependent variable Statistics Chapter 4 Correlatio ad Regressio If we have two (or more) variables we are usually iterested i the relatioship betwee the variables. Associatio betwee Variables Two variables are associated

More information

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample. Statistical Iferece (Chapter 10) Statistical iferece = lear about a populatio based o the iformatio provided by a sample. Populatio: The set of all values of a radom variable X of iterest. Characterized

More information

TAMS24: Notations and Formulas

TAMS24: Notations and Formulas TAMS4: Notatios ad Formulas Basic otatios ad defiitios X: radom variable stokastiska variabel Mea Vätevärde: µ = X = by Xiagfeg Yag kpx k, if X is discrete, xf Xxdx, if X is cotiuous Variace Varias: =

More information

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1. Eco 325/327 Notes o Sample Mea, Sample Proportio, Cetral Limit Theorem, Chi-square Distributio, Studet s t distributio 1 Sample Mea By Hiro Kasahara We cosider a radom sample from a populatio. Defiitio

More information

(all terms are scalars).the minimization is clearer in sum notation:

(all terms are scalars).the minimization is clearer in sum notation: 7 Multiple liear regressio: with predictors) Depedet data set: y i i = 1, oe predictad, predictors x i,k i = 1,, k = 1, ' The forecast equatio is ŷ i = b + Use matrix otatio: k =1 b k x ik Y = y 1 y 1

More information

Correlation Regression

Correlation Regression Correlatio Regressio While correlatio methods measure the stregth of a liear relatioship betwee two variables, we might wish to go a little further: How much does oe variable chage for a give chage i aother

More information

Stat 139 Homework 7 Solutions, Fall 2015

Stat 139 Homework 7 Solutions, Fall 2015 Stat 139 Homework 7 Solutios, Fall 2015 Problem 1. I class we leared that the classical simple liear regressio model assumes the followig distributio of resposes: Y i = β 0 + β 1 X i + ɛ i, i = 1,...,,

More information

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering CEE 5 Autum 005 Ucertaity Cocepts for Geotechical Egieerig Basic Termiology Set A set is a collectio of (mutually exclusive) objects or evets. The sample space is the (collectively exhaustive) collectio

More information

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett Lecture Note 8 Poit Estimators ad Poit Estimatio Methods MIT 14.30 Sprig 2006 Herma Beett Give a parameter with ukow value, the goal of poit estimatio is to use a sample to compute a umber that represets

More information

11 Correlation and Regression

11 Correlation and Regression 11 Correlatio Regressio 11.1 Multivariate Data Ofte we look at data where several variables are recorded for the same idividuals or samplig uits. For example, at a coastal weather statio, we might record

More information

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n. Jauary 1, 2019 Resamplig Methods Motivatio We have so may estimators with the property θ θ d N 0, σ 2 We ca also write θ a N θ, σ 2 /, where a meas approximately distributed as Oce we have a cosistet estimator

More information

ECON 3150/4150, Spring term Lecture 3

ECON 3150/4150, Spring term Lecture 3 Itroductio Fidig the best fit by regressio Residuals ad R-sq Regressio ad causality Summary ad ext step ECON 3150/4150, Sprig term 2014. Lecture 3 Ragar Nymoe Uiversity of Oslo 21 Jauary 2014 1 / 30 Itroductio

More information

Lecture 11 Simple Linear Regression

Lecture 11 Simple Linear Regression Lecture 11 Simple Liear Regressio Fall 2013 Prof. Yao Xie, yao.xie@isye.gatech.edu H. Milto Stewart School of Idustrial Systems & Egieerig Georgia Tech Midterm 2 mea: 91.2 media: 93.75 std: 6.5 2 Meddicorp

More information

Worksheet 23 ( ) Introduction to Simple Linear Regression (continued)

Worksheet 23 ( ) Introduction to Simple Linear Regression (continued) Worksheet 3 ( 11.5-11.8) Itroductio to Simple Liear Regressio (cotiued) This worksheet is a cotiuatio of Discussio Sheet 3; please complete that discussio sheet first if you have ot already doe so. This

More information

Lecture 22: Review for Exam 2. 1 Basic Model Assumptions (without Gaussian Noise)

Lecture 22: Review for Exam 2. 1 Basic Model Assumptions (without Gaussian Noise) Lecture 22: Review for Exam 2 Basic Model Assumptios (without Gaussia Noise) We model oe cotiuous respose variable Y, as a liear fuctio of p umerical predictors, plus oise: Y = β 0 + β X +... β p X p +

More information

Stat 421-SP2012 Interval Estimation Section

Stat 421-SP2012 Interval Estimation Section Stat 41-SP01 Iterval Estimatio Sectio 11.1-11. We ow uderstad (Chapter 10) how to fid poit estimators of a ukow parameter. o However, a poit estimate does ot provide ay iformatio about the ucertaity (possible

More information

7-1. Chapter 4. Part I. Sampling Distributions and Confidence Intervals

7-1. Chapter 4. Part I. Sampling Distributions and Confidence Intervals 7-1 Chapter 4 Part I. Samplig Distributios ad Cofidece Itervals 1 7- Sectio 1. Samplig Distributio 7-3 Usig Statistics Statistical Iferece: Predict ad forecast values of populatio parameters... Test hypotheses

More information

Lecture 3. Properties of Summary Statistics: Sampling Distribution

Lecture 3. Properties of Summary Statistics: Sampling Distribution Lecture 3 Properties of Summary Statistics: Samplig Distributio Mai Theme How ca we use math to justify that our umerical summaries from the sample are good summaries of the populatio? Lecture Summary

More information

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING Lectures MODULE 5 STATISTICS II. Mea ad stadard error of sample data. Biomial distributio. Normal distributio 4. Samplig 5. Cofidece itervals

More information

Statistics 20: Final Exam Solutions Summer Session 2007

Statistics 20: Final Exam Solutions Summer Session 2007 1. 20 poits Testig for Diabetes. Statistics 20: Fial Exam Solutios Summer Sessio 2007 (a) 3 poits Give estimates for the sesitivity of Test I ad of Test II. Solutio: 156 patiets out of total 223 patiets

More information

Chapters 5 and 13: REGRESSION AND CORRELATION. Univariate data: x, Bivariate data (x,y).

Chapters 5 and 13: REGRESSION AND CORRELATION. Univariate data: x, Bivariate data (x,y). Chapters 5 ad 13: REGREION AND CORRELATION (ectios 5.5 ad 13.5 are omitted) Uivariate data: x, Bivariate data (x,y). Example: x: umber of years studets studied paish y: score o a proficiecy test For each

More information

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4 MATH 30: Probability ad Statistics 9. Estimatio ad Testig of Parameters Estimatio ad Testig of Parameters We have bee dealig situatios i which we have full kowledge of the distributio of a radom variable.

More information

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics Explorig Data: Distributios Look for overall patter (shape, ceter, spread) ad deviatios (outliers). Mea (use a calculator): x = x 1 + x 2 + +

More information

Section 14. Simple linear regression.

Section 14. Simple linear regression. Sectio 14 Simple liear regressio. Let us look at the cigarette dataset from [1] (available to dowload from joural s website) ad []. The cigarette dataset cotais measuremets of tar, icotie, weight ad carbo

More information

Efficient GMM LECTURE 12 GMM II

Efficient GMM LECTURE 12 GMM II DECEMBER 1 010 LECTURE 1 II Efficiet The estimator depeds o the choice of the weight matrix A. The efficiet estimator is the oe that has the smallest asymptotic variace amog all estimators defied by differet

More information

MBACATÓLICA. Quantitative Methods. Faculdade de Ciências Económicas e Empresariais UNIVERSIDADE CATÓLICA PORTUGUESA 9. SAMPLING DISTRIBUTIONS

MBACATÓLICA. Quantitative Methods. Faculdade de Ciências Económicas e Empresariais UNIVERSIDADE CATÓLICA PORTUGUESA 9. SAMPLING DISTRIBUTIONS MBACATÓLICA Quatitative Methods Miguel Gouveia Mauel Leite Moteiro Faculdade de Ciêcias Ecoómicas e Empresariais UNIVERSIDADE CATÓLICA PORTUGUESA 9. SAMPLING DISTRIBUTIONS MBACatólica 006/07 Métodos Quatitativos

More information

SIMPLE LINEAR REGRESSION AND CORRELATION ANALYSIS

SIMPLE LINEAR REGRESSION AND CORRELATION ANALYSIS SIMPLE LINEAR REGRESSION AND CORRELATION ANALSIS INTRODUCTION There are lot of statistical ivestigatio to kow whether there is a relatioship amog variables Two aalyses: (1) regressio aalysis; () correlatio

More information

Lecture 2: Monte Carlo Simulation

Lecture 2: Monte Carlo Simulation STAT/Q SCI 43: Itroductio to Resamplig ethods Sprig 27 Istructor: Ye-Chi Che Lecture 2: ote Carlo Simulatio 2 ote Carlo Itegratio Assume we wat to evaluate the followig itegratio: e x3 dx What ca we do?

More information

t distribution [34] : used to test a mean against an hypothesized value (H 0 : µ = µ 0 ) or the difference

t distribution [34] : used to test a mean against an hypothesized value (H 0 : µ = µ 0 ) or the difference EXST30 Backgroud material Page From the textbook The Statistical Sleuth Mea [0]: I your text the word mea deotes a populatio mea (µ) while the work average deotes a sample average ( ). Variace [0]: The

More information

Circle the single best answer for each multiple choice question. Your choice should be made clearly.

Circle the single best answer for each multiple choice question. Your choice should be made clearly. TEST #1 STA 4853 March 6, 2017 Name: Please read the followig directios. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directios This exam is closed book ad closed otes. There are 32 multiple choice questios.

More information

STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. Comments:

STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. Comments: Recall: STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS Commets:. So far we have estimates of the parameters! 0 ad!, but have o idea how good these estimates are. Assumptio: E(Y x)! 0 +! x (liear coditioal

More information

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen)

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen) Goodess-of-Fit Tests ad Categorical Data Aalysis (Devore Chapter Fourtee) MATH-252-01: Probability ad Statistics II Sprig 2019 Cotets 1 Chi-Squared Tests with Kow Probabilities 1 1.1 Chi-Squared Testig................

More information

Expectation and Variance of a random variable

Expectation and Variance of a random variable Chapter 11 Expectatio ad Variace of a radom variable The aim of this lecture is to defie ad itroduce mathematical Expectatio ad variace of a fuctio of discrete & cotiuous radom variables ad the distributio

More information

Simple Regression. Acknowledgement. These slides are based on presentations created and copyrighted by Prof. Daniel Menasce (GMU) CS 700

Simple Regression. Acknowledgement. These slides are based on presentations created and copyrighted by Prof. Daniel Menasce (GMU) CS 700 Simple Regressio CS 7 Ackowledgemet These slides are based o presetatios created ad copyrighted by Prof. Daiel Measce (GMU) Basics Purpose of regressio aalysis: predict the value of a depedet or respose

More information

Continuous Data that can take on any real number (time/length) based on sample data. Categorical data can only be named or categorised

Continuous Data that can take on any real number (time/length) based on sample data. Categorical data can only be named or categorised Questio 1. (Topics 1-3) A populatio cosists of all the members of a group about which you wat to draw a coclusio (Greek letters (μ, σ, Ν) are used) A sample is the portio of the populatio selected for

More information

CHAPTER 8 FUNDAMENTAL SAMPLING DISTRIBUTIONS AND DATA DESCRIPTIONS. 8.1 Random Sampling. 8.2 Some Important Statistics

CHAPTER 8 FUNDAMENTAL SAMPLING DISTRIBUTIONS AND DATA DESCRIPTIONS. 8.1 Random Sampling. 8.2 Some Important Statistics CHAPTER 8 FUNDAMENTAL SAMPLING DISTRIBUTIONS AND DATA DESCRIPTIONS 8.1 Radom Samplig The basic idea of the statistical iferece is that we are allowed to draw ifereces or coclusios about a populatio based

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science APRIL/MAY 2009 EXAMINATIONS ECO220Y1Y PART 1 OF 2 SOLUTIONS

UNIVERSITY OF TORONTO Faculty of Arts and Science APRIL/MAY 2009 EXAMINATIONS ECO220Y1Y PART 1 OF 2 SOLUTIONS PART of UNIVERSITY OF TORONTO Faculty of Arts ad Sciece APRIL/MAY 009 EAMINATIONS ECO0YY PART OF () The sample media is greater tha the sample mea whe there is. (B) () A radom variable is ormally distributed

More information

Chapter 1 Simple Linear Regression (part 6: matrix version)

Chapter 1 Simple Linear Regression (part 6: matrix version) Chapter Simple Liear Regressio (part 6: matrix versio) Overview Simple liear regressio model: respose variable Y, a sigle idepedet variable X Y β 0 + β X + ε Multiple liear regressio model: respose Y,

More information

Final Examination Solutions 17/6/2010

Final Examination Solutions 17/6/2010 The Islamic Uiversity of Gaza Faculty of Commerce epartmet of Ecoomics ad Political Scieces A Itroductio to Statistics Course (ECOE 30) Sprig Semester 009-00 Fial Eamiatio Solutios 7/6/00 Name: I: Istructor:

More information

Asymptotic Results for the Linear Regression Model

Asymptotic Results for the Linear Regression Model Asymptotic Results for the Liear Regressio Model C. Fli November 29, 2000 1. Asymptotic Results uder Classical Assumptios The followig results apply to the liear regressio model y = Xβ + ε, where X is

More information

Statistical and Mathematical Methods DS-GA 1002 December 8, Sample Final Problems Solutions

Statistical and Mathematical Methods DS-GA 1002 December 8, Sample Final Problems Solutions Statistical ad Mathematical Methods DS-GA 00 December 8, 05. Short questios Sample Fial Problems Solutios a. Ax b has a solutio if b is i the rage of A. The dimesio of the rage of A is because A has liearly-idepedet

More information

[ ] ( ) ( ) [ ] ( ) 1 [ ] [ ] Sums of Random Variables Y = a 1 X 1 + a 2 X 2 + +a n X n The expected value of Y is:

[ ] ( ) ( ) [ ] ( ) 1 [ ] [ ] Sums of Random Variables Y = a 1 X 1 + a 2 X 2 + +a n X n The expected value of Y is: PROBABILITY FUNCTIONS A radom variable X has a probabilit associated with each of its possible values. The probabilit is termed a discrete probabilit if X ca assume ol discrete values, or X = x, x, x 3,,

More information

ST 305: Exam 3 ( ) = P(A)P(B A) ( ) = P(A) + P(B) ( ) = 1 P( A) ( ) = P(A) P(B) σ X 2 = σ a+bx. σ ˆp. σ X +Y. σ X Y. σ X. σ Y. σ n.

ST 305: Exam 3 ( ) = P(A)P(B A) ( ) = P(A) + P(B) ( ) = 1 P( A) ( ) = P(A) P(B) σ X 2 = σ a+bx. σ ˆp. σ X +Y. σ X Y. σ X. σ Y. σ n. ST 305: Exam 3 By hadig i this completed exam, I state that I have either give or received assistace from aother perso durig the exam period. I have used o resources other tha the exam itself ad the basic

More information

Correlation. Two variables: Which test? Relationship Between Two Numerical Variables. Two variables: Which test? Contingency table Grouped bar graph

Correlation. Two variables: Which test? Relationship Between Two Numerical Variables. Two variables: Which test? Contingency table Grouped bar graph Correlatio Y Two variables: Which test? X Explaatory variable Respose variable Categorical Numerical Categorical Cotigecy table Cotigecy Logistic Grouped bar graph aalysis regressio Mosaic plot Numerical

More information

Direction: This test is worth 150 points. You are required to complete this test within 55 minutes.

Direction: This test is worth 150 points. You are required to complete this test within 55 minutes. Term Test 3 (Part A) November 1, 004 Name Math 6 Studet Number Directio: This test is worth 10 poits. You are required to complete this test withi miutes. I order to receive full credit, aswer each problem

More information

First, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So,

First, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So, 0 2. OLS Part II The OLS residuals are orthogoal to the regressors. If the model icludes a itercept, the orthogoality of the residuals ad regressors gives rise to three results, which have limited practical

More information

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator Ecoomics 24B Relatio to Method of Momets ad Maximum Likelihood OLSE as a Maximum Likelihood Estimator Uder Assumptio 5 we have speci ed the distributio of the error, so we ca estimate the model parameters

More information

Sample Size Determination (Two or More Samples)

Sample Size Determination (Two or More Samples) Sample Sie Determiatio (Two or More Samples) STATGRAPHICS Rev. 963 Summary... Data Iput... Aalysis Summary... 5 Power Curve... 5 Calculatios... 6 Summary This procedure determies a suitable sample sie

More information

DS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10

DS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10 DS 00: Priciples ad Techiques of Data Sciece Date: April 3, 208 Name: Hypothesis Testig Discussio #0. Defie these terms below as they relate to hypothesis testig. a) Data Geeratio Model: Solutio: A set

More information

Statistics 511 Additional Materials

Statistics 511 Additional Materials Cofidece Itervals o mu Statistics 511 Additioal Materials This topic officially moves us from probability to statistics. We begi to discuss makig ifereces about the populatio. Oe way to differetiate probability

More information

Big Picture. 5. Data, Estimates, and Models: quantifying the accuracy of estimates.

Big Picture. 5. Data, Estimates, and Models: quantifying the accuracy of estimates. 5. Data, Estimates, ad Models: quatifyig the accuracy of estimates. 5. Estimatig a Normal Mea 5.2 The Distributio of the Normal Sample Mea 5.3 Normal data, cofidece iterval for, kow 5.4 Normal data, cofidece

More information

A statistical method to determine sample size to estimate characteristic value of soil parameters

A statistical method to determine sample size to estimate characteristic value of soil parameters A statistical method to determie sample size to estimate characteristic value of soil parameters Y. Hojo, B. Setiawa 2 ad M. Suzuki 3 Abstract Sample size is a importat factor to be cosidered i determiig

More information

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting Lecture 6 Chi Square Distributio (χ ) ad Least Squares Fittig Chi Square Distributio (χ ) Suppose: We have a set of measuremets {x 1, x, x }. We kow the true value of each x i (x t1, x t, x t ). We would

More information

Output Analysis (2, Chapters 10 &11 Law)

Output Analysis (2, Chapters 10 &11 Law) B. Maddah ENMG 6 Simulatio Output Aalysis (, Chapters 10 &11 Law) Comparig alterative system cofiguratio Sice the output of a simulatio is radom, the comparig differet systems via simulatio should be doe

More information

Linear Regression Models, OLS, Assumptions and Properties

Linear Regression Models, OLS, Assumptions and Properties Chapter 2 Liear Regressio Models, OLS, Assumptios ad Properties 2.1 The Liear Regressio Model The liear regressio model is the sigle most useful tool i the ecoometricia s kit. The multiple regressio model

More information

Understanding Samples

Understanding Samples 1 Will Moroe CS 109 Samplig ad Bootstrappig Lecture Notes #17 August 2, 2017 Based o a hadout by Chris Piech I this chapter we are goig to talk about statistics calculated o samples from a populatio. We

More information

Topic 10: Introduction to Estimation

Topic 10: Introduction to Estimation Topic 0: Itroductio to Estimatio Jue, 0 Itroductio I the simplest possible terms, the goal of estimatio theory is to aswer the questio: What is that umber? What is the legth, the reactio rate, the fractio

More information

II. Descriptive Statistics D. Linear Correlation and Regression. 1. Linear Correlation

II. Descriptive Statistics D. Linear Correlation and Regression. 1. Linear Correlation II. Descriptive Statistics D. Liear Correlatio ad Regressio I this sectio Liear Correlatio Cause ad Effect Liear Regressio 1. Liear Correlatio Quatifyig Liear Correlatio The Pearso product-momet correlatio

More information

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors ECONOMETRIC THEORY MODULE XIII Lecture - 34 Asymptotic Theory ad Stochastic Regressors Dr. Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Asymptotic theory The asymptotic

More information

Chapter 13, Part A Analysis of Variance and Experimental Design

Chapter 13, Part A Analysis of Variance and Experimental Design Slides Prepared by JOHN S. LOUCKS St. Edward s Uiversity Slide 1 Chapter 13, Part A Aalysis of Variace ad Eperimetal Desig Itroductio to Aalysis of Variace Aalysis of Variace: Testig for the Equality of

More information

In this section we derive some finite-sample properties of the OLS estimator. b is an estimator of β. It is a function of the random sample data.

In this section we derive some finite-sample properties of the OLS estimator. b is an estimator of β. It is a function of the random sample data. 17 3. OLS Part III I this sectio we derive some fiite-sample properties of the OLS estimator. 3.1 The Samplig Distributio of the OLS Estimator y = Xβ + ε ; ε ~ N[0, σ 2 I ] b = (X X) 1 X y = f(y) ε is

More information

Module 1 Fundamentals in statistics

Module 1 Fundamentals in statistics Normal Distributio Repeated observatios that differ because of experimetal error ofte vary about some cetral value i a roughly symmetrical distributio i which small deviatios occur much more frequetly

More information

MA Advanced Econometrics: Properties of Least Squares Estimators

MA Advanced Econometrics: Properties of Least Squares Estimators MA Advaced Ecoometrics: Properties of Least Squares Estimators Karl Whela School of Ecoomics, UCD February 5, 20 Karl Whela UCD Least Squares Estimators February 5, 20 / 5 Part I Least Squares: Some Fiite-Sample

More information

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting Lecture 6 Chi Square Distributio (χ ) ad Least Squares Fittig Chi Square Distributio (χ ) Suppose: We have a set of measuremets {x 1, x, x }. We kow the true value of each x i (x t1, x t, x t ). We would

More information

The Method of Least Squares. To understand least squares fitting of data.

The Method of Least Squares. To understand least squares fitting of data. The Method of Least Squares KEY WORDS Curve fittig, least square GOAL To uderstad least squares fittig of data To uderstad the least squares solutio of icosistet systems of liear equatios 1 Motivatio Curve

More information

Data Analysis and Statistical Methods Statistics 651

Data Analysis and Statistical Methods Statistics 651 Data Aalysis ad Statistical Methods Statistics 651 http://www.stat.tamu.edu/~suhasii/teachig.html Suhasii Subba Rao Review of testig: Example The admistrator of a ursig home wats to do a time ad motio

More information

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

Discrete Mathematics for CS Spring 2008 David Wagner Note 22 CS 70 Discrete Mathematics for CS Sprig 2008 David Wager Note 22 I.I.D. Radom Variables Estimatig the bias of a coi Questio: We wat to estimate the proportio p of Democrats i the US populatio, by takig

More information

Probability and statistics: basic terms

Probability and statistics: basic terms Probability ad statistics: basic terms M. Veeraraghava August 203 A radom variable is a rule that assigs a umerical value to each possible outcome of a experimet. Outcomes of a experimet form the sample

More information

Final Review. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech

Final Review. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech Fial Review Fall 2013 Prof. Yao Xie, yao.xie@isye.gatech.edu H. Milto Stewart School of Idustrial Systems & Egieerig Georgia Tech 1 Radom samplig model radom samples populatio radom samples: x 1,..., x

More information

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015 ECE 8527: Itroductio to Machie Learig ad Patter Recogitio Midterm # 1 Vaishali Ami Fall, 2015 tue39624@temple.edu Problem No. 1: Cosider a two-class discrete distributio problem: ω 1 :{[0,0], [2,0], [2,2],

More information

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes.

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes. Term Test October 3, 003 Name Math 56 Studet Number Directio: This test is worth 50 poits. You are required to complete this test withi 50 miutes. I order to receive full credit, aswer each problem completely

More information

Lecture 7: Properties of Random Samples

Lecture 7: Properties of Random Samples Lecture 7: Properties of Radom Samples 1 Cotiued From Last Class Theorem 1.1. Let X 1, X,...X be a radom sample from a populatio with mea µ ad variace σ

More information

Linear Regression Demystified

Linear Regression Demystified Liear Regressio Demystified Liear regressio is a importat subject i statistics. I elemetary statistics courses, formulae related to liear regressio are ofte stated without derivatio. This ote iteds to

More information

Lesson 11: Simple Linear Regression

Lesson 11: Simple Linear Regression Lesso 11: Simple Liear Regressio Ka-fu WONG December 2, 2004 I previous lessos, we have covered maily about the estimatio of populatio mea (or expected value) ad its iferece. Sometimes we are iterested

More information

Binomial Distribution

Binomial Distribution 0.0 0.5 1.0 1.5 2.0 2.5 3.0 0 1 2 3 4 5 6 7 0.0 0.5 1.0 1.5 2.0 2.5 3.0 Overview Example: coi tossed three times Defiitio Formula Recall that a r.v. is discrete if there are either a fiite umber of possible

More information

1 Models for Matched Pairs

1 Models for Matched Pairs 1 Models for Matched Pairs Matched pairs occur whe we aalyse samples such that for each measuremet i oe of the samples there is a measuremet i the other sample that directly relates to the measuremet i

More information

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d Liear regressio Daiel Hsu (COMS 477) Maximum likelihood estimatio Oe of the simplest liear regressio models is the followig: (X, Y ),..., (X, Y ), (X, Y ) are iid radom pairs takig values i R d R, ad Y

More information

TMA4245 Statistics. Corrected 30 May and 4 June Norwegian University of Science and Technology Department of Mathematical Sciences.

TMA4245 Statistics. Corrected 30 May and 4 June Norwegian University of Science and Technology Department of Mathematical Sciences. Norwegia Uiversity of Sciece ad Techology Departmet of Mathematical Scieces Corrected 3 May ad 4 Jue Solutios TMA445 Statistics Saturday 6 May 9: 3: Problem Sow desity a The probability is.9.5 6x x dx

More information

Inferential Statistics. Inference Process. Inferential Statistics and Probability a Holistic Approach. Inference Process.

Inferential Statistics. Inference Process. Inferential Statistics and Probability a Holistic Approach. Inference Process. Iferetial Statistics ad Probability a Holistic Approach Iferece Process Chapter 8 Poit Estimatio ad Cofidece Itervals This Course Material by Maurice Geraghty is licesed uder a Creative Commos Attributio-ShareAlike

More information

If, for instance, we were required to test whether the population mean μ could be equal to a certain value μ

If, for instance, we were required to test whether the population mean μ could be equal to a certain value μ STATISTICAL INFERENCE INTRODUCTION Statistical iferece is that brach of Statistics i which oe typically makes a statemet about a populatio based upo the results of a sample. I oesample testig, we essetially

More information

Probability and Statistics

Probability and Statistics ICME Refresher Course: robability ad Statistics Staford Uiversity robability ad Statistics Luyag Che September 20, 2016 1 Basic robability Theory 11 robability Spaces A probability space is a triple (Ω,

More information

Statistics 203 Introduction to Regression and Analysis of Variance Assignment #1 Solutions January 20, 2005

Statistics 203 Introduction to Regression and Analysis of Variance Assignment #1 Solutions January 20, 2005 Statistics 203 Itroductio to Regressio ad Aalysis of Variace Assigmet #1 Solutios Jauary 20, 2005 Q. 1) (MP 2.7) (a) Let x deote the hydrocarbo percetage, ad let y deote the oxyge purity. The simple liear

More information

Overview. p 2. Chapter 9. Pooled Estimate of. q = 1 p. Notation for Two Proportions. Inferences about Two Proportions. Assumptions

Overview. p 2. Chapter 9. Pooled Estimate of. q = 1 p. Notation for Two Proportions. Inferences about Two Proportions. Assumptions Chapter 9 Slide Ifereces from Two Samples 9- Overview 9- Ifereces about Two Proportios 9- Ifereces about Two Meas: Idepedet Samples 9-4 Ifereces about Matched Pairs 9-5 Comparig Variatio i Two Samples

More information

Estimation for Complete Data

Estimation for Complete Data Estimatio for Complete Data complete data: there is o loss of iformatio durig study. complete idividual complete data= grouped data A complete idividual data is the oe i which the complete iformatio of

More information

University of California, Los Angeles Department of Statistics. Simple regression analysis

University of California, Los Angeles Department of Statistics. Simple regression analysis Uiversity of Califoria, Los Ageles Departmet of Statistics Statistics 100C Istructor: Nicolas Christou Simple regressio aalysis Itroductio: Regressio aalysis is a statistical method aimig at discoverig

More information

Open book and notes. 120 minutes. Cover page and six pages of exam. No calculators.

Open book and notes. 120 minutes. Cover page and six pages of exam. No calculators. IE 330 Seat # Ope book ad otes 120 miutes Cover page ad six pages of exam No calculators Score Fial Exam (example) Schmeiser Ope book ad otes No calculator 120 miutes 1 True or false (for each, 2 poits

More information

Homework for 4/9 Due 4/16

Homework for 4/9 Due 4/16 Name: ID: Homework for 4/9 Due 4/16 1. [ 13-6] It is covetioal wisdom i military squadros that pilots ted to father more girls tha boys. Syder 1961 gathered data for military fighter pilots. The sex of

More information