Lecture 1, Jan 19. i=1 p i = 1.
|
|
- Sherman Wheeler
- 5 years ago
- Views:
Transcription
1 Lecture 1, Ja 19 Review of the expected value, covariace, correlatio coefficiet, mea, ad variace. Radom variable. A variable that takes o alterative values accordig to chace. More specifically, a radom variable assumes differet values, each with probability less tha or equal to 1. A cotiuous radom variable may take o ay value o the real umber lie. A discrete radom variable may take oly a specific umber of values. Probability mass fuctio (pmf) ad probability desity fuctio (pdf). The process that geerates the values of a radom variable. It lists all possible outcomes ad the probability that each will occur. Expected values. The mea, or the expected value, of a radom variable X is a weighted average of the possible outcomes, where the probabilities of the outcomes serve as the weights. Expectatio operator deoted E, mea of X deoted µ X. I the discrete case, N µ X = E(X) = p 1 X 1 + p 2 X p N X N = p i X i, where p i = P (X = X i ) ad N p i = 1. I the cotiuous case, µ X = E(X) = Ω f(x)xdx, where Ω is the support of X, ad f(x) is the pdf of X. The sample mea of a set of outcomes o X is deoted by X. X = 1 x i, where x 1,..., x are realizatios of X. 1
2 Variace. The variace of a radom variable provides a measure of the spread, or dispersio, aroud the mea. Var(X) = σx 2 = E[X E(X)] 2. I the discrete case, N Var(X) = σx 2 = p i [X i E(X)] 2. The positive square root of the variace is called the stadard deviatio. Joit distributio. I the discrete case, joit distributios of X ad Y are described by a list of probabilities of occurrece of all possible outcomes o both X ad Y. The covariace of X ad Y is Cov(X, Y ) = E[X E(X)(Y E(Y )]. I the discrete case, N M Cov(X, Y ) = p ij [X i E(X)][Y j E(Y )], j=1 where p ij represets the joit probability of X = X i ad Y = Y j occurrig. The covariace is a measure of the liear associatio betwee X ad Y. If both variables are always above ad below their meas at the same time, the covariace will be positive. If X is above its mea whe Y is below its mea ad vice versa, the covariace will be egative. The value of covariace depeds upo the uits i which X ad Y are measured. The correlatio coefficiet ρ XY is a measure of the associatio which has bee ormalized ad is scale-free. Cov(X, Y) ρ XY =, σ X σ Y where σ X ad σ Y represet the stadard deviatio of X ad Y respectively. The correlatio coefficiet is always betwee -1 ad 1. A positive correlatio idicates that the variables move i the same directio, while a egative correlatio implies that they move i opposite directios. Properties. Suppose X ad Y are radom variables, a ad b are costats: 1. E(aX + by ) = ae(x) + be(y ); 2. Var(aX + b) = a 2 Var(X); 3. Var(X + Y ) = Var(X) + Var(Y ) + 2Cov(X, Y ); 4. If X ad Y are idepedet, the Cov(X, Y ) = 0. 2
3 Estimatio. Meas, variaces ad covariaces ca be measured with certaity oly if we kow all there is to kow about all possible outcomes. I practice, we may obtai a sample of the relevat iformatio eeded. Give x 1,..., x are radom observatios of X, we wat to estimate a populatio parameter (like the mea or the variace). The sample mea X is a ubiased estimator of the populatio mea µ X. E( X) = E( 1 x i ) = 1 E(x i ) = 1 µ X = µ X. The sample variace s 2 X = 1 (x 1 i X) 2 is a ubiased estimator of the populatio variace Var(X). The sample covariace s XY = 1 (x 1 i X)(y i Ȳ ) is a ubiased estimator of the populatio covariace Cov(X, Y ). The sample correlatio coefficiet is defied as r XY = (x i X)(y i Ȳ ) (x i X) 2 (y i Ȳ )2. Desired properties of estimators: 1. Lack of bias. The bias associated with a estimated parameter is defied to be: Bias( ˆβ) = E( ˆβ) β. ˆβ is a ubiased estimator if the mea or the expected value of ˆβ is equal to the true value, that is, E( ˆβ) = β. 2. Cosistecy. ˆβ is a cosistet estimator of β if for ay δ > 0, lim P ( β ˆβ < δ) = 1. As sample size approaches ifiity, the probability that ˆβ will differ from β will get very small. 3. Efficiecy. We say that ˆβ is a efficiet ubiased estimator if for a give sample size, the variace of ˆβ is smaller tha the variace of ay other ubiased estimators. 3
4 Lecture 2, Ja 26 Tradeoff betwee bias ad variace of estimators. Whe the goal is to maximize the precisio of the predictios, a estimator with low variace ad some bias may be more desirable tha a ubiased estimator with high variace. We may wat to miimize the mea square error, defied as Mea square error( ˆβ) = E( ˆβ β) 2. It ca be show that Mea square error( ˆβ) = [Bias( ˆβ)] 2 +Var( ˆβ). The criterio of miimizig mea square error take ito accout of both the variace ad the bias of the estimator. A alterative criterio to cosistecy is that the mea square error of the estimator approaches zero as the sample size icreases. This implies asymptotically, or whe the sample size is very large, the estimator is ubiased ad its variace goes to zero. A estimator with a mea square error that approaches zero will be cosistet estimator but that the reverse eed ot be true. 1
5 Probability distributios. The Normal distributio is a cotiuous bell-shaped probability distributio. It ca be fully described by its mea ad its variace. If X is ormally distributed, we write X N(µ X, σ 2 X), which is read X is ormally distributed with mea µ X ad variace σ 2 X. The probability that a sigle observatio of a ormally distributed variable will lie withi 1.96 stadard deviatios of its mea is approximately The probability that a sigle observatio of a ormally distributed variable will lie withi 2.57 stadard deviatios of its mea is approximately P (µ X 1.96σ X < X < µ X σ X ) 0.95, P (µ X 2.57σ X < X < µ X σ X ) Compariso of ormal distributios Desity Distributios Variace=1 Variace=4 Variace= x value The weighted sum of ormal radom variables is still ormal. 2
6 The chi square distributio with N degrees of freedom is the sum of the squares of N idepedetly distributed stadard ormal radom variables (with mea 0 ad variace 1). The chi square distributio starts at the origi, is skewed to the right, ad has a tail which exteds ifiitely to the right. The distributio becomes more ad more symmetric as the umber of degrees of freedom gets larger. It becomes close to ormal distributio whe degrees of freedom is very large. Compariso of chi square distributios Desity Distributios df=3 df=8 df= x value As a example, whe we calculate the sample variaces s 2 of observatios from a ormal distributio with variaces σ 2, ( 1)s 2 /σ 2 is chi square with 1 degrees of freedom. 3
7 The t distributio. Assume X is ormal with mea 0 ad variace 1, Z is chi square with N degrees of freedom, X ad Z are idepedet. The X/ Z/N has a t distributio with N degrees of freedom. The t distributio is symmetric, has fatter tails tha the ormal distributio, ad approximates the ormal distributio whe the degrees of freedom is large. Compariso of t distributios Desity Distributios df=1 df=3 df=8 df=30 ormal x value As a example, X N(µ X, σx), 2 ad X = 1 x i, is the sample mea based o a sample x 1,..., x of size. The ( X µ X )/(σ X / ) is ormal with mea 0 ad variace 1. For ukow σx, 2 we replace σx 2 by sample variace s 2 X, the ( X µx ) ( = X µ X )/(σ X / ) s X [( 1)s 2 X/σX]/( 2 1), which is a t distributio with 1 degrees of freedom. 4
8 The F distributio. If X ad Z are idepedet ad distributed as chi square with N 1 ad N 2 degrees of freedom, respectively, the (X/N 1 )/(Z/N 2 ) is distributed as a F distributio with N 1 ad N 2 degrees of freedom. These two parameters are called the umerator ad the deomiator degrees of freedom. F Distributios Desity Distributios F(33,10) F(5,10) F(2,7) x value 5
9 Forecastig: to calculate or predict some future evet or coditio, usually as a result of ratioal study or aalysis of pertiet data. We all make forecasts: a perso waitig for a bus; parets expectig a telephoe call from their childre; bak maager predicts cash flow for the ext quarter; compay maager predicts sales or estimates umber of ma-hours required to meet a give productio schedule. Future evets ivolve ucertaity. Forecasts are ot perfect. The objective of forecastig is to reduce the forecast error. Each forecast has its ow specificatios, ad solutios to oe are ot solutios i aother situatio. Geeral priciples for ay forecast system: Model specificatio; Model estimatio; Diagostic checkig; Forecast geeratio; Stability checkig; Forecast updatig. Choice of forecast model. (1) Degrees of accuracy required; (2) forecast horizo; (3) budget; (4) what data is available. Oe may ot costruct accurate empirical forecast models from limited ad icomplete data base. Forecast criteria. Actual observatio at time t is z t. Its forecast, which uses the iformatio up to ad icludig time t 1, is z t 1 (1). Objective is to make the future forecast error z t z t 1 (1) as small as possible. However, z t is ukow, we ca oly talk about its expected value, coditioal o the observed data up to ad icludig time t 1. We miimize the mea absolute error E z t z t 1 (1) or the mea square error E[z t z t 1 (1)] 2. We use the mea square error criterio for simpler mathematical calculatios. What we will discuss. I sigle-variable forecastig, we use past history of the series, z t, where t is the time idex, to extrapolate ito the future. I regressio forecastig, we use the relatioships betwee the variable to be forecast ad the other variables. I the most geeral form, the regressio model ca be writte as y t = f(x t1,..., x tp ; β 1,..., β m ) + ε t Describes relatioship betwee oe depedet variable Y ad p idepedet variables X 1,..., X p. Idex t meas at time t or for subject t. At idex t, we observe y t ad x t1,..., x tp. Ukow parameters β 1,..., β m. Kow mathematical fuctio form f. Radomess for this model rises from the error term ε t. 6
10 Lecture 3, Feb 2 I the most geeral form, the regressio model ca be writte as y t = f(x t1,..., x tp ; β 1,..., β m ) + ε t Describes relatioship betwee oe depedet variable Y ad p idepedet variables X 1,..., X p. Idex t meas at time t or for subject t. At idex t, we observe y t ad x t1,..., x tp. Ukow parameters β 1,..., β m. Kow mathematical fuctio form f. Radomess for this model rises from the error term ε t. Models liear i the parameter: 1. y = β 0 + β 1 x 1 + ε; 2. y = β 0 + β 1 x 1 + β 2 x 2 + ε; 3. y = β 0 + β 1 x 1 + β 2 x ε; Models liear i the idepedet variable: 1. y = β 0 + β 1 x 1 + ε; 2. y = β 0 + β 1 x 1 + β 2 x 2 + ε. Regressio through origi. y: salary of a sales perso; x: umber of products sold. No base salary! y: merit I get for icrease i salary; x: umber of papers published. Model: y = βx + ε. 1
11 Give observatios (x 1, y 1 ),..., (x, y ), we wat to miimize S(β) = (y i βx i ) 2. Take derivatives ad set to 0, the we have S (β) = ds(β) dβ = 2 (y i βx i )x i = 0 The least squares estimator of β is ˆβ = x iy i. x2 i Take secod order derivative to make sure this is a miimizer. Check S (β) = d2 S(β) dβ 2 = ds (β) dβ = 2 x 2 i 0. Simple liear regressio. y: resale price of a preowed car; x: milage! y: ruig time for 10-km road race; x: maximal aerobic capacity (oxyge uptake, milliliter per kilogram per miite, ml/(kg mi) Show MINITAB example! Scatter plot! Model: y = β 0 + β 1 x + ε. Give observatios (x 1, y 1 ),..., (x, y ), we wat to miimize S(β 0, β 1 ) = (y i β 0 β 1 x i ) 2. Take derivatives ad set to 0, the we have β 0 S(β 0, β 1 ) = 2 (y i β 0 β 1 x i ) = 0 2
12 β 1 S(β 0, β 1 ) = 2 (y i β 0 β 1 x i )x i = 0 We have two equatios ad two ukows, solve for β 0 ad β 1. The we have ˆβ 1 = s xy = (x i x)(y i ȳ) s 2 x (x, i x) 2 ˆβ0 = ȳ ˆβ 1 x, where ( s xy = 1 x i y i x ) i y i = (x i x)(y i ȳ) 1 1 [ ] s 2 x = 1 x 2 i ( x i) 2 = (x i x) We still eed to take secod order derivatives to make sure ( ˆβ 0, ˆβ 1 ) are actual miimizers. Model assumptios: 1. The relatioship betwee x ad y is liear, y i = β 0 + β 1 x i + ε i ; 2. The x are o-stochastic variable whose values x 1,..., x are fixed; 3. Normality The error term ε is ormally distributed; Homoscedasticity The error term ε has mea zero ad costat variace for all observatios, E(ε i ) = 0, Var(ε i ) = σ 2 ; Idepedecy The radom error ε i ad ε j are idepedet of each other for i j, or errors correspodig to differet observatios are idepedet. How do we check those assumptios? DO AN EXAMPLE BY HAND. Show MINITAB RESULTS! ˆβ 1, ˆβ 0 as weighted sum x Let w i = i x, the ˆβ (x i x) 2 1 = w iy i. We otice that w i = 0, ad w i x i = w i (x i x) = 1. 3
13 Thus, E( ˆβ 1 ) = E( w i y i ) = w i E(y i ) = = β 0 w i + β 1 w i x i = β 1, w i (β 0 + β 1 x i ) which meas ˆβ 1 is ubiased estimator of β 1. Furthermore wi 2 = (x i x) 2 [ (x i x) 2 ] 2 = (x i x) 2 [ (x i x) 2 ] = 1 2 (x i x), 2 ad Var( ˆβ 1 ) = Var( w i y i ) = wi 2 Var(y i ) = σ 2 wi 2 = σ 2 (x i x) 2. Similarly, ˆβ 0 = ȳ ˆβ 1 x = 1 E( ˆβ 0 ) = y i x w iy i = ( 1 xw i)y i. Thus ( 1 xw i)e(y i ) = ( 1 xw i)(β 0 + β 1 x i ) = β 0 xβ 0 w i + β 1 x β 1 x w i x i = β 0. which meas ˆβ 0 is ubiased estimator of β 0. Furthermore Var( ˆβ 0 ) = Var[ ( 1 xw i)y i ] = ( 1 xw i) 2 Var(y i ) = σ 2 ( 1 xw i) 2 = σ 2 ( x2 wi 2 2 xw i) = σ 2 ( 1 + x 2 (x i x) ) = 2 We ca also show that Cov( ˆβ 0, ˆβ xσ 2 1 ) = (x i x) 2 σ2 x2 i (x i x) 2 4
14 Gauss-Markov Theorem: If the previous model assumptios are satisfied, the amog all the liear ubiased estimators of β 0 ad β 1, the least squares estimators ˆβ 0 ad ˆβ 1 have the smallest variace. The Theorem implies that for ay ubiased estimator of β 1 with form wi y i, its variace should be Var( ˆβ σ 1 ) = 2. (x i x) 2 The fitted least squares lie: ŷ = ˆβ 0 + ˆβ 1 x is used to get the fitted value y. The i-th residual is the differece betwee the i-th observatio of depedet variable ad its fitted value: r i = y i ŷ i. I practice, Var(ε) = σ 2 is ukow ad we eed to estimate it. The sum of squares of error (SSE; also kow as sum of squares of residual) is SSE = (y i ŷ i ) 2 = (y i ˆβ 0 ˆβ 1 x i ) 2. It has 2 degrees of freedom. We have radom observatios, that s degrees of freedom. Estimatig ˆβ 0 ad ˆβ 1 uses 2 degrees of freedom. The mea squares error (MSE) is defied to be MSE = SSE 2 = (y i ˆβ 0 ˆβ 1 x i ) 2. 2 MSE is also deoted by ˆσ 2. It is a ubiased estimator of σ 2. It ca be show that E(MSE) = E(ˆσ 2 ) = σ 2. The i-th stadardized residual sr i = r i /ˆσ. I summarizatio, we have ˆβ 1 N ˆβ 0 N SSE σ 2 ( ) σ 2 β 1, (x. i x) 2 ( σ 2 ) β 0, x2 i (x. i x) 2 = ( 2)ˆσ2 σ 2 χ
15 Lecture 4, Feb 16 Simple liear regressio. Model: y = β 0 + β 1 x + ε. Model assumptios: 1. The relatioship betwee x ad y is liear, y i = β 0 + β 1 x i + ε i ; 2. The x are o-stochastic variable whose values x 1,..., x are fixed; 3. Normality The error term ε is ormally distributed; Homoscedasticity The error term ε has mea zero ad costat variace for all observatios, E(ε i ) = 0, Var(ε i ) = σ 2 ; Idepedecy The radom error ε i ad ε j are idepedet of each other for i j, or errors correspodig to differet observatios are idepedet. Give observatios (x 1, y 1 ),..., (x, y ), we wat to miimize S(β 0, β 1 ) = (y i β 0 β 1 x i ) 2. Take derivatives ad set to 0, the we have S(β 0, β 1 ) = 2 (y i β 0 β 1 x i ) = 0 β 0 β 1 S(β 0, β 1 ) = 2 (y i β 0 β 1 x i )x i = 0 We have two equatios ad two ukows, solve for β 0 ad β 1. The we have ˆβ 1 = s xy = (x i x)(y i ȳ) s 2 x (x, i x) 2 ˆβ0 = ȳ ˆβ 1 x, where 1
16 ( s xy = 1 x i y i x i y i 1 [ s 2 x = 1 1 x 2 i ( x i) 2 ) ] = = (x i x)(y i ȳ) 1 (x i x) 2 1 We still eed to take secod order derivatives to make sure ( ˆβ 0, ˆβ 1 ) are actual miimizers. The fitted least squares lie: ŷ = ˆβ 0 + ˆβ 1 x is used to get the fitted value y. The i-th residual is the differece betwee the i-th observatio of depedet variable ad its fitted value: r i = y i ŷ i. I practice, Var(ε) = σ 2 is ukow ad we eed to estimate it. The sum of squares of error (SSE; also kow as sum of squares of residual) is SSE = (y i ŷ i ) 2 = (y i ˆβ 0 ˆβ 1 x i ) 2. It has 2 degrees of freedom. We have radom observatios, that s degrees of freedom. Estimatig ˆβ 0 ad ˆβ 1 uses 2 degrees of freedom. The mea squares error (MSE) is defied to be MSE = SSE 2 = (y i ˆβ 0 ˆβ 1 x i ) 2. 2 MSE is also deoted by ˆσ 2. It is a ubiased estimator of σ 2. It ca be show that E(MSE) = E(ˆσ 2 ) = σ 2. The i-th stadardized residual sr i = r i /ˆσ. I summarizatio, we have ˆβ 1 N ˆβ 0 N SSE σ 2 ( ) σ 2 β 1, (x. i x) 2 ( σ 2 ) β 0, x2 i (x. i x) 2 = ( 2)ˆσ2 σ 2 χ
17 Cofidece iterval Deote σ 2ˆβ1 = σ 2 (x i x) 2, s 2ˆβ1 = ˆσ 2 (x i x) 2, the ˆβ 1 β 1 σ ˆβ1 N(0, 1). For kow σ 2, the 100(1 α)% cofidece iterval for β 1 is ˆβ 1 ± z α/2 σ ˆβ1. For ukow σ 2, plug i s ˆβ1 as a estimate of ukow σ ˆβ1 ad we have t = ˆβ 1 β 1 s ˆβ1 = ( ˆβ 1 β 1 )/σ ˆβ1 = s ( ˆβ1 β 1 )/σ ˆβ 1 2ˆβ1 /σ 2ˆβ1 ( 2)ˆσ 2 /( 2) σ 2 N(0, 1) χ 2 2 /( 2) = t 2 Deote the upper percetile t 2,α/2, which satisfies P (t > t 2,α/2 ) = α/2 for radom variable t t 2. Thus ( ) P rob t 2,α/2 < ˆβ 1 β 1 s ˆβ1 < t 2,α/2 = 100(1 α)% P rob( ˆβ 1 t 2,α/2 s ˆβ1 < β 1 < ˆβ 1 + t 2,α/2 s ˆβ1 ) = 100(1 α)% The 100(1 α)% cofidece iterval for β 1 is ˆβ 1 ± t 2,α/2 s ˆβ1. Hypothesis test We are iterested whether there is sigificat predictor effect. Null hypothesis: H 0 : β 1 = 0 v.s. H 1 : β 1 0. Test statistics: t = ˆβ 1 s ˆβ1. Uder H 0, t t 2. At 1 α sigificace level, reject H 0 if t > t 2,α/2. We are iterested whether there is sigificat positive predictor effect. Null hypothesis: H 0 : β 1 = 0 v.s. H 0 : β 1 > 0. Test statistics: t = ˆβ 1 s ˆβ1. Uder H 0, t t 2. At 1 α sigificace level, reject H 0 if t > t 2,α. We are iterested whether the slope is at a certai specified level c. Null hypothesis: H 0 : β 1 = c v.s. H 0 : β 1 c. Test statistics: t = ˆβ 1 c s ˆβ1. Uder H 0, t t 2. At 1 α sigificace level, reject H 0 if t > t 2,α/2. 3
18 Lecture 5, Feb 19 Please check with your classmates for missed lecture otes. The materials listed here are the key poits for ANOVA. Make sure you uderstad EV- ERYTHING listed here. We are goig to get back to ANOVA whe we deal with multiple liear regressio. This versio does ot have ay examples or demostratio i MINITAB. TUCAPTURE should be available soo. ANOVA The variatio of y has the followig decompositio (y i ȳ) 2 = (ŷ i ȳ) 2 + (y i ŷ i ) 2 Total Sum of Squares (SST)= Sum of Squares of Regressio (SSR)+ Sum of Squares of Error(SSE) R 2, or the R squared, coefficiet of determiatio, is the proportio of variatio i y explaied by the regressio equatio. R 2 = SSR SST = 1 SSE SST The sample correlatio coefficiet R = ± R 2. R > 0 whe ˆβ 1 > 0; ad R < 0 whe ˆβ 1 < 0. R measures how closely the data fit a straight lie. ANOVA table Source df Sum of Squares Mea Squares F Regressio 1 SSR = (ŷ i ȳ) 2 MSR=SSR/1 F=MSR/MSE Error -2 SSE = (y i ŷ i ) 2 MSE=SSE/(-2) Total -1 SST = (y i ȳ) 2 1
19 To test whether there is sigificat predictor effect, we do t use R 2. We use F, which takes ito cosideratio of the degrees of freedom. Null hypothesis: H 0 : β 1 = 0 v.s. H 1 : β 1 0. If test statistic F is large, which meas a large proportio of variatio i y is explaied by the regressio, the ituitio is that we reject H 0. It ca be show that uder H 0, F = MSR/MSE F 1, 2. Deote the upper percetile F 1, 2,α, which satisfies P (F > F 1, 2,α ) = α for radom variable F F 1, 2. At 1 α sigificace level, reject H 0 if F = MSR/MSE > F 1, 2,α. Notice that F = MSR MSE = SSR SSE/( 2) = ( 2)SSR SST SSR = F = ( 2)R2 1 R 2 ( 2)SSR/SST SST/SST SSR/SST Large R 2 with small may ot be sigificat; but moderate R 2 with large sample size ca be highly sigificat. It ca be show that the t test ad the F test agree with each other. 2
20 Lecture 6, Feb 23 Model assumptios: 1. The relatioship betwee x ad y is liear, y i = β 0 + β 1 x i + ε i ; 2. The x are o-stochastic variable whose values x 1,..., x are fixed; 3. Normality The error term ε is ormally distributed; Homoscedasticity The error term ε has mea zero ad costat variace for all observatios, E(ε i ) = 0, Var(ε i ) = σ 2 ; Idepedecy The radom error ε i ad ε j are idepedet of each other for i j, or errors correspodig to differet observatios are idepedet. Gauss-Markov Theorem: If the previous model assumptios are satisfied, the amog all the liear ubiased estimators of β 0 ad β 1, the least squares estimators ˆβ 0 ad ˆβ 1 have the smallest variace. The Theorem implies that for ay ubiased estimator of β 1 with form wi y i, its variace should be Var( ˆβ σ 1 ) = 2. (x i x) 2 Predictio Iterval Cosider a college tryig to predict first year GPA of a ew studet based o the studet s high school GPA. Based o preset studets, the regressio equatio is ŷ = x, where x =high school GPA ad y = first year college GPA. For a studet with high school GPA x ew = 3.5, we estimate its first year college GPA by ŷ ew = ˆβ 0 + ˆβ 1 x ew = = 3.2. The mea of the forecast error is E(y ew ŷ ew ) = E[(β 0 + β 1 x ew + ε ew ) ( ˆβ 0 + ˆβ 1 x ew )] = 0. 1
21 By the Gauss-Markov theorem, ŷ ew is the miimum mea square error forecast amog all liear ubiased forecasts. The variace of the forecast error is Var(y ew ŷ ew ) = Var[(β 0 + β 1 x ew + ε ew ) ( ˆβ 0 + ˆβ 1 x ew )] = σ 2 + Var( ˆβ 0 ) + x 2 ewvar( ˆβ 1 ) + 2x ew Cov( ˆβ 0, ˆβ 1 ) = σ 2 + σ2 x2 i (x i x) + x2 ewσ 2 2 (x i x) 2x ew xσ 2 2 (x i x) ( 2 = σ (x ) ew x) 2 (x i x) 2 For kow σ 2, the predictio iterval for y ew is ( ŷ ew ± z α/2 σ (x ew x) 2 ). (x i x) 2 is For ukow σ 2, plug i MSE = ˆσ 2 ad the predictio iterval for y ew ( ŷ ew ± t 2,α/2 ˆσ (x ew x) 2 ). (x i x) 2 Which predictio iterval is wider? Why? Cofidece Iterval for expectatio of y If we are ot iterested i just oe particular studet with high school GPA=3.5, but we are iterested i the expectatio of college GPA for all studets with high school GPA=3.5. With kow σ 2, the Cofidece Iterval for y 0 = E(y x 0 ) is ŷ 0 ± z α/2 σ 2 ( 1 + (x 0 x) 2 (x i x) 2 ). For ukow σ 2, plug i MSE = ˆσ 2 ad the predictio iterval for y 0 is ( 1 ŷ 0 ± t 2,α/2 ˆσ 2 + (x 0 x) 2 ). (x i x) 2 2
22 Here ŷ 0 = ˆβ 0 + ˆβ 1 x 0. Is the predictio iterval wider or the cofidece iterval wider? Why? Multiple liear regressio Istead of studyig the relatioship betwee respose y ad oe idepedet variable (predictor) x, we may also use liear model to study the relatioship betwee respose y ad multiple idepedet variables x 1,..., x p. y = β 0 + β 1 x β p x p + ε. We have set of observatios (y i ; x i1, x i2,..., x ip ) for i = 1,...,. We obtai least square estimators of the ukow parameters by miimizig the SSE: S(β 1,..., β p ) = [y i (β 0 + β 1 x i1 + + β p x ip )] 2. Set the first order derivatives equal to 0: S = 0, S = 0,, S = 0. β 0 β 1 β p The solutios are deoted by ˆβ 0, ˆβ 1,..., ˆβ p. The fitted least squares lie is ŷ = ˆβ 0 + ˆβ 1 x ˆβ p x p. The i-th residual is r i = y i ŷ i. ANOVA The variatio of y has the same decompositio as i the simple liear regressio (y i ȳ) 2 = (ŷ i ȳ) 2 + (y i ŷ i ) 2 SST=SSR+ SSE 3
23 ANOVA table Source df Sum of Squares Mea Squares F Regressio p SSR = (ŷ i ȳ) 2 MSR=SSR/p F=MSR/MSE Error -p-1 SSE = (y i ŷ i ) 2 MSE=SSE/(-p-1) Total -1 SST = (y i ȳ) 2 To test whether there is sigificat predictor effect, simultaeous test Null hypothesis H 0 : β 1 = β 2 = = β p = 0 v.s. Alterative hypothesis H 1 : At least oe coefficiet ot equal to zero. If test statistic F is large, which meas a large proportio of variatio i y is explaied by the regressio, the ituitio is that we reject H 0. It ca be show that uder H 0, F = MSR/MSE F p, p 1. Deote the upper percetile F p, p 1,α, which satisfies P (F > F p, p 1,α ) = α for radom variable F F p, p 1. At 1 α sigificace level, reject H 0 if F = MSR/MSE > F p, p 1,α, ad this meas the regressio is sigificat. Fail to reject H 0 if F = MSR/MSE < F p, p 1,α, ad this meas oe of the regressio coefficiets sigificatly differs from 0. Notice that F = MSR MSE = = p 1 p SSR/(p) SSE/( p 1) = p 1 SSR p SST SSR SSR/SST SST/SST SSR/SST = p 1 R 2 p 1 R. 2 The R 2, or coefficiet of determiatio, is R 2 = SSR/SST = 1 SSE/SST. R 2 always icrease as the umber of predictors p icrease, which may result i model overfittig. We eed a ew criterio. The corrected R 2, or R 2, is R 2 = 1 SSE/( p 1). It takes degrees of SST/( 1) freedom ito cosideratio ad pealizes overfittig. 4
24 Lecture 7, March 2 Please read Chapter 2 of the textbook, from page 8 to page 41. You may skip Sectio ad Sectio For the basics of liear algebra about matrix calculatio, you may fid the etry from Wikipedia useful: http : //e.wikipedia.org/wiki/matrix ( mathematics) Matrix represetatio of multiple liear regressio We have set of idepedet observatios (y i ; x i1, x i2,..., x ip ) from model Let y = β 0 + β 1 x β p x p + ε. y 1 β 0 + β 1 x β p x 1p + ε 1 y 2. = β 0 + β 1 x β p x 2p + ε 2. y β 0 + β 1 x β p x p + ε y 1 β 0 1 x 11 x 1p ε 1 y 2 Y =., β = β 1., X = 1 x 21 x 2p...., ε = ε 2.. y β p 1 x 1 x p ε The we have Y = Xβ + ε, where ε is multivariate ormal with E(ε) = 0 ad Var(ε) = σ 2 I The SSE ca be writte as S(β) = [y i (β 0 + β 1 x i1 + + β p x ip )] 2 = (Y Xβ) (Y Xβ). 1
25 Take derivatives with respect to β ad we have the followig ormal equatio: X Xβ = X Y, which leads to least squares estimator ˆβ = (X X) 1 X Y. Because E(ˆβ) = E[(X X) 1 X Y] = (X X) 1 X E(Y) = (X X) 1 X E(Xβ + ε) = (X X) 1 X [Xβ + E(ε)] = (X X) 1 X Xβ = I p+1 β = β, ˆβ is ubiased estimator of β. Because Var(AX) = AVar(X)A for radom vector X ad costat matrix A, we have Var(ˆβ) = Var[(X X) 1 X Y] = [(X X) 1 X ]Var(Y)[(X X) 1 X ] = Var[(X X) 1 X Y] = [(X X) 1 X ]Var(Xβ + ε)[(x X) 1 X ] = σ 2 (X X) 1 X I X(X X) 1 = σ 2 (X X) 1. Thus we kow ˆβ N p+1 (β, σ 2 (X X) 1 ). Thus ˆβ i N(β i, σ 2 d i ), where d i is the i-th elemet o the diagoal of matrix (X X) 1. 2
1 Inferential Methods for Correlation and Regression Analysis
1 Iferetial Methods for Correlatio ad Regressio Aalysis I the chapter o Correlatio ad Regressio Aalysis tools for describig bivariate cotiuous data were itroduced. The sample Pearso Correlatio Coefficiet
More informationRandom Variables, Sampling and Estimation
Chapter 1 Radom Variables, Samplig ad Estimatio 1.1 Itroductio This chapter will cover the most importat basic statistical theory you eed i order to uderstad the ecoometric material that will be comig
More informationTopic 9: Sampling Distributions of Estimators
Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be
More informationProperties and Hypothesis Testing
Chapter 3 Properties ad Hypothesis Testig 3.1 Types of data The regressio techiques developed i previous chapters ca be applied to three differet kids of data. 1. Cross-sectioal data. 2. Time series data.
More informationSimple Linear Regression
Simple Liear Regressio 1. Model ad Parameter Estimatio (a) Suppose our data cosist of a collectio of pairs (x i, y i ), where x i is a observed value of variable X ad y i is the correspodig observatio
More information3/3/2014. CDS M Phil Econometrics. Types of Relationships. Types of Relationships. Types of Relationships. Vijayamohanan Pillai N.
3/3/04 CDS M Phil Old Least Squares (OLS) Vijayamohaa Pillai N CDS M Phil Vijayamoha CDS M Phil Vijayamoha Types of Relatioships Oly oe idepedet variable, Relatioship betwee ad is Liear relatioships Curviliear
More informationTopic 9: Sampling Distributions of Estimators
Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be
More informationLinear Regression Models
Liear Regressio Models Dr. Joh Mellor-Crummey Departmet of Computer Sciece Rice Uiversity johmc@cs.rice.edu COMP 528 Lecture 9 15 February 2005 Goals for Today Uderstad how to Use scatter diagrams to ispect
More informationTopic 9: Sampling Distributions of Estimators
Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be
More informationEcon 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara
Poit Estimator Eco 325 Notes o Poit Estimator ad Cofidece Iterval 1 By Hiro Kasahara Parameter, Estimator, ad Estimate The ormal probability desity fuctio is fully characterized by two costats: populatio
More informationThe variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.
SAMPLE STATISTICS A radom sample x 1,x,,x from a distributio f(x) is a set of idepedetly ad idetically variables with x i f(x) for all i Their joit pdf is f(x 1,x,,x )=f(x 1 )f(x ) f(x )= f(x i ) The sample
More informationStatistical Properties of OLS estimators
1 Statistical Properties of OLS estimators Liear Model: Y i = β 0 + β 1 X i + u i OLS estimators: β 0 = Y β 1X β 1 = Best Liear Ubiased Estimator (BLUE) Liear Estimator: β 0 ad β 1 are liear fuctio of
More informationRegression, Inference, and Model Building
Regressio, Iferece, ad Model Buildig Scatter Plots ad Correlatio Correlatio coefficiet, r -1 r 1 If r is positive, the the scatter plot has a positive slope ad variables are said to have a positive relatioship
More informationResponse Variable denoted by y it is the variable that is to be predicted measure of the outcome of an experiment also called the dependent variable
Statistics Chapter 4 Correlatio ad Regressio If we have two (or more) variables we are usually iterested i the relatioship betwee the variables. Associatio betwee Variables Two variables are associated
More informationStatistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.
Statistical Iferece (Chapter 10) Statistical iferece = lear about a populatio based o the iformatio provided by a sample. Populatio: The set of all values of a radom variable X of iterest. Characterized
More informationTAMS24: Notations and Formulas
TAMS4: Notatios ad Formulas Basic otatios ad defiitios X: radom variable stokastiska variabel Mea Vätevärde: µ = X = by Xiagfeg Yag kpx k, if X is discrete, xf Xxdx, if X is cotiuous Variace Varias: =
More informationEcon 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.
Eco 325/327 Notes o Sample Mea, Sample Proportio, Cetral Limit Theorem, Chi-square Distributio, Studet s t distributio 1 Sample Mea By Hiro Kasahara We cosider a radom sample from a populatio. Defiitio
More information(all terms are scalars).the minimization is clearer in sum notation:
7 Multiple liear regressio: with predictors) Depedet data set: y i i = 1, oe predictad, predictors x i,k i = 1,, k = 1, ' The forecast equatio is ŷ i = b + Use matrix otatio: k =1 b k x ik Y = y 1 y 1
More informationCorrelation Regression
Correlatio Regressio While correlatio methods measure the stregth of a liear relatioship betwee two variables, we might wish to go a little further: How much does oe variable chage for a give chage i aother
More informationStat 139 Homework 7 Solutions, Fall 2015
Stat 139 Homework 7 Solutios, Fall 2015 Problem 1. I class we leared that the classical simple liear regressio model assumes the followig distributio of resposes: Y i = β 0 + β 1 X i + ɛ i, i = 1,...,,
More informationCEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering
CEE 5 Autum 005 Ucertaity Cocepts for Geotechical Egieerig Basic Termiology Set A set is a collectio of (mutually exclusive) objects or evets. The sample space is the (collectively exhaustive) collectio
More informationLecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett
Lecture Note 8 Poit Estimators ad Poit Estimatio Methods MIT 14.30 Sprig 2006 Herma Beett Give a parameter with ukow value, the goal of poit estimatio is to use a sample to compute a umber that represets
More information11 Correlation and Regression
11 Correlatio Regressio 11.1 Multivariate Data Ofte we look at data where several variables are recorded for the same idividuals or samplig uits. For example, at a coastal weather statio, we might record
More informationResampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.
Jauary 1, 2019 Resamplig Methods Motivatio We have so may estimators with the property θ θ d N 0, σ 2 We ca also write θ a N θ, σ 2 /, where a meas approximately distributed as Oce we have a cosistet estimator
More informationECON 3150/4150, Spring term Lecture 3
Itroductio Fidig the best fit by regressio Residuals ad R-sq Regressio ad causality Summary ad ext step ECON 3150/4150, Sprig term 2014. Lecture 3 Ragar Nymoe Uiversity of Oslo 21 Jauary 2014 1 / 30 Itroductio
More informationLecture 11 Simple Linear Regression
Lecture 11 Simple Liear Regressio Fall 2013 Prof. Yao Xie, yao.xie@isye.gatech.edu H. Milto Stewart School of Idustrial Systems & Egieerig Georgia Tech Midterm 2 mea: 91.2 media: 93.75 std: 6.5 2 Meddicorp
More informationWorksheet 23 ( ) Introduction to Simple Linear Regression (continued)
Worksheet 3 ( 11.5-11.8) Itroductio to Simple Liear Regressio (cotiued) This worksheet is a cotiuatio of Discussio Sheet 3; please complete that discussio sheet first if you have ot already doe so. This
More informationLecture 22: Review for Exam 2. 1 Basic Model Assumptions (without Gaussian Noise)
Lecture 22: Review for Exam 2 Basic Model Assumptios (without Gaussia Noise) We model oe cotiuous respose variable Y, as a liear fuctio of p umerical predictors, plus oise: Y = β 0 + β X +... β p X p +
More informationStat 421-SP2012 Interval Estimation Section
Stat 41-SP01 Iterval Estimatio Sectio 11.1-11. We ow uderstad (Chapter 10) how to fid poit estimators of a ukow parameter. o However, a poit estimate does ot provide ay iformatio about the ucertaity (possible
More information7-1. Chapter 4. Part I. Sampling Distributions and Confidence Intervals
7-1 Chapter 4 Part I. Samplig Distributios ad Cofidece Itervals 1 7- Sectio 1. Samplig Distributio 7-3 Usig Statistics Statistical Iferece: Predict ad forecast values of populatio parameters... Test hypotheses
More informationLecture 3. Properties of Summary Statistics: Sampling Distribution
Lecture 3 Properties of Summary Statistics: Samplig Distributio Mai Theme How ca we use math to justify that our umerical summaries from the sample are good summaries of the populatio? Lecture Summary
More informationFACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures
FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING Lectures MODULE 5 STATISTICS II. Mea ad stadard error of sample data. Biomial distributio. Normal distributio 4. Samplig 5. Cofidece itervals
More informationStatistics 20: Final Exam Solutions Summer Session 2007
1. 20 poits Testig for Diabetes. Statistics 20: Fial Exam Solutios Summer Sessio 2007 (a) 3 poits Give estimates for the sesitivity of Test I ad of Test II. Solutio: 156 patiets out of total 223 patiets
More informationChapters 5 and 13: REGRESSION AND CORRELATION. Univariate data: x, Bivariate data (x,y).
Chapters 5 ad 13: REGREION AND CORRELATION (ectios 5.5 ad 13.5 are omitted) Uivariate data: x, Bivariate data (x,y). Example: x: umber of years studets studied paish y: score o a proficiecy test For each
More informationMATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4
MATH 30: Probability ad Statistics 9. Estimatio ad Testig of Parameters Estimatio ad Testig of Parameters We have bee dealig situatios i which we have full kowledge of the distributio of a radom variable.
More informationTABLES AND FORMULAS FOR MOORE Basic Practice of Statistics
TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics Explorig Data: Distributios Look for overall patter (shape, ceter, spread) ad deviatios (outliers). Mea (use a calculator): x = x 1 + x 2 + +
More informationSection 14. Simple linear regression.
Sectio 14 Simple liear regressio. Let us look at the cigarette dataset from [1] (available to dowload from joural s website) ad []. The cigarette dataset cotais measuremets of tar, icotie, weight ad carbo
More informationEfficient GMM LECTURE 12 GMM II
DECEMBER 1 010 LECTURE 1 II Efficiet The estimator depeds o the choice of the weight matrix A. The efficiet estimator is the oe that has the smallest asymptotic variace amog all estimators defied by differet
More informationMBACATÓLICA. Quantitative Methods. Faculdade de Ciências Económicas e Empresariais UNIVERSIDADE CATÓLICA PORTUGUESA 9. SAMPLING DISTRIBUTIONS
MBACATÓLICA Quatitative Methods Miguel Gouveia Mauel Leite Moteiro Faculdade de Ciêcias Ecoómicas e Empresariais UNIVERSIDADE CATÓLICA PORTUGUESA 9. SAMPLING DISTRIBUTIONS MBACatólica 006/07 Métodos Quatitativos
More informationSIMPLE LINEAR REGRESSION AND CORRELATION ANALYSIS
SIMPLE LINEAR REGRESSION AND CORRELATION ANALSIS INTRODUCTION There are lot of statistical ivestigatio to kow whether there is a relatioship amog variables Two aalyses: (1) regressio aalysis; () correlatio
More informationLecture 2: Monte Carlo Simulation
STAT/Q SCI 43: Itroductio to Resamplig ethods Sprig 27 Istructor: Ye-Chi Che Lecture 2: ote Carlo Simulatio 2 ote Carlo Itegratio Assume we wat to evaluate the followig itegratio: e x3 dx What ca we do?
More informationt distribution [34] : used to test a mean against an hypothesized value (H 0 : µ = µ 0 ) or the difference
EXST30 Backgroud material Page From the textbook The Statistical Sleuth Mea [0]: I your text the word mea deotes a populatio mea (µ) while the work average deotes a sample average ( ). Variace [0]: The
More informationCircle the single best answer for each multiple choice question. Your choice should be made clearly.
TEST #1 STA 4853 March 6, 2017 Name: Please read the followig directios. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directios This exam is closed book ad closed otes. There are 32 multiple choice questios.
More informationSTATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. Comments:
Recall: STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS Commets:. So far we have estimates of the parameters! 0 ad!, but have o idea how good these estimates are. Assumptio: E(Y x)! 0 +! x (liear coditioal
More informationGoodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen)
Goodess-of-Fit Tests ad Categorical Data Aalysis (Devore Chapter Fourtee) MATH-252-01: Probability ad Statistics II Sprig 2019 Cotets 1 Chi-Squared Tests with Kow Probabilities 1 1.1 Chi-Squared Testig................
More informationExpectation and Variance of a random variable
Chapter 11 Expectatio ad Variace of a radom variable The aim of this lecture is to defie ad itroduce mathematical Expectatio ad variace of a fuctio of discrete & cotiuous radom variables ad the distributio
More informationSimple Regression. Acknowledgement. These slides are based on presentations created and copyrighted by Prof. Daniel Menasce (GMU) CS 700
Simple Regressio CS 7 Ackowledgemet These slides are based o presetatios created ad copyrighted by Prof. Daiel Measce (GMU) Basics Purpose of regressio aalysis: predict the value of a depedet or respose
More informationContinuous Data that can take on any real number (time/length) based on sample data. Categorical data can only be named or categorised
Questio 1. (Topics 1-3) A populatio cosists of all the members of a group about which you wat to draw a coclusio (Greek letters (μ, σ, Ν) are used) A sample is the portio of the populatio selected for
More informationCHAPTER 8 FUNDAMENTAL SAMPLING DISTRIBUTIONS AND DATA DESCRIPTIONS. 8.1 Random Sampling. 8.2 Some Important Statistics
CHAPTER 8 FUNDAMENTAL SAMPLING DISTRIBUTIONS AND DATA DESCRIPTIONS 8.1 Radom Samplig The basic idea of the statistical iferece is that we are allowed to draw ifereces or coclusios about a populatio based
More informationUNIVERSITY OF TORONTO Faculty of Arts and Science APRIL/MAY 2009 EXAMINATIONS ECO220Y1Y PART 1 OF 2 SOLUTIONS
PART of UNIVERSITY OF TORONTO Faculty of Arts ad Sciece APRIL/MAY 009 EAMINATIONS ECO0YY PART OF () The sample media is greater tha the sample mea whe there is. (B) () A radom variable is ormally distributed
More informationChapter 1 Simple Linear Regression (part 6: matrix version)
Chapter Simple Liear Regressio (part 6: matrix versio) Overview Simple liear regressio model: respose variable Y, a sigle idepedet variable X Y β 0 + β X + ε Multiple liear regressio model: respose Y,
More informationFinal Examination Solutions 17/6/2010
The Islamic Uiversity of Gaza Faculty of Commerce epartmet of Ecoomics ad Political Scieces A Itroductio to Statistics Course (ECOE 30) Sprig Semester 009-00 Fial Eamiatio Solutios 7/6/00 Name: I: Istructor:
More informationAsymptotic Results for the Linear Regression Model
Asymptotic Results for the Liear Regressio Model C. Fli November 29, 2000 1. Asymptotic Results uder Classical Assumptios The followig results apply to the liear regressio model y = Xβ + ε, where X is
More informationStatistical and Mathematical Methods DS-GA 1002 December 8, Sample Final Problems Solutions
Statistical ad Mathematical Methods DS-GA 00 December 8, 05. Short questios Sample Fial Problems Solutios a. Ax b has a solutio if b is i the rage of A. The dimesio of the rage of A is because A has liearly-idepedet
More information[ ] ( ) ( ) [ ] ( ) 1 [ ] [ ] Sums of Random Variables Y = a 1 X 1 + a 2 X 2 + +a n X n The expected value of Y is:
PROBABILITY FUNCTIONS A radom variable X has a probabilit associated with each of its possible values. The probabilit is termed a discrete probabilit if X ca assume ol discrete values, or X = x, x, x 3,,
More informationST 305: Exam 3 ( ) = P(A)P(B A) ( ) = P(A) + P(B) ( ) = 1 P( A) ( ) = P(A) P(B) σ X 2 = σ a+bx. σ ˆp. σ X +Y. σ X Y. σ X. σ Y. σ n.
ST 305: Exam 3 By hadig i this completed exam, I state that I have either give or received assistace from aother perso durig the exam period. I have used o resources other tha the exam itself ad the basic
More informationCorrelation. Two variables: Which test? Relationship Between Two Numerical Variables. Two variables: Which test? Contingency table Grouped bar graph
Correlatio Y Two variables: Which test? X Explaatory variable Respose variable Categorical Numerical Categorical Cotigecy table Cotigecy Logistic Grouped bar graph aalysis regressio Mosaic plot Numerical
More informationDirection: This test is worth 150 points. You are required to complete this test within 55 minutes.
Term Test 3 (Part A) November 1, 004 Name Math 6 Studet Number Directio: This test is worth 10 poits. You are required to complete this test withi miutes. I order to receive full credit, aswer each problem
More informationFirst, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So,
0 2. OLS Part II The OLS residuals are orthogoal to the regressors. If the model icludes a itercept, the orthogoality of the residuals ad regressors gives rise to three results, which have limited practical
More informationEconomics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator
Ecoomics 24B Relatio to Method of Momets ad Maximum Likelihood OLSE as a Maximum Likelihood Estimator Uder Assumptio 5 we have speci ed the distributio of the error, so we ca estimate the model parameters
More informationSample Size Determination (Two or More Samples)
Sample Sie Determiatio (Two or More Samples) STATGRAPHICS Rev. 963 Summary... Data Iput... Aalysis Summary... 5 Power Curve... 5 Calculatios... 6 Summary This procedure determies a suitable sample sie
More informationDS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10
DS 00: Priciples ad Techiques of Data Sciece Date: April 3, 208 Name: Hypothesis Testig Discussio #0. Defie these terms below as they relate to hypothesis testig. a) Data Geeratio Model: Solutio: A set
More informationStatistics 511 Additional Materials
Cofidece Itervals o mu Statistics 511 Additioal Materials This topic officially moves us from probability to statistics. We begi to discuss makig ifereces about the populatio. Oe way to differetiate probability
More informationBig Picture. 5. Data, Estimates, and Models: quantifying the accuracy of estimates.
5. Data, Estimates, ad Models: quatifyig the accuracy of estimates. 5. Estimatig a Normal Mea 5.2 The Distributio of the Normal Sample Mea 5.3 Normal data, cofidece iterval for, kow 5.4 Normal data, cofidece
More informationA statistical method to determine sample size to estimate characteristic value of soil parameters
A statistical method to determie sample size to estimate characteristic value of soil parameters Y. Hojo, B. Setiawa 2 ad M. Suzuki 3 Abstract Sample size is a importat factor to be cosidered i determiig
More informationLecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting
Lecture 6 Chi Square Distributio (χ ) ad Least Squares Fittig Chi Square Distributio (χ ) Suppose: We have a set of measuremets {x 1, x, x }. We kow the true value of each x i (x t1, x t, x t ). We would
More informationOutput Analysis (2, Chapters 10 &11 Law)
B. Maddah ENMG 6 Simulatio Output Aalysis (, Chapters 10 &11 Law) Comparig alterative system cofiguratio Sice the output of a simulatio is radom, the comparig differet systems via simulatio should be doe
More informationLinear Regression Models, OLS, Assumptions and Properties
Chapter 2 Liear Regressio Models, OLS, Assumptios ad Properties 2.1 The Liear Regressio Model The liear regressio model is the sigle most useful tool i the ecoometricia s kit. The multiple regressio model
More informationUnderstanding Samples
1 Will Moroe CS 109 Samplig ad Bootstrappig Lecture Notes #17 August 2, 2017 Based o a hadout by Chris Piech I this chapter we are goig to talk about statistics calculated o samples from a populatio. We
More informationTopic 10: Introduction to Estimation
Topic 0: Itroductio to Estimatio Jue, 0 Itroductio I the simplest possible terms, the goal of estimatio theory is to aswer the questio: What is that umber? What is the legth, the reactio rate, the fractio
More informationII. Descriptive Statistics D. Linear Correlation and Regression. 1. Linear Correlation
II. Descriptive Statistics D. Liear Correlatio ad Regressio I this sectio Liear Correlatio Cause ad Effect Liear Regressio 1. Liear Correlatio Quatifyig Liear Correlatio The Pearso product-momet correlatio
More informationECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors
ECONOMETRIC THEORY MODULE XIII Lecture - 34 Asymptotic Theory ad Stochastic Regressors Dr. Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Asymptotic theory The asymptotic
More informationChapter 13, Part A Analysis of Variance and Experimental Design
Slides Prepared by JOHN S. LOUCKS St. Edward s Uiversity Slide 1 Chapter 13, Part A Aalysis of Variace ad Eperimetal Desig Itroductio to Aalysis of Variace Aalysis of Variace: Testig for the Equality of
More informationIn this section we derive some finite-sample properties of the OLS estimator. b is an estimator of β. It is a function of the random sample data.
17 3. OLS Part III I this sectio we derive some fiite-sample properties of the OLS estimator. 3.1 The Samplig Distributio of the OLS Estimator y = Xβ + ε ; ε ~ N[0, σ 2 I ] b = (X X) 1 X y = f(y) ε is
More informationModule 1 Fundamentals in statistics
Normal Distributio Repeated observatios that differ because of experimetal error ofte vary about some cetral value i a roughly symmetrical distributio i which small deviatios occur much more frequetly
More informationMA Advanced Econometrics: Properties of Least Squares Estimators
MA Advaced Ecoometrics: Properties of Least Squares Estimators Karl Whela School of Ecoomics, UCD February 5, 20 Karl Whela UCD Least Squares Estimators February 5, 20 / 5 Part I Least Squares: Some Fiite-Sample
More informationLecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting
Lecture 6 Chi Square Distributio (χ ) ad Least Squares Fittig Chi Square Distributio (χ ) Suppose: We have a set of measuremets {x 1, x, x }. We kow the true value of each x i (x t1, x t, x t ). We would
More informationThe Method of Least Squares. To understand least squares fitting of data.
The Method of Least Squares KEY WORDS Curve fittig, least square GOAL To uderstad least squares fittig of data To uderstad the least squares solutio of icosistet systems of liear equatios 1 Motivatio Curve
More informationData Analysis and Statistical Methods Statistics 651
Data Aalysis ad Statistical Methods Statistics 651 http://www.stat.tamu.edu/~suhasii/teachig.html Suhasii Subba Rao Review of testig: Example The admistrator of a ursig home wats to do a time ad motio
More informationDiscrete Mathematics for CS Spring 2008 David Wagner Note 22
CS 70 Discrete Mathematics for CS Sprig 2008 David Wager Note 22 I.I.D. Radom Variables Estimatig the bias of a coi Questio: We wat to estimate the proportio p of Democrats i the US populatio, by takig
More informationProbability and statistics: basic terms
Probability ad statistics: basic terms M. Veeraraghava August 203 A radom variable is a rule that assigs a umerical value to each possible outcome of a experimet. Outcomes of a experimet form the sample
More informationFinal Review. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech
Fial Review Fall 2013 Prof. Yao Xie, yao.xie@isye.gatech.edu H. Milto Stewart School of Idustrial Systems & Egieerig Georgia Tech 1 Radom samplig model radom samples populatio radom samples: x 1,..., x
More informationECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015
ECE 8527: Itroductio to Machie Learig ad Patter Recogitio Midterm # 1 Vaishali Ami Fall, 2015 tue39624@temple.edu Problem No. 1: Cosider a two-class discrete distributio problem: ω 1 :{[0,0], [2,0], [2,2],
More informationDirection: This test is worth 250 points. You are required to complete this test within 50 minutes.
Term Test October 3, 003 Name Math 56 Studet Number Directio: This test is worth 50 poits. You are required to complete this test withi 50 miutes. I order to receive full credit, aswer each problem completely
More informationLecture 7: Properties of Random Samples
Lecture 7: Properties of Radom Samples 1 Cotiued From Last Class Theorem 1.1. Let X 1, X,...X be a radom sample from a populatio with mea µ ad variace σ
More informationLinear Regression Demystified
Liear Regressio Demystified Liear regressio is a importat subject i statistics. I elemetary statistics courses, formulae related to liear regressio are ofte stated without derivatio. This ote iteds to
More informationLesson 11: Simple Linear Regression
Lesso 11: Simple Liear Regressio Ka-fu WONG December 2, 2004 I previous lessos, we have covered maily about the estimatio of populatio mea (or expected value) ad its iferece. Sometimes we are iterested
More informationBinomial Distribution
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0 1 2 3 4 5 6 7 0.0 0.5 1.0 1.5 2.0 2.5 3.0 Overview Example: coi tossed three times Defiitio Formula Recall that a r.v. is discrete if there are either a fiite umber of possible
More information1 Models for Matched Pairs
1 Models for Matched Pairs Matched pairs occur whe we aalyse samples such that for each measuremet i oe of the samples there is a measuremet i the other sample that directly relates to the measuremet i
More informationLinear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d
Liear regressio Daiel Hsu (COMS 477) Maximum likelihood estimatio Oe of the simplest liear regressio models is the followig: (X, Y ),..., (X, Y ), (X, Y ) are iid radom pairs takig values i R d R, ad Y
More informationTMA4245 Statistics. Corrected 30 May and 4 June Norwegian University of Science and Technology Department of Mathematical Sciences.
Norwegia Uiversity of Sciece ad Techology Departmet of Mathematical Scieces Corrected 3 May ad 4 Jue Solutios TMA445 Statistics Saturday 6 May 9: 3: Problem Sow desity a The probability is.9.5 6x x dx
More informationInferential Statistics. Inference Process. Inferential Statistics and Probability a Holistic Approach. Inference Process.
Iferetial Statistics ad Probability a Holistic Approach Iferece Process Chapter 8 Poit Estimatio ad Cofidece Itervals This Course Material by Maurice Geraghty is licesed uder a Creative Commos Attributio-ShareAlike
More informationIf, for instance, we were required to test whether the population mean μ could be equal to a certain value μ
STATISTICAL INFERENCE INTRODUCTION Statistical iferece is that brach of Statistics i which oe typically makes a statemet about a populatio based upo the results of a sample. I oesample testig, we essetially
More informationProbability and Statistics
ICME Refresher Course: robability ad Statistics Staford Uiversity robability ad Statistics Luyag Che September 20, 2016 1 Basic robability Theory 11 robability Spaces A probability space is a triple (Ω,
More informationStatistics 203 Introduction to Regression and Analysis of Variance Assignment #1 Solutions January 20, 2005
Statistics 203 Itroductio to Regressio ad Aalysis of Variace Assigmet #1 Solutios Jauary 20, 2005 Q. 1) (MP 2.7) (a) Let x deote the hydrocarbo percetage, ad let y deote the oxyge purity. The simple liear
More informationOverview. p 2. Chapter 9. Pooled Estimate of. q = 1 p. Notation for Two Proportions. Inferences about Two Proportions. Assumptions
Chapter 9 Slide Ifereces from Two Samples 9- Overview 9- Ifereces about Two Proportios 9- Ifereces about Two Meas: Idepedet Samples 9-4 Ifereces about Matched Pairs 9-5 Comparig Variatio i Two Samples
More informationEstimation for Complete Data
Estimatio for Complete Data complete data: there is o loss of iformatio durig study. complete idividual complete data= grouped data A complete idividual data is the oe i which the complete iformatio of
More informationUniversity of California, Los Angeles Department of Statistics. Simple regression analysis
Uiversity of Califoria, Los Ageles Departmet of Statistics Statistics 100C Istructor: Nicolas Christou Simple regressio aalysis Itroductio: Regressio aalysis is a statistical method aimig at discoverig
More informationOpen book and notes. 120 minutes. Cover page and six pages of exam. No calculators.
IE 330 Seat # Ope book ad otes 120 miutes Cover page ad six pages of exam No calculators Score Fial Exam (example) Schmeiser Ope book ad otes No calculator 120 miutes 1 True or false (for each, 2 poits
More informationHomework for 4/9 Due 4/16
Name: ID: Homework for 4/9 Due 4/16 1. [ 13-6] It is covetioal wisdom i military squadros that pilots ted to father more girls tha boys. Syder 1961 gathered data for military fighter pilots. The sex of
More information