F Tests and F statistics

Similar documents
σ σ MLR Models: Estimation and Inference v.3 SLR.1: Linear Model MLR.1: Linear Model Those (S/M)LR Assumptions MLR3: No perfect collinearity

Ordinary Least Squares (OLS): Multiple Linear Regression (MLR) Assessment I What s New? & Goodness-of-Fit

Ordinary Least Squares (OLS): Multiple Linear Regression (MLR) Analytics What s New? Not Much!

Lecture 4: Multivariate Regression, Part 2

Multiple Regression Analysis

Economics 326 Methods of Empirical Research in Economics. Lecture 14: Hypothesis testing in the multiple regression model, Part 2

Multiple Regression: Inference

2.1. Consider the following production function, known in the literature as the transcendental production function (TPF).

Lecture 4: Multivariate Regression, Part 2

Empirical Application of Simple Regression (Chapter 2)

1 Independent Practice: Hypothesis tests for one parameter:

Lecture 5: Hypothesis testing with the classical linear model

Statistical Inference with Regression Analysis

Problem Set 1 ANSWERS

ECO220Y Simple Regression: Testing the Slope

Group Comparisons: Differences in Composition Versus Differences in Models and Effects

Handout 12. Endogeneity & Simultaneous Equation Models

Rockefeller College University at Albany

1 Warm-Up: 2 Adjusted R 2. Introductory Applied Econometrics EEP/IAS 118 Spring Sylvan Herskowitz Section #

Lab 10 - Binary Variables

Section Least Squares Regression

(a) Briefly discuss the advantage of using panel data in this situation rather than pure crosssections

1 The basics of panel data

At this point, if you ve done everything correctly, you should have data that looks something like:

STAT 3A03 Applied Regression With SAS Fall 2017

Statistical Modelling in Stata 5: Linear Models

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Acknowledgements. Outline. Marie Diener-West. ICTR Leadership / Team INTRODUCTION TO CLINICAL RESEARCH. Introduction to Linear Regression

sociology 362 regression

Hypothesis Tests and Confidence Intervals in Multiple Regression

Nonrecursive Models Highlights Richard Williams, University of Notre Dame, Last revised April 6, 2015

General Linear Model (Chapter 4)

Lecture 13 Extra Sums of Squares

Nonlinear Regression Functions

Interpreting coefficients for transformed variables

Question 1 [17 points]: (ch 11)

Problem Set #5-Key Sonoma State University Dr. Cuellar Economics 317- Introduction to Econometrics

Chapter 7. Hypothesis Tests and Confidence Intervals in Multiple Regression

ECON Introductory Econometrics. Lecture 7: OLS with Multiple Regressors Hypotheses tests

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Econometrics Midterm Examination Answers

Problem Set 10: Panel Data

Measurement Error. Often a data set will contain imperfect measures of the data we would ideally like.

Lecture 7: OLS with qualitative information

Lecture#12. Instrumental variables regression Causal parameters III

Computer Exercise 3 Answers Hypothesis Testing

Lab 07 Introduction to Econometrics

Inference. ME104: Linear Regression Analysis Kenneth Benoit. August 15, August 15, 2012 Lecture 3 Multiple linear regression 1 1 / 58

Regression #8: Loose Ends

Introduction to Econometrics. Multiple Regression (2016/2017)

1 Linear Regression Analysis The Mincer Wage Equation Data Econometric Model Estimation... 11

Lecture 3: Multivariate Regression

Binary Dependent Variables

Answer all questions from part I. Answer two question from part II.a, and one question from part II.b.

Specification Error: Omitted and Extraneous Variables

Autocorrelation. Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time

Econometrics. 8) Instrumental variables

sociology 362 regression

Lecture 3: Inference in SLR

STAT 350: Summer Semester Midterm 1: Solutions

ECON3150/4150 Spring 2016

Lab 6 - Simple Regression

SMA 6304 / MIT / MIT Manufacturing Systems. Lecture 10: Data and Regression Analysis. Lecturer: Prof. Duane S. Boning

Answers: Problem Set 9. Dynamic Models

1 Multiple Regression

10) Time series econometrics

STATISTICS 110/201 PRACTICE FINAL EXAM

Handout 11: Measurement Error

Lecture 14. More on using dummy variables (deal with seasonality)

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

Applied Statistics and Econometrics

Instrumental Variables, Simultaneous and Systems of Equations

Using the same data as before, here is part of the output we get in Stata when we do a logistic regression of Grade on Gpa, Tuce and Psi.

Week 3: Simple Linear Regression

Practice exam questions

Applied Statistics and Econometrics

Lecture 10 Multiple Linear Regression

1. The shoe size of five randomly selected men in the class is 7, 7.5, 6, 6.5 the shoe size of 4 randomly selected women is 6, 5.

Section I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem -

1: a b c d e 2: a b c d e 3: a b c d e 4: a b c d e 5: a b c d e. 6: a b c d e 7: a b c d e 8: a b c d e 9: a b c d e 10: a b c d e

Sampling Distributions in Regression. Mini-Review: Inference for a Mean. For data (x 1, y 1 ),, (x n, y n ) generated with the SRM,

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

sociology sociology Scatterplots Quantitative Research Methods: Introduction to correlation and regression Age vs Income

ECON2228 Notes 7. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 41

Problem 4.1. Problem 4.3

Multicollinearity Richard Williams, University of Notre Dame, Last revised January 13, 2015

Lecture 11: Simple Linear Regression

Introduction to Econometrics. Review of Probability & Statistics

SOCY5601 Handout 8, Fall DETECTING CURVILINEARITY (continued) CONDITIONAL EFFECTS PLOTS

Review of Statistics 101

Econometrics. 4) Statistical inference

Problem Set #3-Key. wage Coef. Std. Err. t P> t [95% Conf. Interval]

Introduction to Econometrics. Multiple Regression

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

Functional Form. So far considered models written in linear form. Y = b 0 + b 1 X + u (1) Implies a straight line relationship between y and X

Self-Assessment Weeks 8: Multiple Regression with Qualitative Predictors; Multiple Comparisons

Motivation for multiple regression

Heteroskedasticity. Occurs when the Gauss Markov assumption that the residual variance is constant across all observations in the data set

Introductory Econometrics. Lecture 13: Hypothesis testing in the multiple regression model, Part 1

Transcription:

F Tests and F statistics Testing Linear estrictions F Stats and F Tests F Distributions F stats (w/ ) F Stats and tstat s eported F Stat's in OLS Output Example I: Bodyfat Babies and Bathwater F Stats, Adjusted -squared, MSE and tstats Adding and Dropping HS Variables Example II: More Bodyfat Testing Linear estrictions 1. Why do we test? Because we want to know! Maybe we thought the world was flat and it certainly looks flat! but maybe we have some evidence to the contrary. So run the test!. We use the F test to test linear restrictions in SL and ML models. What's a linear restriction?, you ask well here are some examples: 1 1) β1 β =, ) β1+ β = 0, 3) β1 = 0and β = 0, and 4) β1 = 1and β =. a. We ll be interested in counting the number of restrictions. To do that, just count the number of equals signs ('='s). So in the previous examples, 1) and ) have one restriction, and 3) and 4) have two. i. For some unknown reason we use the letter q to count the number of restrictions. 3. unning the F test: a. Step 1: Start with the Null hypothesis that it s A-OK to impose some linear restrictions on the OLS model. b. Step :Estimate the model with and without the restrictions and focus on the SSs and how they change. Since we ve imposed a restriction (or restrictions) on the estimated coefficients, the SSs will almost always increase: SS SSU. c. Step 3: OK, so SSs increased. But by how much? a lot? or maybe not so much? 1 Linear restrictions are linear functions of the unknown parameters. The subscripts refer to the estricted and Unrestricted models.

i. Big increase in SSs: If SSs increase by a lot (whatever that is) then the restrictions severely impacted the performance of the model, and so we reject the Null Hypothesis (which was that the restrictions were A-OK). ii. It's not so big: But if not so much, then maybe those restrictions weren t so bad after all, and we might fail to reject. Which is to say that it really was A-OK to impose those restrictions. 4. As you'll see below, running F tests in Stata is a snap. For example, to test the linear restrictions above, you would run the following Stata commands just after estimating your OLS model ( reg y x1 x ): a. β1 β b. β1 β = : test ( x1 = x) + = 0: test ( x1+ x = 0) = 0 = 0: test ( x1 = 0) ( x = 0) or just test x1 x ( ' = 0' is assumed if no value is specified) c. β1 and β d. β1 and β = 1 = : test ( x1 = 1) ( x = ) F Stats and F Tests 5. We use the F statistic (and the F distribution) to do the F test. a. The F statistic is defined by: Fstat = F = ( ) / SS SSU q, SS / ( n k 1) U where q is the number of restrictions (e.g. the number of '='s), and n k 1 is the number of degrees of freedom in the unrestricted (U) model. b. By construction F 0, assuming that F is well defined (since SS SSU ). 6. The F statistic as an elasticity! Who knew? a. The F statistic is really just an elasticity. We can rewrite the equation for F as: ( ) / % SS SSU SSU SS F = = q / ( n k 1) % dofs b. So the F statistics tells you the %change in SSs for a given %change in degrees of freedom (you might call this bang per buck).

7. We do the F test following the classical hypothesis testing protocol: a. The Null Hypothesis is that the restrictions are A-OK. b. Focus on Type I errors and the probability of a False ejection. c. Pick a small significance level, α, say α =.05, which will be the maximum acceptable probability of a False ejection. d. For the given Fstat, compute the p-value as the probability in the tail to the right of the Fstat: p = P( F( q, n k 1) > Fstat). 3 e. The p value is the probability that you d observe an F statistic at least as large as Fstat when the Null Hypothesis was in fact true. f. eject the Null Hypothesis if the p < α and fail to reject otherwise. i. So reject the Null Hypothesis if the F statistic is large ( % SS % dof ) and the associated probability level (p-value) is small (less than the significance level α ). In other words: the restrictions are rejected because when they were imposed, the world got a whole lot worse. ii. If the Fstat (elasticity) is small ( % SS % dof ) so that relatively speaking, SSs did not increase by very much, then the associated probability level (p-value) is high, and the restrictions are not rejected and are deemed to be A-OK. 8. And so we reject the Null Hypothesis if the probability value associated with this test statistic p < α or equivalently, if the F statistic is above is below the desired significance level ( ) the critical value c: Fstat > c, where { (, 1) } prob F q n k > c = α. 3 Here's Stata syntax for the right tail F probability: Ftail(q, n-k-1, Fstat). 3

F Distributions 9. Under ML.1-ML.6, the F statistic will have a Fqn (, k 1) distribution. Here are some F distributions for different parameter values: F(,0) Distribution F(3,0) Distribution F(10,0) Distribution F(15,0) Distribution F stats (w/ ) 10. You can also derive the F statistic using 's. a. Define SS = 1 and SST U SSU = 1, and so SST ( U ) ( U ) SST 1 1 / q / q ( n k 1) F = = = SST 1 / ( n k 1) 1 / ( n k 1) q 1 where ( U ) =. ( ) U U U, 4

F Stats and tstat s: Testing a single parameter 11. If you apply the F test to the case of testing just one parameter value (say, testing β x = 0 ), the F test will effectively be the same as the t test. So there is no inconsistency between the two tests. In fact, for these tests Fstat = tstat, and the p-values for the two statistics will be the same, since the F distribution will essentially be the square of the t distribution: t ˆ F β x 0 x β = (1, 1) < = < n k 1 < for x > 0. = and prob{ F n k x } prob{ x t x} (See examples below.) 1. In fact, you've already seen these sorts of Fstats in action. You may recall that quite a ways back (ML Assessment) we observed that in SL and ML models, a variable's t stat reflected it's incremental contribution to : 1 ˆ = dofs x βx t, where x is a HS variable's incremental contribution to a. If you consider the full model to be unrestricted, and the restricted model to restrict the x coefficient to be zero (so effectively dropping x from the model), the F test statistic is: ( ) ( n k 1) F = = 1 1 1 U x dofs U b. And since t ˆx = F, we have t β ˆ = dofs x βx. 1 13. So the connection between t stats and incremental, which probably seemed to you to have come out of nowhere, was in fact just an example of F stats in action... 5

eported F Stat's in OLS Output 14. Standard regression packages report the F statistic associated with the particular F test of the Null Hypothesis that all of the (non-intercept) parameter values are zero. This is sometimes referred to as testing the overall significance of the regression. 15. Although there was no mention of statistical tests at the time, you may recall the following discussion of F statistics from ML Assessment: The F statistic, or F stat for short, is useful in seeing whether the whole lot of HS variables have explanatory power so it has something to say about the collective precision of estimation of the HS variables as a group. This topic will be discussed in much more detail when we get to Inference, but for now here's the definition of the F stat: dofs dofs SSE Fstat = = alternatively, k 1 k SS n 1 Fstat = 1+ k 1 In practice, the reported F stats are almost always quite sizable (in double, if not triple, digits). If your F stat is even close to single digits, you probably have a crummy model and need to find a better group of explanatory variables or better data or maybe both!. 16. So now you know. That F test can be used to test the Null hypothesis that all of the (non-intercept) parameter values are zero. And as I said before: If you cannot reject that hypothesis, you have a really crummy model. 17. In conducting the test, the restricted model has = 0, since the dependent variable is regressed on just the constant term. And since U = (the reported -squared for the / k regression), we have F = the reported 1 / ( n k 1) F statistic, used to assess the overall statistical significance of the regression.. 6

Example I: Bodyfat 18. Testing the Null hypothesis that the true hgt parameter is zero in the following SL model:. reg Brozek wgt abd hgt Source SS df MS Number of obs = 5 -------------+---------------------------------- F(3, 48) = 13.67 Model 1087.5504 3 364.18347 Prob > F = 0.0000 esidual 406.4663 48 16.9615574 -squared = 0.710 -------------+---------------------------------- Adj -squared = 0.7177 Total 15079.0166 51 60.0757635 oot MSE = 4.1184 ------------------------------------------------------------------------------ Brozek Coef. Std. Err. t P> t [95% Conf. Interval] -------------+---------------------------------------------------------------- wgt -.10415.0516-5.41 0.000 -.164411 -.0765888 abd.879846.0579164 15.19 0.000.7657751.9939168 hgt -.1181607.08419-1.43 0.153 -.804915.0441701 _cons -3.6647 6.51936-5.01 0.000-45.5085-19.81 ------------------------------------------------------------------------------. test hgt ( 1) hgt = 0 F( 1, 48) =.06 Prob > F = 0.159 Notice the equivalence of the F test and the t test: 1.43 =.04 (rounding error) and Prob > F = 0.159 = P > t = 0.153. Since the p value is.15 (well above.1 and.05), we cannot reject the null hypothesis that the true hgt parameter is 0, at any standard (and attractive) level of statistical significance.. test wgt abd hgt ( 1) wgt = 0 ( ) abd = 0 ( 3) hgt = 0 F( 3, 48) = 13.67 Prob > F = 0.0000 Notice the agreement with the reported F(3.48) for the regression as well as Prob > F. Here are some examples of other F tests, testing for equality of parameters:. test wgt = hgt ( 1) wgt - hgt = 0 F( 1, 48) = 0.00 Prob > F = 0.981. test wgt = abd ( 1) wgt - abd = 0 F( 1, 48) = 161.61 Prob > F = 0.0000 7

And here's a replication of the F stat for the regression, F(3.48): / k.710 / 3 F = 13.63 1 / ( n k 1) = (1.710) / 48 = close enough Finally here's the by-hand F test for testing the null hypothesis that the hgt parameter is 0. We need to run the restricted model:. reg Brozek wgt abd Source SS df MS Number of obs = 5 -------------+---------------------------------- F(, 49) = 318.13 Model 10837.6881 5418.84407 Prob > F = 0.0000 esidual 441.3849 49 17.0334478 -squared = 0.7187 -------------+---------------------------------- Adj -squared = 0.7165 Total 15079.0166 51 60.0757635 oot MSE = 4.17 ------------------------------------------------------------------------------ Brozek Coef. Std. Err. t P> t [95% Conf. Interval] -------------+---------------------------------------------------------------- wgt -.1364535.019756-7.08 0.000 -.1744175 -.0984895 abd.915138.055355 17.4 0.000.8116675 1.018609 _cons -41.3481.41986-17.14 0.000-46.10059-36.59566 ------------------------------------------------------------------------------ Now that we have the SSs ( SS and this test: ( SS SS ) SS ( ) SS U ), we can compute the F stat associated with U / U 4, 41.33 4, 06.47 / 4, 06.47 F = = =.055 and the Prob > F: q/ ( n k 1) 1/ 48 Stata:. di Ftail(1,48,.055).1596678 F(1,48) Distribution 0 1 3 4 Ftail(1, 48,.055) =.1596678 8

Babies and Bathwater 19. Be careful about throwing out the baby with the bath water you don t want to exclude a significant explanatory variable from your model just because it happens to be associated with a set of variables that are jointly insignificant. Or put differently: F tests judge variables by the friends they keep! Example: Here s an example using a sample from a sovereign debt dataset. The F test does not reject at the 10% level the Null Hypothesis that the inflation and deficit_gdp parameters are zero even though deficit_gdp is statistically significant at almost the 5% level (and has p < 0.05 when inflation is dropped from the model). But inflation is so statistically insignificant, it pulls down deficit_gdp in the process when the F test (of the Null Hypothesis that βinflation = βdeficit _ gdp = 0 ) is conducted.. reg NSate corrupt gdp inflation deficit_gdp debt_gdp eurozone if _n < 30 Source SS df MS Number of obs = 9 -------------+---------------------------------- F(6, ) = 1.99 Model 88.81105 6 14.800351 Prob > F = 0.0000 esidual 5.065705 1.13935093 -squared = 0.7799 -------------+---------------------------------- Adj -squared = 0.7199 Total 113.877931 8 4.06706897 oot MSE = 1.0674 ------------------------------------------------------------------------------ NSate Coef. Std. Err. t P> t [95% Conf. Interval] -------------+---------------------------------------------------------------- corrupt.644807.1078044 5.98 0.000.4134.8683797 gdp.000144.0000765.80 0.010.0000557.0003731 inflation.0361479.0846488 0.43 0.674 -.139403.116988 deficit_gdp -.073749.035707 -.05 0.05 -.147366.0007768 debt_gdp -.0078.0094606 -.33 0.09 -.041698 -.004581 eurozone.999665.471874.1 0.046.003699 1.978883 _cons 4.69966 1.6366 3.48 0.00 1.76638 6.81395 ------------------------------------------------------------------------------. test inflation deficit_gdp ( 1) inflation = 0 ( ) deficit_gdp = 0 F(, ) =. Prob > F = 0.130 9

F stats, Adjusted -squared, MSE and t Stats 0. We previously considered adjusted ( ) in ML Assessment Goodness-of-Fit. At that time we were focused on the impacts of adding and subtracting individual explanatory variables one-by-one. 1. ecall that was developed in an attempt to adjust the coefficient of determination for the fact that when you add HS variables to a model, cannot decline since SS cannot increase. Small decreases in SSs will not generate a higher adj ; larger decreases will and what is small or large will also depend on how many changes were made.. From ML Assessment: a. Adjusted is defined: ( n 1) i. Since > 0, ( n k 1) SS ( n 1) MSE = 1 = 1. SST ( n k 1) S, with the difference inversely related to k. ii. For given S YY adding or subtracting HS variables from a ML model, exactly opposite directions., adjusted increases if and only if MSE decreases. So if you are YY and MSE will move in 3. We now move from the world of one-by-one inclusions/exclusions to the more general cases of adding or subtracting multiple explanatory variables. 4. Consider two models: U unrestricted and restricted. You can think of the process of adding HS variables to a model as being equivalent to moving from a restricted model (in which the coefficients for all of the added variables are zero) to an unrestricted model (in which their estimated coefficients are no longer constrained to be zero). adj from adding HS var(s) to the model Model: fewer HS vars U Model: more HS vars F statistic from dropping HS var(s) (restricting coeffs to be zero) 10

5. It turns out that the change in adj in dropping the restrictions (adding the variables to the model) is closely related to the F-statistic associated with imposing the restrictions (dropping the variables from the model). The change in adjusted in going from one model to the other is defined by: 1 1 SS SSU U = ( MSE MSEU ) =. SYY SYY ( n k 1) + q n k 1 q 1 ( n k 1) + q U After much algebra, this is: = [ F 1] U a. The sign of this expression will depend on whether the F statistic is greater or less than 1: increases if and only if the F statistic associated with the restrictions F test exceeds 1. b. ecall that for a single restriction, the F statistic is the square of the t stat, and so, as you saw in ML Assessment, increases if and only if the magnitude of the t stat of the added HS variable exceeds 1.. Adding and Dropping HS Variables 6. This last fact is useful when considering dropping HS variables from a model. a. One HS variable: i. t stat > 1: If a HS variable has a t stat > 1 then dropping that variable from the model will cause the adjusted -squared to decline. 1. Put differently: Adding HS variables with t stat > 1 will lead to higher. ii. t stat < 1: And if t stat < 1, then adjusted -squared will increase when that variable is dropped from the model. 4 b. Multiple HS variables: i. Fstat > 1: If the Fstat associated with dropping multiple HS variables is > 1 then dropping those variables from the model will cause adjusted -squared to decline. 1. Put differently: Adding HS variables with Fstat > 1 will lead to higher. ii. Fstat < 1: And if Fstat < 1, then adjusted -squared will increase when those variables are dropped from the model. 7. elation to statistical significance: It could be that additional HS variables that lead to a higher are not statistically significant. Adding a HS variable with t stat > 1 will be associated with higher ; but statistical significance will generally want to see a t stat > or so. 4 And yes, if t stat = 0, then there will be no change to adjusted -squared when the variable is dropped from the model. 11

a. If the added variable is statistically significant (so its t stat is above 1), then when that variable was added to the model. increased b. But it could be that the new variable is not statistically significant, even though adjusted -squared increased when the variable was added to the model. c. Example: Spoze that you have a HS variable with a t stat of 1.5. We would normally say that that variable was not statistically significant at standard significance levels. And so you might be tempted to toss that variable from your ML model. But if you do that, will decline, since the t stat > 1. d. This highlights a tension between two approaches to ML model building. You could be focused on maximizing or perhaps you are more focused on statistical significance. These two objectives can conflict at times. Deal with it! prob (F(q, n-k-1) > 1) prob (F(q, n-k-1) > ) prob (F(q, n-k-1) > 4) q: #restrictions q: #restrictions q: #restrictions dofs 1 3 4 5 10 1 3 4 5 10 1 3 4 5 10 dofs 10 34% 40% 43% 45% 47% 19% 19% 18% 17% 16% 7% 5% 4% 3% 3% 10 0 33% 39% 41% 43% 44% 48% 17% 16% 15% 13% 1% 9% 6% 3% % % 1% 0% 0 50 3% 38% 40% 4% 43% 46% 16% 15% 13% 11% 9% 5% 5% % 1% 1% 0% 0% 50 100 3% 37% 40% 41% 4% 45% 16% 14% 1% 10% 9% 4% 5% % 1% 0% 0% 0% 100 50 3% 37% 39% 41% 4% 44% 16% 14% 11% 10% 8% 3% 5% % 1% 0% 0% 0% 50 500 3% 37% 39% 41% 4% 44% 16% 14% 11% 9% 8% 3% 5% % 1% 0% 0% 0% 500 1

Example II: More Bodyfat -------------------------------------------------------------- (1) () (3) (4) Brozek Brozek Brozek Brozek -------------------------------------------------------------- wgt -0.151*** -0.19*** -0.154*** -0.13*** (-5.1) (-3.68) (-4.9) (-3.56) abd 0.937*** 0.911*** 0.940*** 0.914*** (17.19) (15.37) (16.7) (15.01) hip -0.154-0.18-0.153-0.181 (-1.) (-1.4) (-1.1) (-1.41) thigh 0.77* 0.55* 0.77* 0.55* (.43) (.1) (.4) (.0) hgt -0.0983-0.098 (-1.13) (-1.1) ankle 0.0477 0.0459 (0.4) (0.3) _cons -41.80*** -3.33** -4.73*** -33.4** (-6.45) (-3.05) (-5.65) (-.93) -------------------------------------------------------------- N 5 5 5 5 -sq 0.753 0.767 0.754 0.768 adj. -sq 0.708 0.71 0.7198 0.701 rmse 4.095 4.093 4.103 4.101 -------------------------------------------------------------- t statistics in parentheses * p<0.05, ** p<0.01, *** p<0.001 tstats and t tests: (1) to () (add hgt): hgt tstat > 1, and adj increase, MSE decreases (1) to (3) (add ankle): ankle tstat < 1, increases, adj decreases, and MSE increases F stats and F tests: (4) to (1) (drop hgt and ankle): Since adj decreases, the F test associated with dropping hgt and ankle from (4) will have an Fstat<1 here are the F test results:. reg Brozek wgt abd hip thigh hgt ankle. test hgt ankle ( 1) hgt = 0 ( ) ankle = 0 F(, 45) = 0.66 Prob > F = 0.5180 13