EDF 7405 Advanced Quantitative Methods in Educational Research. Data are available on IQ of the child and seven potential predictors.

Similar documents
EDF 7405 Advanced Quantitative Methods in Educational Research MULTR.SAS

Advanced Quantitative Data Analysis

Regression. Notes. Page 1. Output Created Comments 25-JAN :29:55

Investigating Models with Two or Three Categories

Correlations. Notes. Output Created Comments 04-OCT :34:52

*************NO YOGA!!!!!!!************************************.

MULTINOMIAL LOGISTIC REGRESSION

Multiple linear regression S6

Topic 1. Definitions

Class Notes: Week 8. Probit versus Logit Link Functions and Count Data

Chapter 9 - Correlation and Regression

36-309/749 Experimental Design for Behavioral and Social Sciences. Sep. 22, 2015 Lecture 4: Linear Regression

In Class Review Exercises Vartanian: SW 540

Regression: Main Ideas Setting: Quantitative outcome with a quantitative explanatory variable. Example, cont.

4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES

Review of Multiple Regression

Using SPSS for One Way Analysis of Variance

Chapter 19: Logistic regression

Sociology 593 Exam 2 March 28, 2002

2/26/2017. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

SPSS Output. ANOVA a b Residual Coefficients a Standardized Coefficients

Logistic Regression. Continued Psy 524 Ainsworth

Research Design - - Topic 19 Multiple regression: Applications 2009 R.C. Gardner, Ph.D.

2 Prediction and Analysis of Variance

Dependent Variable Q83: Attended meetings of your town or city council (0=no, 1=yes)

Chapter 19: Logistic regression

Basic Medical Statistics Course

One-Way ANOVA. Some examples of when ANOVA would be appropriate include:

Sociology 593 Exam 2 Answer Key March 28, 2002

Simple Linear Regression: One Quantitative IV

ESP 178 Applied Research Methods. 2/23: Quantitative Analysis

SPSS LAB FILE 1

Practical Biostatistics

7. Assumes that there is little or no multicollinearity (however, SPSS will not assess this in the [binary] Logistic Regression procedure).

LAB 3 INSTRUCTIONS SIMPLE LINEAR REGRESSION

Multiple Regression. More Hypothesis Testing. More Hypothesis Testing The big question: What we really want to know: What we actually know: We know:

Daniel Boduszek University of Huddersfield

Analysis of Covariance. The following example illustrates a case where the covariate is affected by the treatments.

SOS3003 Applied data analysis for social science Lecture note Erling Berge Department of sociology and political science NTNU.

Interactions and Centering in Regression: MRC09 Salaries for graduate faculty in psychology

Models for Binary Outcomes

x3,..., Multiple Regression β q α, β 1, β 2, β 3,..., β q in the model can all be estimated by least square estimators

Chapter Goals. To understand the methods for displaying and describing relationship among variables. Formulate Theories.

1 Correlation and Inference from Regression

Area1 Scaled Score (NAPLEX) .535 ** **.000 N. Sig. (2-tailed)

1. BINARY LOGISTIC REGRESSION

CHAPTER 10. Regression and Correlation

STAT 350 Final (new Material) Review Problems Key Spring 2016

Simple Linear Regression: One Qualitative IV

Lecture 12: Effect modification, and confounding in logistic regression

Prepared by: Prof. Dr Bahaman Abu Samah Department of Professional Development and Continuing Education Faculty of Educational Studies Universiti

McGill University. Faculty of Science MATH 204 PRINCIPLES OF STATISTICS II. Final Examination

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

Introduction to Regression

Model Estimation Example

PSY 216. Assignment 9 Answers. Under what circumstances is a t statistic used instead of a z-score for a hypothesis test

Three Factor Completely Randomized Design with One Continuous Factor: Using SPSS GLM UNIVARIATE R. C. Gardner Department of Psychology

Independent Samples ANOVA

Chapter 15 - Multiple Regression

Step 2: Select Analyze, Mixed Models, and Linear.

Sociology Research Statistics I Final Exam Answer Key December 15, 1993

STA6938-Logistic Regression Model

Multiple linear regression

Equation Number 1 Dependent Variable.. Y W's Childbearing expectations

A Re-Introduction to General Linear Models (GLM)

1 A Review of Correlation and Regression

Tests for the Odds Ratio in Logistic Regression with One Binary X (Wald Test)

LOGISTICS REGRESSION FOR SAMPLE SURVEYS

UNIVERSITY OF TORONTO. Faculty of Arts and Science APRIL 2010 EXAMINATIONS STA 303 H1S / STA 1002 HS. Duration - 3 hours. Aids Allowed: Calculator

Logistic Regression Analysis

Binary Logistic Regression

STA 431s17 Assignment Eight 1

One-Way ANOVA Source Table J - 1 SS B / J - 1 MS B /MS W. Pairwise Post-Hoc Comparisons of Means

Chapter 4 Regression with Categorical Predictor Variables Page 1. Overview of regression with categorical predictors

y response variable x 1, x 2,, x k -- a set of explanatory variables

Simple, Marginal, and Interaction Effects in General Linear Models: Part 1

(Where does Ch. 7 on comparing 2 means or 2 proportions fit into this?)

Regression ( Kemampuan Individu, Lingkungan kerja dan Motivasi)

Designing Multilevel Models Using SPSS 11.5 Mixed Model. John Painter, Ph.D.

More Accurately Analyze Complex Relationships

Chapter 15 - Multiple Regression

Retrieve and Open the Data

Longitudinal Modeling with Logistic Regression

" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2

NATIONAL UNIVERSITY OF SINGAPORE EXAMINATION. ST3241 Categorical Data Analysis. (Semester II: ) April/May, 2011 Time Allowed : 2 Hours

Analysis of Covariance (ANCOVA) with Two Groups

Inferences for Regression

EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7

Correlation and simple linear regression S5

SPSS Guide For MMI 409

Correlation and regression

STA 303 H1S / 1002 HS Winter 2011 Test March 7, ab 1cde 2abcde 2fghij 3

Simple logistic regression

Model Based Statistics in Biology. Part V. The Generalized Linear Model. Chapter 18.1 Logistic Regression (Dose - Response)

Assoc.Prof.Dr. Wolfgang Feilmayr Multivariate Methods in Regional Science: Regression and Correlation Analysis REGRESSION ANALYSIS

Utilization of Addictions Services

Simple Linear Regression

Module 8: Linear Regression. The Applied Research Center

Contrasts (in general)

Multiple Linear Regression II. Lecture 8. Overview. Readings

Transcription:

EDF 7405 Advanced Quantitative Methods in Educational Research Data are available on IQ of the child and seven potential predictors. Four are medical variables available at the birth of the child: Birthweight (BW) the weight of the new born in grams; very large or very small weights may indicate health problems ( X ). APGAR a quick assessment of overall newborn well being. Low numbers may indicate problems ( X ). Intrapartum factors (IF) a measure of the quality of the delivery. Low numbers indicate problems occurred in the delivery ( X 3 ). Neonatal factors (NF) a measure of the health of the newborn. Low numbers may indicate health problems ( X 4 ). Three are sociological variables measured at the time of the baby s birth. Socioeconomic status a measure of the material resources available to the mother. Higher scores indicate more resources ( X 5 ). Social support a measure of the social resources available to the mother. Higher scores indicate more resources ( X 6 ). Stressful life events (SLE) a measure of the extent of stressors in the mother s life in the prior year. Higher scores indicate more stressors ( X 7 ) We want to consider the medical variables as control variables and test whether or not the sociological variables are related to IQ when the medical variables are controlled. To do this we need results for two models:

The full model: Y X X X X X X X 3 3 4 4 5 5 6 6 7 7 The reduced model: Y X X X X 3 3 4 4 In terms of the abbreviations for the variables these models are The full model: Y BW APGAR IF NF SES SS SLE 3 4 5 6 7 The reduced model: Y BW APGAR IF NF 3 4 I present the regression results only. The residual plots for the full model should be examined before using the results for the models. Full Results See pages 8-9 of the first section for directions to conduct a regression analysis. Regression Variables Entered/Removed b Variables Entered SLE, IF, Variables Removed Method BW, SES,. Enter APGAR, SS, NF a a. All requested variables entered. b. Dependent Variable: IQ

3 Summary Std. Error Adjusted of the R R Square R Square Estimate.60 a.36.3 6.3 a. Predictors: (Constant), SLE, IF, BW, SES, APGAR, SS, NF ANOVA b Sum of Squares df Mean Square F Sig. Regression 3380.436 7 9.49 7.344.000 a Residual 3686.069 9 60.86 Total 37066.505 98 a. Predictors: (Constant), SLE, IF, BW, SES, APGAR, SS, NF b. Dependent Variable: IQ Unstandardized Coefficients Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig. (Constant) 33.085 6.709.980.05 BW 5.950E-03.00.74.569.0 APGAR.680.5.43.38.70 IF -.48.86 -.070 -.795.49 NF 8.393E-0.64.06.53.609 SES.55.490.450 4.607.000 SS -.08.366 -.06 -.570.570 SLE.68.57.097.04.30 a. Dependent Variable: IQ Reduced Results

4 See pages 8-9 of the first section for directions to conduct a regression analysis. Regression Variables Entered/Removed b Variables Entered NF, IF, Variables Removed Method APGAR, BW a. Enter a. All requested variables entered. b. Dependent Variable: IQ Summary Std. Error R R Square Adjusted R Square of the Estimate.380 a.44.08 8.37 a. Predictors: (Constant), NF, IF, APGAR, BW ANOVA b Sum of Squares df Mean Square F Sig. Regression 5347.70 4 336.793 3.96.005 a Residual 379.335 94 337.440 Total 37066.505 98 a. Predictors: (Constant), NF, IF, APGAR, BW b. Dependent Variable: IQ Unstandardized Coefficients Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig. (Constant) 55.69 5.559 3.579.00 BW 5.979E-03.003.76.366.00 APGAR.775.35.36.053.043 IF -.9.0 -.090 -.908.366 NF.5.80.093.695.489 a. Dependent Variable: IQ

5 EDF 7405 Advanced Quantitative Methods in Educational Research SEQUENT.SAS In this handout a procedure for obtaining control-preceding variables tests is presented. Here is the SPSS Windows editor with part of the data displayed. To conduct the analysis press analyze, regression and linear

You get the following screen: 6

7 Move iq into the Dependent slot by highlighting iq and pressing the arrow to the left of the Dependent slot and bw into the Independent(s) slot by highlighting bw and pressing the arrow to the left of the independents slot. You get Now press next. You get

8 Notice that Block of has changed to Block of and the Independent(s) slot is empty. Move SES into the Independent(s) slot. You get Press next. You get

9 Notice that Block of has changed to Block 3 of 3 and the Independent(s) slot is empty. Move MAGE in the Independent(s) slot. Continue in this fashion until you have moved all independent variables into the Independent(s) and in the appropriate order. The final screen looks like this

0 Block 8 of 8 indicates that eight independent variables have been moved to the Independent(s) slot. Now click Statistics. The screen looks like this Unclick estimates and click R squared change. Then click continue and when the new screen opens, click OK. These results follow over several pages. Regression [DataSet] C:\7405\Spss.Programs\Third\sequent.sav Variables Entered/Removed b Variables Entered Variables Removed Method bw a. Enter ses a. Enter 3 mage a. Enter 4 apgar a. Enter 5 nf a. Enter 6 if a. Enter 7 ss a. Enter 8 sle a. Enter a. All requested variables entered. b. Dependent Variable: iq

In the following you will find the up to j R X in the R square column, the Type I squared semi partial correlation coefficients in the R square change column, the control-preceding-variables F statistic in the F change column and the control-precedingvariables p value in the Sig F change column. Summary 3 4 5 6 7 8 Change Statistics Adjusted Std. Error of R Square R R Square R Square the Estimate Change F Change df df Sig. F Change.30 a.09.08 8.636.09 9.76 97.00.575 b.330.36 6.08.39 34.5 96.000.576 c.33.3 6.45.00.5 95.67.589 d.347.39 6.044.05.04 94.4.59 e.349.34 6.09.00.47 93.60.595 f.354.3 6.3.005.730 9.395.595 g.354.305 6.8.000.04 9.877.60 h.36.306 6.07.008.4 90.9 a. Predictors: (Constant), bw b. Predictors: (Constant), bw, ses c. Predictors: (Constant), bw, ses, mage d. Predictors: (Constant), bw, ses, mage, apgar e. Predictors: (Constant), bw, ses, mage, apgar, nf f. Predictors: (Constant), bw, ses, mage, apgar, nf, if g. Predictors: (Constant), bw, ses, mage, apgar, nf, if, ss h. Predictors: (Constant), bw, ses, mage, apgar, nf, if, ss, sle

The following results contain the F statistics for testing the omnibus multivariable hypothesis for each successive model. We do not use these results. 3 4 5 6 7 8 Regression Residual Total Regression Residual Total Regression Residual Total Regression Residual Total Regression Residual Total Regression Residual Total Regression Residual Total Regression Residual Total a. Predictors: (Constant), bw ANOVA i Sum of Squares df Mean Square F Sig. 3377.765 3377.765 9.76.00 a 33688.740 97 347.307 37066.505 98 36.756 68.378 3.656.000 b 489.749 96 58.643 37066.505 98 30.46 3 400.809 5.73.000 c 4764.079 95 60.675 37066.505 98 869.7 4 37.48.499.000 d 496.794 94 57.43 37066.505 98 933.774 5 586.755 9.969.000 e 43.73 93 59.49 37066.505 98 33.68 6 87.7 8.405.000 f 394.877 9 60.49 37066.505 98 39.957 7 875.708 7.3.000 g 3936.548 9 63.039 37066.505 98 345.98 8 678.6 6.389.000 h 364.07 90 6.680 37066.505 98 b. Predictors: (Constant), bw, ses c. Predictors: (Constant), bw, ses, mage d. Predictors: (Constant), bw, ses, mage, apgar e. Predictors: (Constant), bw, ses, mage, apgar, nf f. Predictors: (Constant), bw, ses, mage, apgar, nf, if g. Predictors: (Constant), bw, ses, mage, apgar, nf, if, ss h. Predictors: (Constant), bw, ses, mage, apgar, nf, if, ss, sle i. Dependent Variable: iq

3 EDF 7405 Advanced Quantitative Methods in Educational Research Fathers are randomly assigned to either an experimental or a control group. The experimental treatment is designed to increase fathers' authoritative towards child rearing. The fathers were post tested with the Eversoll Father Role Questionnaire. In the data the control group is coded Z for the experimental group Z 0 for the control group This kind of coding is called dummy coding. The variable is denoted by a Z to emphasize that it is a categorical variable rather than a quantitative variable.

4 The Data Group EFRQ 0 97 0 0 0 94 0 0 85 0 08 0 07 0 0 86 0 03 0 9 0 0 09 97 95 03 04 8 07 5 94 00 90 08 The model is Y Z or in terms of the abbreviations for the variables EFRQ Group The symbol is used in place of to emphasize that it is a regression coefficient for a categorical variable rather than for a quantitative variable.

5 The same steps that were used in simple regression analyses in the first section of the course are used again (see pages 8-9 of the first section to conduct a regression analysis). However before doing the regression analysis we must compute descriptive statistics for each group. Descriptive Statistics by Treatment Group

See pages 0- of the first section for directions to obtain descriptive statistics. 6

7 Descriptives GROUP = 0 Descriptive Statistics a EFRQ Valid N (listwise) a. GROUP = 0 Std. N Minimum Maximum Mean Deviation 3 88 0 03.85 9.69 3 GROUP = Descriptive Statistics a EFRQ Valid N (listwise) a. GROUP = Std. N Minimum Maximum Mean Deviation 3 03 8.46 8.75 3 Before conducting the regression analysis we must turn off the split-file option. To do this follow the steps used to turn it on until you get to the screen Select Analyze all cases, do not create groups and then click OK.

8 Regression Analysis Regression Variables Entered/Removed b Variables Entered Variables Removed Method GROUP a. Enter a. All requested variables entered. b. Dependent Variable: EFRQ Summary Std. Error R R Square Adjusted R Square of the Estimate.437 a.9.57 9.4 a. Predictors: (Constant), GROUP ANOVA b Sum of Squares df Mean Square F Sig. Regression 48.46 48.46 5.657.06 a Residual 046.93 4 85.88 Total 59.385 5 a. Predictors: (Constant), GROUP b. Dependent Variable: EFRQ Unstandardized Coefficients Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig. (Constant) 03.846.56 40.543.000 GROUP 8.65 3.6.437.378.06 a. Dependent Variable: EFRQ

9 EDF 7405 Advanced Quantitative Methods in Educational Research QUAL.SAS Data are available on 5 teenage mothers and their children. The data consist of dummy variables indicating the prenatal care program in which the mother took part and mental development index (MDI) scores derived from the Bayley scales of infant development. The prenatal care was delivered by the teenage pregnancy team, private physicians, or the Shands high risk clinic. The coding of the groups is presented in the following table Prenatal Care Z Z Teenage Pregnancy Team 0 Private Physician 0 Shands High Risk Clinic 0 0 The MDI scores were obtained at age six months. The data are used to illustrate the use of regression to conduct a one-way between-subjects ANOVA. Descriptive Statistics for Each Group To get descriptive statistics on the groups, before using the Descriptive Statistics option within the Analyze option it is necessary to split the file. To review, splitting the file is done by using the Split File option within the Data option in the SPSS for Windows Data Editor.

0

Descriptives Z = 0, Z = 0 Descriptive Statistics a MDI Valid N (listwise) a. Z = 0, Z = 0 Std. N Minimum Maximum Mean Deviation 3 9 5 8.38.3 3 Z = 0, Z = Descriptive Statistics a MDI Valid N (listwise) a. Z = 0, Z = Std. N Minimum Maximum Mean Deviation 9 88 55 34.37 6.55 9 Z =, Z = 0

Descriptive Statistics a MDI Valid N (listwise) a. Z =, Z = 0 Std. N Minimum Maximum Mean Deviation 9 94 44 0.84 4.74 9 Before running the regression analysis the file must be unsplit. Just select Analyze all cases, do not select groups in the following screen: Regression Analysis The following is the result of running the SPSS regression analysis (see pages 8-9 of the first section to conduct a regression analysis). It provided the test of the omnibus hypothesis and of the comparison of Teenage Pregnancy Team to Shands High Risk Clinic and of Private Physicians to Shands High Risk Clinic. A method for obtaining the comparison of Teenage Pregnancy Team and Private Physicians is presented after the initial regression results.

3 Regression Variables Entered/Removed b Variables Variables Entered Removed Method Z, Z a. Enter a. All requested variables entered. b. Dependent Variable: MDI Summary Std. Error Adjusted of the R R Square R Square Estimate.39 a.53.8 7.0 a. Predictors: (Constant), Z, Z Regression Residual Total a. Predictors: (Constant), Z, Z b. Dependent Variable: MDI ANOVA b Sum of Mean Squares df Square F Sig. 56.66 80.83 4.39.09 a 40.04 48 95.876 6763.686 50 (Constant) Z Z a. Dependent Variable: MDI Unstandardized Coefficients Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig. 8.385 4.77 4.85.000.457 6.9.066.397.693 5.984 6.9.46.58.03

4 Unfortunately the SPSS regression program does not automatically produce a t statistic for comparing Teenage Pregnancy Team and Private Physicians. The t statistic can be computed by using the SPSS line code. In the following screen for the regression program press Paste: The SPSS syntax editor displays the following code: REGRESSION /MISSING LISTWISE /STATISTICS COEFF OUTS R ANOVA /CRITERIA=PIN(.05) POUT(.0) /NOORIGIN /DEPENDENT mdi /METHOD=ENTER z z Add BCOV to the statistics line: REGRESSION /MISSING LISTWISE /STATISTICS COEFF OUTS R ANOVA BCOV /CRITERIA=PIN(.05) POUT(.0) /NOORIGIN

5 /DEPENDENT mdi /METHOD=ENTER z z Press Run and then press All. The results will be the same as for the new full model with the exception that the following results will be added at the end. Coefficient Correlations(a) z z Correlations z.000.594 z.594.000 Covariances z 38.33.760 z.760 38.33 a Dependent Variable: mdi These results are confusingly labeled because they are not correlation coefficients and covariances for the variables. Rather they are sampling correlations and covariances. The t statistic for H0 : 0 is t S S C The results required for the denominator are Name of Statistic Sampling variance for Sampling variance for Sampling covariance for and Symbol S 38.33 S 38.33 C.760 Numeric value Each sampling variance is the square of a standard error, a term with which you are more familiar. The coefficients and can be found in the results labeled coefficients. These results are on page 3. Substituting in the t statistic for testing H 0 : 0

6 t S S C yields.457 5.984 38.3 38..760 3.57.4 3.4

7 EDF 7405 Advanced Quantitative Methods in Educational Research ANCORAN.SAS In the investigation of the treatment designed to increase fathers' authoritative towards child rearing, fathers were pre and post tested with the Eversoll Father Role Questionnaire. This handout illustrates how to incorporate the pretest and test for the covariate x treatment interaction. The groups are again coded The data Z for the experimental group Z 0 for the control group Group EFPOST EFPRE 0 00 97 0 96 0 0 96 94 0 3 0 9 85 0 08 0 3 07 0 0 0 88 86 0 07 03 0 99 9 0 0 0 03 3 09 8 97 95 03 03 03 04 4 8 6 07 08 5 07 94 3 00 04 90 8 08

8 for testing the interaction: The model is Y X XZ or in terms of the abbreviations of the variables EFPOST ( EFPRE) Group ( Group EFPRE). The symbol emphasizes it is a regression coefficient for the product term. We must use the TRANSFORM and COMPUTE option to add the product (GROUP EFPRE) to the data set (see pages 37-39 of the first section to use the compute feature). The product is denoted by PR is the SPSS results. The following shows the resulting data See pages 6-9 of the first section for directions to produce case summaries. Summarize Case Processing Summary a Cases Included Excluded Total N Percent N Percent N Percent EFPOST 6 00.0% 0.0% 6 00.0% EFPRE 6 00.0% 0.0% 6 00.0% PR 6 00.0% 0.0% 6 00.0% a. Limited to first 00 cases.

9 Case Summaries a EFPOST EFPRE PR 03 3 09 09 3 8 97 97 4 95 95 5 03 03 03 6 03 04 04 7 4 8 8 8 6 07 07 9 08 5 5 0 07 94 94 3 00 00 04 90 90 3 8 08 08 4 00 97 0 5 96 0 6 0 7 96 94 0 8 3 0 9 9 85 0 0 08 0 3 07 0 0 0 3 88 86 0 4 07 03 0 5 99 9 0 6 0 0 0 Total N 6 6 6 a. Limited to first 00 cases.

30 See pages 8-9 of the first section for directions to conduct a regression analysis. Regression Variables Entered/Removed b Variables Entered PR, Variables Removed Method EFPRE, GROUP a. Enter a. All requested variables entered. b. Dependent Variable: EFPOST Summary Std. Error R R Square Adjusted R Square of the Estimate.70 a.505.437 7.55 a. Predictors: (Constant), PR, EFPRE, GROUP ANOVA b Sum of Squares df Mean Square F Sig. Regression 76.88 3 45.396 7.468.00 a Residual 53.97 56.964 Total 59.385 5 a. Predictors: (Constant), PR, EFPRE, GROUP b. Dependent Variable: EFPOST Unstandardized Coefficients Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig. (Constant) 30.876 0.36.56.44 GROUP 55.645 33.55.8.660. EFPRE.709.97.684 3.603.00 PR -.460.3 -.438 -.46.68 a. Dependent Variable: EFPOST

3 The evidence to this point indicates lack of support for a covariate x treatment interaction. I next show how to test the treatment effect, under the assumption that there is no covariate x treatment interaction. The groups are again coded for testing the treatment effect The model is Z for the experimental group Z 0 for the control group Y X or in terms of the abbreviations of the variables EFRQ _ POST ( EFRQ _ PRE) Group. Note that the product term has been removed. See pages 8-9 of the first section for directions to conduct a regression analysis. Regression Variables Entered/Removed b Variables Entered Variables Removed Method EFPRE, GROUP a. Enter a. All requested variables entered. b. Dependent Variable: EFPOST Summary Std. Error R R Square Adjusted R Square of the Estimate.677 a.459.4 7.7 a. Predictors: (Constant), EFPRE, GROUP

3 Regression Residual Total ANOVA b Sum of Mean Squares df Square F Sig. 60.408 580.04 9.748.00 a 368.976 3 59.5 59.385 5 a. Predictors: (Constant), EFPRE, GROUP b. Dependent Variable: EFPOST (Constant) GROUP EFPRE Unstandardized Coefficients a. Dependent Variable: EFPOST Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig. 48.505 6.537.933.007 8.036 3.03.407.65.04.538.59.59 3.375.003

33 EDF 7405 Advanced Quantitative Methods in Educational Research ANCORAN.SAS Adult children of alcoholic parents were tested using a healthy coping behaviors scale, randomly assigned to one of three treatments:. Cognitive treatment. Experiential treatment 3. Control treatment and then tested again using the healthy coping behaviors scale. Our first purpose is to test for a covariate x treatment interaction. That is, we want to determine if there is evidence for the claim that the direction and/or size of the treatment effects depends on degree of healthy coping behavior as measured by the pretest. The following are the data Treatment GROUP PRETEST POSTTEST 7 38 46 5 40 6 50 6 54 33 43 ` 3 3 33 45 36 38 47 50 3 8 44 49 7 39 3 5 35 40 46 5 37 6 3 35 38 3 45 4

34 for testing the interaction: 3 8 9 3 49 43 3 4 39 3 3 3 3 44 43 3 9 0 3 6 3 3 3 34 3 5 5 3 9 8 3 4 The model that includes the interaction (product) terms is Y X Z Z XZ XZ or in terms of the abbreviations of the variables Post Pre Z Z Pre Z Pre Z () We must use the TRANSFORM and COMPUTE option to add the product terms to the data set (see pages 37-39 of the first section to use the compute feature). These products are denoted by PR and PR in the SPSS results. The hypothesis we want to test is H : 0. 0 This is a hypothesis on a subset of the coefficients and requires us to use a full model and a reduced model. The full model is equation (). The reduced model is obtained by eliminating the coefficients in the hypothesis Post Pre Z Z. () The following are the results for the full model. See pages 8-9 of the first section for directions to conduct a regression analysis.from the model summary section we find R FM.546.

35 Regression Variables Entered/Removed b Variables Entered PR, PRE, Variables Removed Method Z, PR, Z a. Enter a. All requested variables entered. b. Dependent Variable: POST Summary b R R Square Adjusted R Square Std. Error of the Estimate.739 a.546.458 8.49 a. Predictors: (Constant), PR, PRE, Z, PR, Z b. Dependent Variable: POST ANOVA b Sum of Squares df Mean Square F Sig. Regression 7.860 5 443.57 6.43.00 a Residual 847.358 6 7.05 Total 4065.9 3 a. Predictors: (Constant), PR, PRE, Z, PR, Z b. Dependent Variable: POST Unstandardized Coefficients Coefficients a Standardized Coefficients B Std. Error Beta t Sig. (Constant) 4.566 8.79.55.586 Z 5.947.690.093.0.035 Z -.953 4.46 -.037 -.066.948 PRE.839.46.70 3.406.00 PR -.449.360 -.60 -.47.3 PR.4.45.88.334.74 a. Dependent Variable: POST

36 Residuals Statistics a Minimum Maximum Mean Std. Deviation N Predicted Value 0.5 48.85 36.34 8.458 3 Residual -3.93 7.96.00 7.70 3 Std. Predicted Value -.87.478.000.000 3 Std. Residual -.838.3.000.96 3 a. Dependent Variable: POST The following are the results for the reduced model. See pages 8-9 of the first section for directions to conduct a regression analysis. From the model summary section we find R RM.504. Regression Variables Entered/Removed b Variables Entered Variables Removed Method PRE, Z, Z a. Enter a. All requested variables entered. b. Dependent Variable: POST Summary b R R Square Adjusted R Square Std. Error of the Estimate.70 a.504.45 8.485 a. Predictors: (Constant), PRE, Z, Z b. Dependent Variable: POST ANOVA b Sum of Squares df Mean Square F Sig. Regression 049.336 3 683. 9.488.000 a Residual 05.883 8 7.996 Total 4065.9 3 a. Predictors: (Constant), PRE, Z, Z b. Dependent Variable: POST

37 (Constant) Z Z PRE a. Dependent Variable: POST Unstandardized Coefficients Coefficients a Standardized Coefficients B Std. Error Beta t Sig. 8.937 5.685.57.7.00 3.496.54 3.490.00 3.843 3.85.48.007.3.704.6.589 4.383.000 Residuals Statistics a Minimum Maximum Mean Std. Deviation N Predicted Value.3 54. 36.34 8.3 3 Residual -.8 7.60.00 8.064 3 Std. Predicted Value -.77.97.000.000 3 Std. Residual -.689.075.000.950 3 a. Dependent Variable: POST To test a subset hypothesis H : 0. 0 we use the test statistic F R FM R RM nk k g R FM. In the current example we get F 3 5.546.504.0 5 3.546 The critical value is F,, F.,,,6 3.37 and we fail to reject the null kg nk or hypothesis. Had we rejected the null hypothesis, we would have been interested in plotting the regression lines for the three groups. To do so, on the following screen

press Save to obtain 38

39 and select the unstandardized predicted values and then plot these against the pretest. See pages -3 of the first section for directions to construct a scatter plot. The following is the plot with annotations added:

40 Plot of Predicted Postest Score vs. Pretest 50 Score for with Product Terms Cognitive Predicted Posttest 40 30 Experiential Control 0 0 0 30 40 50 Pretest Since the null hypothesis was not rejected, we conclude that the following model is adequate for the data Post Pre Z Z. (3) This becomes our new full model. The hypothesis of interest is H : 0 0 Since this a hypothesis on a subset of the parameters we need a new reduced mode: Post Pre (4) We already have results for the new full model since it was out old reduced model. For this model, R FM.504. The following are the results for the new reduced model. From the model summary section we find R RM.85.

4 Regression Variables Entered/Removed b Variables Entered Variables Removed Method PRE a. Enter a. All requested variables entered. b. Dependent Variable: POST Summary b R R Square Adjusted R Square Std. Error of the Estimate.533 a.85.6 9.847 a. Predictors: (Constant), PRE b. Dependent Variable: POST ANOVA b Sum of Squares df Mean Square F Sig. Regression 56.580 56.580.99.00 a Residual 908.638 30 96.955 Total 4065.9 3 a. Predictors: (Constant), PRE b. Dependent Variable: POST Unstandardized Coefficients Coefficients a Standardized Coefficients B Std. Error Beta t Sig. (Constant) 6.8 6.09.656.03 PRE.638.85.533 3.454.00 a. Dependent Variable: POST Residuals Statistics a Minimum Maximum Mean Std. Deviation N Predicted Value 6.38 47.4 36.34 6.08 3 Residual -3.77 3.6.00 9.686 3 Std. Predicted Value -.63.84.000.000 3 Std. Residual -.44.399.000.984 3 a. Dependent Variable: POST

4 To test the subset hypothesis H : 0 0 we use the test statistic F R FM R RM nk k g R FM. In the current example we get F 3 3.504.85 6.8 3.504 The critical value is F,, F.,,,8 3.34 and we reject the null hypothesis. kg nk or Since we have rejected the null hypothesis H : 0 0 we want to test specific hypotheses to determine which pairs of treatments had different vertical separation between the regression lines. That is we want to test Groups Compared Cognitive vs. Control Experiential vs. Control Cognitive vs. Experiential Hypothesis H0 : 0 H0 : 0 H : 0 0 The t statistics for the first two hypotheses can be found in the coefficients section of the printout for the new full model (which was the original reduced model). These are shown in the following table:

43 Groups Compared Cognitive vs. Control Experiential vs. Control Cognitive vs. Experiential Hypothesis H0 : 0 3.490 H0 : 0.007 H : 0 0 t Unfortunately the SPSS regression program does not automatically produce a t statistic for a third hypothesis. The t statistic can be computed by using the SPSS line code. In the following screen for the regression program press Paste:

44 The SPSS syntax editor displays the following code REGRESSION /MISSING LISTWISE /STATISTICS COEFF OUTS R ANOVA /CRITERIA=PIN(.05) POUT(.0) /NOORIGIN /DEPENDENT post /METHOD=ENTER pre z z /SAVE PRED. Add BCOV to the statistics line: REGRESSION /MISSING LISTWISE /STATISTICS COEFF OUTS R ANOVA BCOV /CRITERIA=PIN(.05) POUT(.0) /NOORIGIN /DEPENDENT post /METHOD=ENTER pre z z Press Run and then press All. The results will be the same as for the new full model with the exception that the following results will be added at the end. Correlations Covariances Coefficient Correlations a Z PRE Z Z PRE Z a. Dependent Variable: POST Z PRE Z.000 -.03.4 -.03.000.07.4.07.000 4.553 -.98E-0 5.49 -.98E-0.578E-0 5.984E-0 5.49 5.984E-0. These results are confusingly labeled because they are not correlation coefficients and covariances for the variables. Rather they are sampling correlations and covariances. The t statistic for H0 : 0 is t S S C

45 The results required for the denominator are Name of Statistic Sampling variance for Sampling variance for Sampling covariance for and Symbol S 4.553 S. C 5.49 Numeric value Each sampling variance is the square of a standard error, a term with which you are more familiar. The coefficients and can be found in the results labeled coefficients. These results are on page 4. Substituting in the t statistic for testing H 0 : 0 t S S C yields.00 3.843 4.553. 5.49 8.357 5.79.0 We will use the Bonferroni critical value in order to control the family wise error rate: t /,, C, nk t.05/,3,33.4

46 Collecting all of the results we have Groups Compared Cognitive vs. Control Experiential vs. Control Cognitive vs. Experiential Hypothesis t t /,, C, n k Decison H0 : 0 3.490.4 Reject H0 : 0.007.4 Fail to reject H : 0.0.4 Fail to reject 0 and we conclude that there is a treatment difference between Cognitive and Control, but we do not have sufficient evidence to conclude that there is a treatment difference between Experiential and Control or between Cognitive and Experiential. The regression lines in the following plot are consistent with these results. Plot of Predicted Postest Score vs. Pretest 60 Score for without Product Terms Predicted Posttest 50 40 30 Experiential Cognitive Control 0 0 0 30 40 50 Pretest

47 EDF 7405 Advanced Quantitative Methods in Educational Research STEPWISE.SAS In this handout I show how to use SPSS to do stepwise regression. The data are from a textbook example. The textbook did not include a context. The data Y X X X 3.5 7.0.7 5.7.4 6.8.0 5.0 9.7.7. 3.8.4 3.8. 4.7 0.7 3.8 3.3.7.9 3.3 4. 3.0 0.6 3.3.6 4.3 0.7 3..5 3.5 0.5. 4.0.4.7 5..9 4. To do stepwise regression, follow the steps to conduct a regression analysis (see pages 8-9 of the first section) until you get to the following screen.

48 In addition to declaring the independent and dependent variables, you must change the method to stepwise by selecting it in the drop-down menu

49 To make the results agree with the results in SAS I have changed some of the options. It is not necessary to change the options in order to run a stepwise regression, but SAS and SPSS will give different results if you do not. I changed the entry probability to.5 and the removal probability to.5.

50 Thus a coefficient that is significant at the.5 alpha level will enter the model and a coefficient that is not significant at the.5 alpha level will be removed from the model. Regression

5 Variables Entered/Removed a 3 4 Variables Entered X. X. X3.. X a. Dependent Variable: Y Variables Removed Method Stepwise (Criteria: Probabilit y-of-f-to-e nter <=.50, Probabilit y-of-f-to-r emove >=.5). Stepwise (Criteria: Probabilit y-of-f-to-e nter <=.50, Probabilit y-of-f-to-r emove >=.5). Stepwise (Criteria: Probabilit y-of-f-to-e nter <=.50, Probabilit y-of-f-to-r emove >=.5). Stepwise (Criteria: Probabilit y-of-f-to-e nter <=.50, Probabilit y-of-f-to-r emove >=.5).

5 3 4 Summary Std. Error Adjusted of the R R Square R Square Estimate.607 a.369.90.89.757 b.573.45.70.869 c.756.634.588.845 d.74.63.589 a. Predictors: (Constant), X b. Predictors: (Constant), X, X c. Predictors: (Constant), X, X, X3 d. Predictors: (Constant), X, X3 3 4 Regression Residual Total Regression Residual Total Regression Residual Total Regression Residual Total a. Predictors: (Constant), X b. Predictors: (Constant), X, X ANOVA e Sum of Mean Squares df Square F Sig. 3.38 3.38 4.674.063 a 5.37 8.67 8.509 9 4.878.439 4.70.05 b 3.63 7.59 8.509 9 6.433 3.44 6.96.09 c.076 6.346 8.509 9 6.077 3.038 8.743.0 d.43 7.347 8.509 9 c. Predictors: (Constant), X, X, X3 d. Predictors: (Constant), X, X3 e. Dependent Variable: Y

53 3 4 (Constant) X (Constant) X X (Constant) X X X3 (Constant) X X3 a. Dependent Variable: Y Unstandardized Coefficients Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig. 9.873.67 4.7.000.33.53.607.6.063 7.69.39 5.788.00.467.54.855 3.037.09.599.37.56.83.0.736 3.0.576.585.85.83.339.04.350.554.54.337.967.05.44.540.37.0.078.63.644.099.94.797.467.546 3.847.006.54.37.667 4.49.004 4 X X3 X3 X Excluded Variables d Collinearit y Partial Statistics Beta In t Sig. Correlation Tolerance.56 a.83.0.569.769 -.5 a -.58.579 -.5.459.37 b.0.078.654.9.339 c.04.350.383.363 a. Predictors in the : (Constant), X b. Predictors in the : (Constant), X, X c. Predictors in the : (Constant), X, X3 d. Dependent Variable: Y

54

55 EDF 7405 Advanced Quantitative Methods in Educational Research SUBSET.SAS SPSS does not have an all possible subsets program.

56 EDF 7405 Advanced Quantitative Methods in Educational Research SUBCOMP.SAS SPSS does not have an all possible subsets program.

57 EDF 7405 Advanced Quantitative Methods in Educational Research This shows how to use SPSS to do a multicategory logistic regression. After importing the data into the SRSS Data Editor, click Analyze, Regression, Multinomial Logistic. For my data the result is: Move MATHGRP into the dependent slot because it is the variable indicating math group membership: = Advanced; = Regular; 3 = Remedial. Move SAS into the Covariates slot. Here covariate is being used as a synonym for quantitative independent variable. The results are

58 Nominal Regression MATHGRP Valid Missing Total Subpopulation Case Processing Summary 3 Marginal N Percentage 7550 38.0% 384 57.3% 944 4.7% 9878 00.0% 0 9878 50 a a. The dependent variable has only one value observed in 5 (0.0%) subpopulations. Fitting Information Intercept Only Final Fitting Criteria Likelihood Ratio Tests - Log Likelihood Chi-Square df Sig. 867.63 669.596 98.037.000 Pseudo R-Square Cox and Snell Nagelkerke McFadden.00.0.006 Effect Intercept SES Likelihood Ratio Tests Fitting Criteria Likelihood Ratio Tests - Log Likelihood of Reduced Chi-Square df Sig. 97.000 57.405.000 867.63 98.037.000 The chi-square statistic is the difference in - log-likelihoods between the final model and a reduced model. The reduced model is formed by omitting an effect from the final model. The null hypothesis is that all parameters of that effect are 0.

59 MATHGR a Intercep SES Intercep SES a. The reference category is: 3. Parameter Estimates B Std. Error Wald df Sig. Exp(B) Lower BoundUpper Bound -.38..59.6.047.004 6.685.000.048.039.057.46.07 36.348.000 5% Confidence Interval fo Exp(B).06.004 36.7.000.06.08.035

60

6 EDF 7405 Advanced Quantitative Methods in Educational Research This shows how to use SPSS to do a proportional odds logistic regression. After importing the data into the SRSS Data Editor, click Analyze, Regression, Ordinal. For my data the result is: Move MATHGRP into the dependent slot because it is the variable indicating math group membership: = Advanced; = Regular; 3 = Remedial. Move SAS into the Covariates slot. Here covariate is being used as a synonym for quantitative independent variable. The results are:

6 PLUM - Ordinal Regression Warnings There are 6 (0.7%) cells (i.e., dependent variable levels by combinations of predictor variable values) with zero frequencies. Case Processing Summary MATHGRP Valid Missing Total 3 Marginal N Percentage 7550 38.0% 384 57.3% 944 4.7% 9878 00.0% 0 9878 Fitting Information - Log Likelihood Chi-Square df Sig. Intercept Only 867.63 Final 674.796 9.837.000 Link function: Logit. Goodness-of-Fit Chi-Square df Sig. Pearson 43.499 97.000 Deviance 34.884 97.000 Link function: Logit. Pseudo R-Square Cox and Snell Nagelkerke McFadden Link function: Logit..00.0.006 Parameter Estimates 95% Confidence Interval Estimate Std. Error Wald df Sig. Lower BoundUpper Bound Threshold [MATHGRP = -.77.09 363.4.000 -.905 -.550 [MATHGRP =.783.093 367.73.000.60.966 Location SES -.05.00 9.89.000 -.08 -.0 Link function: Logit.