Density Temp vs Ratio. temp
|
|
- Jodie Jemimah Tucker
- 6 years ago
- Views:
Transcription
1 Temp Ratio Density Density (a) temp ratio The histogram shows that the temperature measures have two peaks, one around 172, and another higher peak around 182. The ratio measures seem slightly skewed, with more observations around 1 or 1.5, and fewer out in the tails around 2.5 or 3. Neither sets of measurements appear to have large outliers. (b) No, the value of the efficiency ratio is not completely and uniquely determined by tank temperature. If this were the case, then every tank temperature would correspond to a unique ratio measure. However, we can see that there are five temperature measure equal to 180, but they correspond to five different ratio measures (1.45, 1.60, 1.61, 2.13, 2.15). If the ratio were completely determined by temperature, the corresponding ratios for the 180 temp measures would all be the same. Temp vs Ratio ratio (c) temp The scatterplot of temperature vs ratio does appear to show an increasing and linear relationship between the two variables. The higher temperature values are generally associated with higher ratio values. It seems reasonable that temperature might predict ratio values. (d) Using statistical software, we can get the estimated regression line: Ŷ i = ˆβ 0 + ˆβ 1 temp i = temp i. The regression line shows that for every degree increase in temperature, the ratio increases less than 0.1 in the efficiency ratio. When the temperature is zero, the efficiency ratio is equal to 1
2 This is not directly interpretable, because the efficiency ratio cannot go below zero. However, since the smallest temperature value is 170 degrees, the intercept is calculated using this value as our baseline value. One way to fix this issue would be to recalibrate the temperature values, subtracting 170 from all of them so that 0 is a meaningful value. We have three assumptions we would like to test for the model: i. The relationship between temperature and efficiency ratio is linear. ii. The ɛ i are normally distributed iii. The ɛ i are normally distributed with the same variance ( Homoscedasticity ) Fitted vs Residuals Standardized residuals Residuals Frequency Fitted e_star The first assumption seems valid based on the scatterplot we created earlier, which shows a reasonable, increasing linear relationship between temperature and the efficiency ratio. The histogram of the standardized residuals e i seems to be normal, so the second assumption is reasonable. The graph of the fitted values vs the standardized residuals shows no pattern (which is what we want), and the values lie between -2 and 2. Because there is not pattern, it seems the homoscedastic assumption is valid. Code and output in R: #### MAKE THE MODEL #### > fit1 = lm(ratio ~ Temp, data = prob1data) # Predict ratio with temp using prob1 data > summary(fit1) Call: lm(formula = Ratio ~ Temp, data = prob1data) Residuals: Min 1Q Median 3Q Max Coefficients: Estimate Std. Error t value Pr(> t ) (Intercept) *** 2
3 Temp *** --- Signif. codes: 0 *** ** 0.01 * Residual standard error: on 22 degrees of freedom Multiple R-squared: ,Adjusted R-squared: F-statistic: 18.1 on 1 and 22 DF, p-value: ### MAKE THE PLOTS ### > e_star = (fit1$resid-mean(fit1$resid))/sd(fit1$resid) > par(mfrow = c(1,2)) > plot(fit1$fitted, e_star, main = Fitted vs Residuals, xlab = Fitted, ylab = Residuals ) > abline(h = 0) > hist(e_star, main = Standardized residuals ) (e) The regression line is used to predict the average efficiency ratio: EY x=182 = = When the temperature equals 182 degrees, the average efficiency ratio equals (f) The residuals for the four observations for which temperature equals 182 are: , , 0.034, and The reason that these do not all have the same sign is because they values do not all lie on the regression line. Some of them are above the regression line, and some are below it. This is due to the random variation in the observed Y i values. (g) The output in part (d) shows that R 2 is equal to , which means that 45% of the variation in the efficiency ratio is explained by temperature. 2. (a) The scatterplot shows that the simple linear regression model appears to be reasonable, as the relationship between SO 2 and steel weight loss seems linear. Steel weight loss S02 (b) Ŷ = SO 2 The estimated regression equation shows that for a 1 mg/m 2 /d increase in SO2, the steel weight loss goes up an average of 9.31 g/m 2. With the sodium chloride level is 0, then the average steel weight loss is g/m 2. 3
4 > summary(fit2) Call: lm(formula = y ~ x, data = prob2data) Residuals: Coefficients: Estimate Std. Error t value Pr(> t ) (Intercept) ** x e-05 *** --- Signif. codes: 0 *** ** 0.01 * Residual standard error: on 4 degrees of freedom Multiple R-squared: ,Adjusted R-squared: F-statistic: 385 on 1 and 4 DF, p-value: 3.978e-05 It is difficult to tell from the plots of the residuals if our assumptions hold since there are so few of them, but it does not appear there are any obvious violations. Fitted vs resid Histogram of estar2 Resid Frequency Fitted estar2 (c) The output above shows that R 2 equals , which means that almost 99% of the variation in steel weight loss can be attributed to SO 2. This value is extremely high, indicating that this is a great predictor. (d) Before this model is even created, we can guess that the slope of the regression line will not change much. We can guess this by looking at the scatterplot; even though the SO 2 measure is quite high, so is the steel weight value, so this point still fits along the regression line and is probably not that influential. Call: lm(formula = y ~ x, data = prob2data[-6, ]) Residuals: Coefficients: Estimate Std. Error t value Pr(> t ) (Intercept) * 4
5 x ** --- Signif. codes: 0 *** ** 0.01 * Residual standard error: on 3 degrees of freedom Multiple R-squared: ,Adjusted R-squared: F-statistic: on 1 and 3 DF, p-value: Results from the two models show that the parameter estimates for β 0 are quite different, with higher values for the original model. The estimates for the slope are slightly different, although the change is not as drastic. A plot of the fitted values from each model shows that they are fairly similar, although the relationship is not perfectly at 45 degrees. Fitted values New model Original model 3. (a) Ŷ = tannin ˆβ 1 = S xy = = ˆβ 0 = ȳ ˆβ 1 x =.549/ /32 = 1.58 Interpretation of the regression line: For every 1-unit increase in tannin concentration, the perceived astringency increases by units. If the tannin concentration is zero, then the perceived astringency is The figure does not show any violations of our assumptions, and the linear model seems to fit very nicely to the data. On the left, we see that there is a strong linear relationship between the tannin level and the perceived astringency. In the middle, there does not appear to be any trend among the residuals, and they are centered around 0, with values between -2 and 2. On the right, we see that the residuals appear to be normally distributed. (b) The confidence interval for β 1 can be calculated in two ways. i. Calculate MSE and use the equation: > MSE = sum(fit$resid^2/30) > qt(.975, 30)*sqrt(MSE/1.482) ˆβ 1 ± t α/2,n 2 MSE/Sxx 5
6 Tannin vs Astringency Fitted vs Residuals Hist of residuals Astringency Residuals Frequency Tannin Fitted estar [1] > qt(.975, 30)*sqrt(MSE/1.482) [1] ii. Use the output, which provides estimates of the slope as well as standard errors: Coefficients: Estimate Std. Error t value Pr(> t ) (Intercept) e-13 *** tannin e-13 *** --- > qt(.975, 30)*.2080 [1] > qt(.975, 30)*.2080 [1] The 95% confidence interval for the slope is (2.16, 3.01), which means that we are 95% confident that the true slope is in this range. (c) We use the estimated regression line to estimate the average astringency when the tannin concentration is 0.6: EY x=0.6 = = However, we want to express how reliable this estimate is, so we can calculate a 95% confidence interval for the average astringency when the tannin level is 0.6: [ ] 1 Ŷ x=x ± t α/2,n 2 MSE [ ] 1 ( )2 = ± = ( 0.128, 0.060) We are 95% confident that the true average perceived astringency level is between and when the tannin concentration is 0.6. (d) The prediction interval will be similar to the confidence interval, except that we are not making the interval for the average outcome, we are making it for the range of all possible outcomes, 6
7 which produces a larger variance: [ Ŷ x=x ± t α/2,n 2 MSE ] = ± 2.04 = ( 0.560, 0.492) [ ] ( ) This means that we predict that 95% of the time, the perceived astringency will be between and when the tannin level is 0.6. (e) We can test the null hypothesis with a confidence interval for the average astringency for a tannin concentration of 0.7, and if 0 is in the interval, then we fail to reject the null hypothesis. H o : EY x=0.7 = 0, H a : EY x=0.7 0 [ ] 1 Ŷ x=x ± t α/2,n 2 MSE [ ] 1 ( )2 = ± = (0.125, 0.324) We are 95% confident that the true average perceived astringency level is between and when the tannin concentration is equal to 0.7. Because zero is not contained in the interval, we can also reject the null hypothesis. There is evidence that this average is significantly different from zero. 4. (a) Based on the given calculations, the estimated regression line is Ŷ = x cf. This means that for every SCCM unit increase in chlorine flow, the etch rate increases A/min. With no chlorine flow, the average etch rate is A/min. A check of our assumptions shows that the assumption of linearity has not been violated. An examination of the residuals is harder to determine because the sample size is small. However, the residuals are between -2 and 2 and seem to show no patterns. R 2 is equal to 0.94, which means that 94% of CF vs etch rate Fitted vs Residuals Histogram of estar Etch rate Residuals Frequency CF Fitted estar the variation in the etch rate is explained by the chlorine flow. It seems that the regression model specifies a useful relationship between chlorine flow and etch rate. HERE DO AN F-TEST? 7
8 (b) The average change in etch rate associated with a 1-SCCM increase in flow rate is the slope, β 1. The estimate for β 1 is equal to 10.6 (given). We can create a 95% confidence interval for this parameter using the equation: (c) (d) ˆβ 1 ± t α/2,n 2 MSE/Sxx = ± /6.5 = (8.24, 12.96). We are 95% confident that the true average change in etch rate is between 8.24 and for every 1-SCCM increase in chlorine flow rate. [ ] 1 Ŷ x=3 ± t α/2,n 2 MSE [ ] 1 (3 2.67)2 = ± = (36.098, ) We are 95% confident that the average etch rate is between 36.1 and A/min when the chlorine flow is 3 SCCM. Because 3 falls in the range of the x values in the data set, it seems reasonable to assume that our estimate of the average etch rate is likely to be accurate. [ Ŷ x=3 ± t α/2,n 2 MSE ] = ± = (35.05, 41.45) 6.48 [ ] (3 2.67) (e) The standard error of the prediction intervals and the confidence intervals contains the term (x x) 2. The value of x that is closer to the average will produce a smaller standard error than a value that is further. The average chlorine flow values is 2.67 SCCM, and because 2.5 is closer to this average than 3.0, the confidence and prediction intervals for EY x=2.5 will be smaller than EY x=3.0. (f) It would not wise to recommend a 95% PI for a flow of 6.0, because this value is so far from any of the recorded x values in the data set. The maximum value is 4.0 SCCM, and because 6.0 is much higher, the interval will be very wide and inaccurate. 5. The estimated regression equation is: Ŷ = NAOH T IME When the NaOH and treatment time are equal to 0, the average specific surface area is 6.05 cm 2 /g. When treatment time is held fixed, a one-percent increase in NaOH causes a increase in cm 2 /g in surface area. When NaOH is held fixed, a one minute increase in treatment time decreases the surface are by cm 2 /g. (a) R 2 = 0.807, which means that time and NaOH account for 80.7% of the variation in surface area. (b) The p-value for the entire model (the F-statistic) is 0.007, which means that there is a useful relationship between the dependent variable and the predictors. (c) Provided that the percentage of NaOH remains in the model, it does not appear the predictor treatment time needs to be eliminated if we use a significance level α = 0.05, since the p- value for that coefficient is (which means that we reject H o : β time = 0). However, if the model were being validated against a higher significance level, such as α = 0.01, then we would recommend possibly eliminating this variable from the model. 8
9 (d) Calculating a 95% CI for the expected change in specific surface area associated with a 1% in NaOH (treatment time is held fixed) means we are calculating a 95% CI for β NaOH. We can use the output to make this confidence interval, using the standard error provided by the output: ± t.975, = ± = (0.061, 0.222) Note that the confidence interval does not contain zero, which means that we can reject the null hypothesis that this parameter is equal to 0 at the 0.05 level. 9
ST430 Exam 1 with Answers
ST430 Exam 1 with Answers Date: October 5, 2015 Name: Guideline: You may use one-page (front and back of a standard A4 paper) of notes. No laptop or textook are permitted but you may use a calculator.
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationInference for Regression
Inference for Regression Section 9.4 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Department of Mathematics University of Houston Lecture 13b - 3339 Cathy Poliak, Ph.D. cathy@math.uh.edu
More informationSimple Linear Regression
Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)
More informationSimple and Multiple Linear Regression
Sta. 113 Chapter 12 and 13 of Devore March 12, 2010 Table of contents 1 Simple Linear Regression 2 Model Simple Linear Regression A simple linear regression model is given by Y = β 0 + β 1 x + ɛ where
More informationNature vs. nurture? Lecture 18 - Regression: Inference, Outliers, and Intervals. Regression Output. Conditions for inference.
Understanding regression output from software Nature vs. nurture? Lecture 18 - Regression: Inference, Outliers, and Intervals In 1966 Cyril Burt published a paper called The genetic determination of differences
More information401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.
401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis
More informationMultiple Linear Regression
Multiple Linear Regression ST 430/514 Recall: a regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates).
More informationLecture 18: Simple Linear Regression
Lecture 18: Simple Linear Regression BIOS 553 Department of Biostatistics University of Michigan Fall 2004 The Correlation Coefficient: r The correlation coefficient (r) is a number that measures the strength
More informationLinear Regression. Simple linear regression model determines the relationship between one dependent variable (y) and one independent variable (x).
Linear Regression Simple linear regression model determines the relationship between one dependent variable (y) and one independent variable (x). A dependent variable is a random variable whose variation
More informationUnit 6 - Simple linear regression
Sta 101: Data Analysis and Statistical Inference Dr. Çetinkaya-Rundel Unit 6 - Simple linear regression LO 1. Define the explanatory variable as the independent variable (predictor), and the response variable
More informationEstimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.
Estimating σ 2 We can do simple prediction of Y and estimation of the mean of Y at any value of X. To perform inferences about our regression line, we must estimate σ 2, the variance of the error term.
More informationUNIVERSITY OF MASSACHUSETTS. Department of Mathematics and Statistics. Basic Exam - Applied Statistics. Tuesday, January 17, 2017
UNIVERSITY OF MASSACHUSETTS Department of Mathematics and Statistics Basic Exam - Applied Statistics Tuesday, January 17, 2017 Work all problems 60 points are needed to pass at the Masters Level and 75
More informationScatter plot of data from the study. Linear Regression
1 2 Linear Regression Scatter plot of data from the study. Consider a study to relate birthweight to the estriol level of pregnant women. The data is below. i Weight (g / 100) i Weight (g / 100) 1 7 25
More informationUnit 6 - Introduction to linear regression
Unit 6 - Introduction to linear regression Suggested reading: OpenIntro Statistics, Chapter 7 Suggested exercises: Part 1 - Relationship between two numerical variables: 7.7, 7.9, 7.11, 7.13, 7.15, 7.25,
More informationAMS 315/576 Lecture Notes. Chapter 11. Simple Linear Regression
AMS 315/576 Lecture Notes Chapter 11. Simple Linear Regression 11.1 Motivation A restaurant opening on a reservations-only basis would like to use the number of advance reservations x to predict the number
More informationAMS 7 Correlation and Regression Lecture 8
AMS 7 Correlation and Regression Lecture 8 Department of Applied Mathematics and Statistics, University of California, Santa Cruz Suumer 2014 1 / 18 Correlation pairs of continuous observations. Correlation
More informationHandout 4: Simple Linear Regression
Handout 4: Simple Linear Regression By: Brandon Berman The following problem comes from Kokoska s Introductory Statistics: A Problem-Solving Approach. The data can be read in to R using the following code:
More informationScatter plot of data from the study. Linear Regression
1 2 Linear Regression Scatter plot of data from the study. Consider a study to relate birthweight to the estriol level of pregnant women. The data is below. i Weight (g / 100) i Weight (g / 100) 1 7 25
More informationTABLES AND FORMULAS FOR MOORE Basic Practice of Statistics
TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics Exploring Data: Distributions Look for overall pattern (shape, center, spread) and deviations (outliers). Mean (use a calculator): x = x 1 + x
More informationInferences for Regression
Inferences for Regression An Example: Body Fat and Waist Size Looking at the relationship between % body fat and waist size (in inches). Here is a scatterplot of our data set: Remembering Regression In
More informationStatistics for Engineers Lecture 9 Linear Regression
Statistics for Engineers Lecture 9 Linear Regression Chong Ma Department of Statistics University of South Carolina chongm@email.sc.edu April 17, 2017 Chong Ma (Statistics, USC) STAT 509 Spring 2017 April
More informationCh 3: Multiple Linear Regression
Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery
More informationCoefficient of Determination
Coefficient of Determination ST 430/514 The coefficient of determination, R 2, is defined as before: R 2 = 1 SS E (yi ŷ i ) = 1 2 SS yy (yi ȳ) 2 The interpretation of R 2 is still the fraction of variance
More informationMATH 644: Regression Analysis Methods
MATH 644: Regression Analysis Methods FINAL EXAM Fall, 2012 INSTRUCTIONS TO STUDENTS: 1. This test contains SIX questions. It comprises ELEVEN printed pages. 2. Answer ALL questions for a total of 100
More information13 Simple Linear Regression
B.Sc./Cert./M.Sc. Qualif. - Statistics: Theory and Practice 3 Simple Linear Regression 3. An industrial example A study was undertaken to determine the effect of stirring rate on the amount of impurity
More informationIntroduction and Single Predictor Regression. Correlation
Introduction and Single Predictor Regression Dr. J. Kyle Roberts Southern Methodist University Simmons School of Education and Human Development Department of Teaching and Learning Correlation A correlation
More informationLecture 11: Simple Linear Regression
Lecture 11: Simple Linear Regression Readings: Sections 3.1-3.3, 11.1-11.3 Apr 17, 2009 In linear regression, we examine the association between two quantitative variables. Number of beers that you drink
More informationMAT2377. Rafa l Kulik. Version 2015/November/26. Rafa l Kulik
MAT2377 Rafa l Kulik Version 2015/November/26 Rafa l Kulik Bivariate data and scatterplot Data: Hydrocarbon level (x) and Oxygen level (y): x: 0.99, 1.02, 1.15, 1.29, 1.46, 1.36, 0.87, 1.23, 1.55, 1.40,
More informationOct Simple linear regression. Minimum mean square error prediction. Univariate. regression. Calculating intercept and slope
Oct 2017 1 / 28 Minimum MSE Y is the response variable, X the predictor variable, E(X) = E(Y) = 0. BLUP of Y minimizes average discrepancy var (Y ux) = C YY 2u C XY + u 2 C XX This is minimized when u
More informationComparing Nested Models
Comparing Nested Models ST 370 Two regression models are called nested if one contains all the predictors of the other, and some additional predictors. For example, the first-order model in two independent
More informationApplied Regression Analysis
Applied Regression Analysis Chapter 3 Multiple Linear Regression Hongcheng Li April, 6, 2013 Recall simple linear regression 1 Recall simple linear regression 2 Parameter Estimation 3 Interpretations of
More informationOrdinary Least Squares Regression Explained: Vartanian
Ordinary Least Squares Regression Explained: Vartanian When to Use Ordinary Least Squares Regression Analysis A. Variable types. When you have an interval/ratio scale dependent variable.. When your independent
More informationLecture 6 Multiple Linear Regression, cont.
Lecture 6 Multiple Linear Regression, cont. BIOST 515 January 22, 2004 BIOST 515, Lecture 6 Testing general linear hypotheses Suppose we are interested in testing linear combinations of the regression
More informationChapter 16: Understanding Relationships Numerical Data
Chapter 16: Understanding Relationships Numerical Data These notes reflect material from our text, Statistics, Learning from Data, First Edition, by Roxy Peck, published by CENGAGE Learning, 2015. Linear
More information22s:152 Applied Linear Regression. Take random samples from each of m populations.
22s:152 Applied Linear Regression Chapter 8: ANOVA NOTE: We will meet in the lab on Monday October 10. One-way ANOVA Focuses on testing for differences among group means. Take random samples from each
More informationHomework 2: Simple Linear Regression
STAT 4385 Applied Regression Analysis Homework : Simple Linear Regression (Simple Linear Regression) Thirty (n = 30) College graduates who have recently entered the job market. For each student, the CGPA
More informationSTAT 350: Summer Semester Midterm 1: Solutions
Name: Student Number: STAT 350: Summer Semester 2008 Midterm 1: Solutions 9 June 2008 Instructor: Richard Lockhart Instructions: This is an open book test. You may use notes, text, other books and a calculator.
More informationStatistical View of Least Squares
May 23, 2006 Purpose of Regression Some Examples Least Squares Purpose of Regression Purpose of Regression Some Examples Least Squares Suppose we have two variables x and y Purpose of Regression Some Examples
More informationReview of Statistics 101
Review of Statistics 101 We review some important themes from the course 1. Introduction Statistics- Set of methods for collecting/analyzing data (the art and science of learning from data). Provides methods
More informationChapter 8: Simple Linear Regression
Chapter 8: Simple Linear Regression Shiwen Shen University of South Carolina 2017 Summer 1 / 70 Introduction A problem that arises in engineering, economics, medicine, and other areas is that of investigating
More informationST430 Exam 2 Solutions
ST430 Exam 2 Solutions Date: November 9, 2015 Name: Guideline: You may use one-page (front and back of a standard A4 paper) of notes. No laptop or textbook are permitted but you may use a calculator. Giving
More informationBiostatistics 380 Multiple Regression 1. Multiple Regression
Biostatistics 0 Multiple Regression ORIGIN 0 Multiple Regression Multiple Regression is an extension of the technique of linear regression to describe the relationship between a single dependent (response)
More informationLinear Regression. In this lecture we will study a particular type of regression model: the linear regression model
1 Linear Regression 2 Linear Regression In this lecture we will study a particular type of regression model: the linear regression model We will first consider the case of the model with one predictor
More informationMath 3330: Solution to midterm Exam
Math 3330: Solution to midterm Exam Question 1: (14 marks) Suppose the regression model is y i = β 0 + β 1 x i + ε i, i = 1,, n, where ε i are iid Normal distribution N(0, σ 2 ). a. (2 marks) Compute the
More informationSLR output RLS. Refer to slr (code) on the Lecture Page of the class website.
SLR output RLS Refer to slr (code) on the Lecture Page of the class website. Old Faithful at Yellowstone National Park, WY: Simple Linear Regression (SLR) Analysis SLR analysis explores the linear association
More informationINFERENCE FOR REGRESSION
CHAPTER 3 INFERENCE FOR REGRESSION OVERVIEW In Chapter 5 of the textbook, we first encountered regression. The assumptions that describe the regression model we use in this chapter are the following. We
More informationRegression, Part I. - In correlation, it would be irrelevant if we changed the axes on our graph.
Regression, Part I I. Difference from correlation. II. Basic idea: A) Correlation describes the relationship between two variables, where neither is independent or a predictor. - In correlation, it would
More informationFigure 1: The fitted line using the shipment route-number of ampules data. STAT5044: Regression and ANOVA The Solution of Homework #2 Inyoung Kim
0.0 1.0 1.5 2.0 2.5 3.0 8 10 12 14 16 18 20 22 y x Figure 1: The fitted line using the shipment route-number of ampules data STAT5044: Regression and ANOVA The Solution of Homework #2 Inyoung Kim Problem#
More informationLinear Regression Model. Badr Missaoui
Linear Regression Model Badr Missaoui Introduction What is this course about? It is a course on applied statistics. It comprises 2 hours lectures each week and 1 hour lab sessions/tutorials. We will focus
More informationL21: Chapter 12: Linear regression
L21: Chapter 12: Linear regression Department of Statistics, University of South Carolina Stat 205: Elementary Statistics for the Biological and Life Sciences 1 / 37 So far... 12.1 Introduction One sample
More informationSTAT Chapter 11: Regression
STAT 515 -- Chapter 11: Regression Mostly we have studied the behavior of a single random variable. Often, however, we gather data on two random variables. We wish to determine: Is there a relationship
More informationOrdinary Least Squares Regression Explained: Vartanian
Ordinary Least Squares Regression Eplained: Vartanian When to Use Ordinary Least Squares Regression Analysis A. Variable types. When you have an interval/ratio scale dependent variable.. When your independent
More informationSimple Linear Regression
Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.
More informationLecture 3: Inference in SLR
Lecture 3: Inference in SLR STAT 51 Spring 011 Background Reading KNNL:.1.6 3-1 Topic Overview This topic will cover: Review of hypothesis testing Inference about 1 Inference about 0 Confidence Intervals
More informationStat 135, Fall 2006 A. Adhikari HOMEWORK 10 SOLUTIONS
Stat 135, Fall 2006 A. Adhikari HOMEWORK 10 SOLUTIONS 1a) The model is cw i = β 0 + β 1 el i + ɛ i, where cw i is the weight of the ith chick, el i the length of the egg from which it hatched, and ɛ i
More informationIntroduction to Linear Regression Rebecca C. Steorts September 15, 2015
Introduction to Linear Regression Rebecca C. Steorts September 15, 2015 Today (Re-)Introduction to linear models and the model space What is linear regression Basic properties of linear regression Using
More information9 Correlation and Regression
9 Correlation and Regression SW, Chapter 12. Suppose we select n = 10 persons from the population of college seniors who plan to take the MCAT exam. Each takes the test, is coached, and then retakes the
More informationSimple Linear Regression
Simple Linear Regression ST 370 Regression models are used to study the relationship of a response variable and one or more predictors. The response is also called the dependent variable, and the predictors
More informationSimple Linear Regression. (Chs 12.1, 12.2, 12.4, 12.5)
10 Simple Linear Regression (Chs 12.1, 12.2, 12.4, 12.5) Simple Linear Regression Rating 20 40 60 80 0 5 10 15 Sugar 2 Simple Linear Regression Rating 20 40 60 80 0 5 10 15 Sugar 3 Simple Linear Regression
More informationA discussion on multiple regression models
A discussion on multiple regression models In our previous discussion of simple linear regression, we focused on a model in which one independent or explanatory variable X was used to predict the value
More informationRegression: Main Ideas Setting: Quantitative outcome with a quantitative explanatory variable. Example, cont.
TCELL 9/4/205 36-309/749 Experimental Design for Behavioral and Social Sciences Simple Regression Example Male black wheatear birds carry stones to the nest as a form of sexual display. Soler et al. wanted
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the
More informationSTA 302 H1F / 1001 HF Fall 2007 Test 1 October 24, 2007
STA 302 H1F / 1001 HF Fall 2007 Test 1 October 24, 2007 LAST NAME: SOLUTIONS FIRST NAME: STUDENT NUMBER: ENROLLED IN: (circle one) STA 302 STA 1001 INSTRUCTIONS: Time: 90 minutes Aids allowed: calculator.
More informationLectures on Simple Linear Regression Stat 431, Summer 2012
Lectures on Simple Linear Regression Stat 43, Summer 0 Hyunseung Kang July 6-8, 0 Last Updated: July 8, 0 :59PM Introduction Previously, we have been investigating various properties of the population
More informationLINEAR REGRESSION ANALYSIS. MODULE XVI Lecture Exercises
LINEAR REGRESSION ANALYSIS MODULE XVI Lecture - 44 Exercises Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Exercise 1 The following data has been obtained on
More information22s:152 Applied Linear Regression. There are a couple commonly used models for a one-way ANOVA with m groups. Chapter 8: ANOVA
22s:152 Applied Linear Regression Chapter 8: ANOVA NOTE: We will meet in the lab on Monday October 10. One-way ANOVA Focuses on testing for differences among group means. Take random samples from each
More information2.4.3 Estimatingσ Coefficient of Determination 2.4. ASSESSING THE MODEL 23
2.4. ASSESSING THE MODEL 23 2.4.3 Estimatingσ 2 Note that the sums of squares are functions of the conditional random variables Y i = (Y X = x i ). Hence, the sums of squares are random variables as well.
More informationCorrelation and the Analysis of Variance Approach to Simple Linear Regression
Correlation and the Analysis of Variance Approach to Simple Linear Regression Biometry 755 Spring 2009 Correlation and the Analysis of Variance Approach to Simple Linear Regression p. 1/35 Correlation
More informationSociology 6Z03 Review II
Sociology 6Z03 Review II John Fox McMaster University Fall 2016 John Fox (McMaster University) Sociology 6Z03 Review II Fall 2016 1 / 35 Outline: Review II Probability Part I Sampling Distributions Probability
More informationChapter 12: Linear regression II
Chapter 12: Linear regression II Timothy Hanson Department of Statistics, University of South Carolina Stat 205: Elementary Statistics for the Biological and Life Sciences 1 / 14 12.4 The regression model
More information36-309/749 Experimental Design for Behavioral and Social Sciences. Sep. 22, 2015 Lecture 4: Linear Regression
36-309/749 Experimental Design for Behavioral and Social Sciences Sep. 22, 2015 Lecture 4: Linear Regression TCELL Simple Regression Example Male black wheatear birds carry stones to the nest as a form
More informationRegression. Marc H. Mehlman University of New Haven
Regression Marc H. Mehlman marcmehlman@yahoo.com University of New Haven the statistician knows that in nature there never was a normal distribution, there never was a straight line, yet with normal and
More informationBIOL 458 BIOMETRY Lab 9 - Correlation and Bivariate Regression
BIOL 458 BIOMETRY Lab 9 - Correlation and Bivariate Regression Introduction to Correlation and Regression The procedures discussed in the previous ANOVA labs are most useful in cases where we are interested
More informationFinal Exam. Name: Solution:
Final Exam. Name: Instructions. Answer all questions on the exam. Open books, open notes, but no electronic devices. The first 13 problems are worth 5 points each. The rest are worth 1 point each. HW1.
More information36-707: Regression Analysis Homework Solutions. Homework 3
36-707: Regression Analysis Homework Solutions Homework 3 Fall 2012 Problem 1 Y i = βx i + ɛ i, i {1, 2,..., n}. (a) Find the LS estimator of β: RSS = Σ n i=1(y i βx i ) 2 RSS β = Σ n i=1( 2X i )(Y i βx
More informationRegression Analysis II
Regression Analysis II Measures of Goodness of fit Two measures of Goodness of fit Measure of the absolute fit of the sample points to the sample regression line Standard error of the estimate An index
More informationSTAT 3022 Spring 2007
Simple Linear Regression Example These commands reproduce what we did in class. You should enter these in R and see what they do. Start by typing > set.seed(42) to reset the random number generator so
More informationModel Specification and Data Problems. Part VIII
Part VIII Model Specification and Data Problems As of Oct 24, 2017 1 Model Specification and Data Problems RESET test Non-nested alternatives Outliers A functional form misspecification generally means
More informationMatrices and vectors A matrix is a rectangular array of numbers. Here s an example: A =
Matrices and vectors A matrix is a rectangular array of numbers Here s an example: 23 14 17 A = 225 0 2 This matrix has dimensions 2 3 The number of rows is first, then the number of columns We can write
More informationR 2 and F -Tests and ANOVA
R 2 and F -Tests and ANOVA December 6, 2018 1 Partition of Sums of Squares The distance from any point y i in a collection of data, to the mean of the data ȳ, is the deviation, written as y i ȳ. Definition.
More informationSSR = The sum of squared errors measures how much Y varies around the regression line n. It happily turns out that SSR + SSE = SSTO.
Analysis of variance approach to regression If x is useless, i.e. β 1 = 0, then E(Y i ) = β 0. In this case β 0 is estimated by Ȳ. The ith deviation about this grand mean can be written: deviation about
More informationSimple linear regression
Simple linear regression Biometry 755 Spring 2008 Simple linear regression p. 1/40 Overview of regression analysis Evaluate relationship between one or more independent variables (X 1,...,X k ) and a single
More informationBusiness Statistics. Lecture 10: Course Review
Business Statistics Lecture 10: Course Review 1 Descriptive Statistics for Continuous Data Numerical Summaries Location: mean, median Spread or variability: variance, standard deviation, range, percentiles,
More informationVariance Decomposition and Goodness of Fit
Variance Decomposition and Goodness of Fit 1. Example: Monthly Earnings and Years of Education In this tutorial, we will focus on an example that explores the relationship between total monthly earnings
More informationAny of 27 linear and nonlinear models may be fit. The output parallels that of the Simple Regression procedure.
STATGRAPHICS Rev. 9/13/213 Calibration Models Summary... 1 Data Input... 3 Analysis Summary... 5 Analysis Options... 7 Plot of Fitted Model... 9 Predicted Values... 1 Confidence Intervals... 11 Observed
More informationLecture 4 Multiple linear regression
Lecture 4 Multiple linear regression BIOST 515 January 15, 2004 Outline 1 Motivation for the multiple regression model Multiple regression in matrix notation Least squares estimation of model parameters
More informationSimple Linear Regression. Material from Devore s book (Ed 8), and Cengagebrain.com
12 Simple Linear Regression Material from Devore s book (Ed 8), and Cengagebrain.com The Simple Linear Regression Model The simplest deterministic mathematical relationship between two variables x and
More informationCorrelation and Simple Linear Regression
Correlation and Simple Linear Regression Sasivimol Rattanasiri, Ph.D Section for Clinical Epidemiology and Biostatistics Ramathibodi Hospital, Mahidol University E-mail: sasivimol.rat@mahidol.ac.th 1 Outline
More informationInference for Regression Inference about the Regression Model and Using the Regression Line
Inference for Regression Inference about the Regression Model and Using the Regression Line PBS Chapter 10.1 and 10.2 2009 W.H. Freeman and Company Objectives (PBS Chapter 10.1 and 10.2) Inference about
More informationAnalysis of Bivariate Data
Analysis of Bivariate Data Data Two Quantitative variables GPA and GAES Interest rates and indices Tax and fund allocation Population size and prison population Bivariate data (x,y) Case corr® 2 Independent
More informationReview of Multiple Regression
Ronald H. Heck 1 Let s begin with a little review of multiple regression this week. Linear models [e.g., correlation, t-tests, analysis of variance (ANOVA), multiple regression, path analysis, multivariate
More informationMidterm 2 - Solutions
Ecn 102 - Analysis of Economic Data University of California - Davis February 23, 2010 Instructor: John Parman Midterm 2 - Solutions You have until 10:20am to complete this exam. Please remember to put
More information9. Linear Regression and Correlation
9. Linear Regression and Correlation Data: y a quantitative response variable x a quantitative explanatory variable (Chap. 8: Recall that both variables were categorical) For example, y = annual income,
More informationUNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, October 2013
UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, October 2013 STAC67H3 Regression Analysis Duration: One hour and fifty minutes Last Name: First Name: Student
More informationData Analysis and Statistical Methods Statistics 651
y 1 2 3 4 5 6 7 x Data Analysis and Statistical Methods Statistics 651 http://www.stat.tamu.edu/~suhasini/teaching.html Lecture 32 Suhasini Subba Rao Previous lecture We are interested in whether a dependent
More informationAssumptions, Diagnostics, and Inferences for the Simple Linear Regression Model with Normal Residuals
Assumptions, Diagnostics, and Inferences for the Simple Linear Regression Model with Normal Residuals 4 December 2018 1 The Simple Linear Regression Model with Normal Residuals In previous class sessions,
More informationIES 612/STA 4-573/STA Winter 2008 Week 1--IES 612-STA STA doc
IES 612/STA 4-573/STA 4-576 Winter 2008 Week 1--IES 612-STA 4-573-STA 4-576.doc Review Notes: [OL] = Ott & Longnecker Statistical Methods and Data Analysis, 5 th edition. [Handouts based on notes prepared
More informationK. Model Diagnostics. residuals ˆɛ ij = Y ij ˆµ i N = Y ij Ȳ i semi-studentized residuals ω ij = ˆɛ ij. studentized deleted residuals ɛ ij =
K. Model Diagnostics We ve already seen how to check model assumptions prior to fitting a one-way ANOVA. Diagnostics carried out after model fitting by using residuals are more informative for assessing
More information1 Multiple Regression
1 Multiple Regression In this section, we extend the linear model to the case of several quantitative explanatory variables. There are many issues involved in this problem and this section serves only
More information