Modules 1-2 are background; they are the same for regression analysis and time series.

Size: px
Start display at page:

Download "Modules 1-2 are background; they are the same for regression analysis and time series."

Transcription

1 Regression Analysis, Module 1: Regression models (The attached PDF file has better formatting.) Required reading: Chapter 1, pages 3 13 (until appendix 1.1). Updated: May 23, 2005 Modules 1-2 are background; they are the same for regression analysis and time series. Jacob: This chapter does not seem like background. It covers much of regression analysis, as if we have already had a course in this subject. Rachel: This chapter provides the basic formulas. Chapter 3 begins the actual course, deriving the formulas and explaining the intuition. This chapter is for candidates who have never dealt with regression analysis, and who are not familiar with the basic equations. Section 1.1 on page 3-7 covers curve fitting. Know the definition of the least squares criterion on page 5: The line of best fit minimizes the sum of squared deviations Figure 1.2 on page 5 shows the graphic interpretation. Jacob: Do we prove that minimizing the sum of square deviations is the best way to fit a curve? Rachel: No; this is a definition. We could use other loss functions instead. For the final exam, you must know the pros and cons of different loss functions. Throughout this course, we discuss deviations. The deviations may be from the regression line (error sum of squares) or from the mean (total sum of squares). To minimize the sum of squared deviations and keep the algebra simple, we solve a regression problem in two steps: (i) move the graph horizontally and vertically so that the means are zero, and (ii) determine the line of best fit passing through the origin. Jacob: What does move the graph horizontally and vertically mean? Do we move the regression line horizontally and vertically or do we move the co-ordinates of the graph? Rachel: Those are the same thing. Moving the line upward (vertically) ten units is moving the origin of the graph downward ten units. Figure 1.3 on page 6 shows alternative loss functions. One of the illustrative test questions compares alternative loss functions. We use the least squares loss function, which has some desirable properties. But we do not say that this loss function is correct, that other loss functions are incorrect, or that this loss function is better than others. We might minimize the absolute error, if it were mathematically tractable. In fact, minimizing absolute

2 errors has become a common alternative with the spread of spreadsheet capabilities. In this course, we use the least squares loss function. Section 1.2 on pages 7-10 is the essence of regression analysis. Equation 1.6 on page 9 in the shaded type is the core equation for this course. In practice, much of regression analysis is finding the value of b, which we refer to in this course as, the ordinary least squares estimator. Once we know b, we solve for a (referred to as ) from equation 1.3 on page 8. You are not tested on the derivation of the regression coefficients a and b; the final exam is multiple choice and does not test derivations. But it is worth your while to review the mathematics. We solve for b repeatedly, and it helps to know what this formula does. Read Example 1.1 on page 10 and 11. You must work similar problems for both the homework assignments and the final exam. One of the continuing illustrations in this course regresses scores on Course C on the hours each candidate spends studying. We show several variations of this problem, corresponding to the topics in each module. The equations for the regression coefficients can be written in terms of X and Y or in terms of x and y. The lower case variables are the deviations, or the upper-case variables minus the means: x = X. Some texts use the formulas in terms of X and Y. You should know those formulas, but we generally use the formulas in terms of the deviations. Appendix 1.1 covers properties of sums. We use these properties in the rest of the course. Know especially Rule 5 and Rule 6 on page 15; these are used in many of the homework assignments and exam questions. Rules 5 and 6 allow us to convert between absolute numbers (X and Y) and deviations (x and y). On an exam problem, you may be given the sum of squares and the mean and asked to derive the sum of squared deviations; this is Rule 6, equation A1.11, on page 15. Appendix 1.2 is optional. You will not be tested on the derivation of the ordinary least squares estimators and, but you must know the formulas for them (A1.23 and A1.24). Come back to this appendix several times: after modules 3, 5, 7, and 22. The homework assignment is similar to the grade point average example in Table 1.2 on page 11. Work through the example in the text, and then do the homework assignment. (The PDF attachment shows the Greek letters for the ordinary least squares estimators.)

3 Time Series, Modules 1, 2: Statistics and Regressions Practice Problems (The attached PDF file has better formatting.) Updated: May 23, 2005 Jacob: To which Module do these practice problems apply? Rachel: Module 2 is statistical background; Module 1 is the basic regression equations; and Regression Analysis Modules 3 and 4 cover the initial concepts. If you already know regression analysis, this material is easy. If you have not had regression analysis, make sure you understand these problems, since the concepts are used in time series as well. Exercise 1.2: Means Given the sample below, what are the estimated means of X and Y? i X Y i i Solution 1.2: = = ( ) / 5 = 10.0 = = ( ) / 5 = 15.0 Jacob: This is obvious; the mean is the average. What is the point of this problem? Rachel: The mean uses a divisor of N; the variance uses a divisor of N-1; see below. Exercise 1.3: Deviations Given the sample below, what are x i and y, i the deviations of X i and Y? i

4 i Xi Yi xi yi Mn Jacob: Is the deviation the same as the residual? Rachel: The deviation is the difference from the mean; the residual is the difference from the fitted value. Exercise 1.4: Sample Variances What are the sample variances of X and Y? Solution 1.4: 2 2 i Xi Yi xi yi xi yi Mn s (X) = = [( 2) + ( 1) ] / 4 = 10 / 4 = s (Y) = = [( 3) + ( 1) ] / 4 = 18 / 4 = 4.5 An alternative method is X = = s (X) = = ( ) / 4 = 2.5 Jacob: Why is the divisor N-1?

5 Rachel: The divisor is the degrees of freedom. Separate postings explain why the degrees of freedom is N-1. Exercise 1.5: Covariance What is estimated covariance between the two random variables? Solution 1.5: 2 2 i Xi Yi xi yi xi yi x i yi Mn = = 13 The estimated covariance of (x,y) is = 13 / 4 = Exercise 1.6: Correlation What is the correlation between the two random variables? Solution 1.6: ½ = = 3.25 / ( ) = {Note: The practice problems above are explained in Module 2, though many actuarial candidates are familiar with these formulas. The practice problem below is for Modules 1, 3, and 4.} Suppose we fit the two-variable regression model to these pairs of points. We assume now that Y is a random variable, but X is not a random variable. Exercise 1.7: Ordinary Least Squares Estimates

6 A. What are the values of and? B. What is the ordinary least squares estimate of? C. What is the ordinary least squares estimate of? D. Using the values of and computed earlier, what is the ordinary least squares estimate of using deviations? E. What are the values of? i F. What is the total sum of squares (TSS)? G. What is the error sum of squares (ESS), or the residual variance? H. What is the regression sum of squares (RSS), or the explained variance? I. What is the variance of the ordinary least squares estimator? J. What is the variance of the ordinary least squares estimator? Solution 1.7: We calculate the sum of squares, the deviations, the sum of squared deviations, the sum of the cross terms, and the sum of the cross deviations. 2 2 i Xi Yi Xi Yi X i Yi Total / Mean Part A: = 50.0, = The mean of X is 50 / 5 = 10; the mean of Y is 75 / 5 = 15. = , = 510.0, = Part B: = = ( / 5) / ( / 5) = Part C: = = 75 / / 5 = 2.0

7 Part D: = 10 and = 13, so = 13/10 = 1.3 Parts E-H: The table below shows the calculations. i Xi Yi i residual ESS Total / Mean The fit is good: the fitted values are close to the actual values and the residuals are small. We determine the fitted values by the estimated regression equation: Y = X. For the first row, = The average residual is zero, as is true by convention. If we work out the ordinary least squares estimators correctly, the average residual is zero. Jacob: Must the rows have different values for X? Rachel: No; several rows may have the same value of X. Suppose we regress Course C scores on the hours of study, and we examine all 2,500 candidates who take the test in 20XX. If we estimate hours of study to the nearest 10 hours, we may have 50 candidates who study 400 hours. The scores for these candidates may vary, though they have the same fitted score. Jacob: Is it possible to run a regression analysis if all data points have the same X value? Rachel: If all X values are the same, the sum of squared deviations is zero, and the ordinary least squares estimator for is not defined. Part F: We work out the total sum of squares two ways:! The deviations of Y are the Y values minus the mean, or 3, 1, 0, 2, 2. The squares of the deviations are 9, 1, 0, 4, 4; the sum of these squares is ! The sum of squared deviations is the sum of the Y N = 1, / 5 = 18. Part G: The error sum of squares is shown in the table as 1.10 Part H: The regression sum of squares is = Part I: R = / = 93.89%

8 2 Jacob: Can we derive R as the square of the correlation between X and Y? Rachel: Yes. The correlation is the covariance divided by the standard deviations of each ½ variable. We worked out the needed figures above. The correlation is 13 / (10 18) = 0.969; the square of the correlation is Part J: For the variances of the ordinary least squares estimators, we must know the variance of the error term. We estimate the variance of the error term from the variance of the residuals, which is the error sum of squares divided by the degrees of freedom. We have five data points, so three degrees of freedom. The estimated variance of the error term is 1.10 / 3 = For the variance of the, we divide the variance of the error term by the sum of squared deviations of the X variable: / 10 = Part K: The variance of is / 5 =

9 Regression Analysis, Module 1, Introduction to the Regression Model (The attached PDF file has better formatting.) Homework Assignment Updated: May 23, 2005 Modules 3 and 4 repeat the information in Module 1 with more explanation. The textbook assumes you know the basic formulas of simple linear regression. The textbook focuses on the concepts and the intuition, not the formulas. You see this from the exercises at the end of each chapter; they review the concepts, not the equations. Read this homework assignment after reading chapter 1; do the assignment after reading chapter 3. If you have not had regression analysis before, it takes a while to get used to the equations. After the first few weeks of this course, the equations come naturally. Problem 1: We examine four items: the ordinary least squares estimators for beta and alpha, an in-range forecast, and an out-of-range forecast. We are examining the effects on study on exam scores. For the eight candidates below, the table shows the number of hours studied and the score on Course C (Exam 4): Candidate Hours Studied Exam Score a b c d e f g h We fit a two-variable regression model (Y = A + B * X) to these observations, where X is hours studied and Y is the Course C exam score. A. What is the ordinary least squares estimator for beta? B. What is the ordinary least squares estimator for alpha? C. How many hours of study are needed to get a 6 on the exam according to the regression equation? Assume that scores are rounded to the nearest integer, so we solve for Y = 5.5, not Y = 6.)

10 D. If the candidate does not study, what is the predicted exam score from the regression equation? E. (Optional) Explain why the regression equation should not be used to estimate the exam score with no hours of study. (Part E is not required for the homework.) {Part E says that we can not use the regression equation to make forecasts about outlying scenarios, since we don t know that the regression equation extends to those points.}

11 Regression Analysis and Time Series, Module 2, Statistical Processes Required Reading (The attached PDF file has better formatting.) Updated: May 25, 2005 Module 2 is background that is the same for regression analysis and time series. If you are taking both courses, one homework assignment suffices for both courses. You learn this material in greater depth in Courses M and C (CAS Exams 3 and 4). Some subjects proceed in linear sequence: you learn Fact A, then Fact B, then Fact C. For regression analysis and time series, you may not understand Fact A until you have learned Facts Y and Z. These courses are frustrating the first several modules, until you understand the general themes. There is much material in Module 2. There is nothing that you must master now; you will understand the material as you see the statistical applications in later modules. If you have never had a statistics course, Module 2 is hard, since you can t understand the concepts with no context. This module summarizes the mathematics; as you learn the material in later modules, come back to Module 2 to review the mathematics. VARIANCES Read sections 2.1 and 2.2 on pages 19-28; know equations 2.1 through These are background knowledge: equations for the mean, variance, covariance, and correlation of a sample. Know especially equation 2.5 (correlations and covariances). We use this relation also in the corporate finance course for the CAPM beta. (The CAPM beta is the slope parameter of the regression equation where Y is the individual stock return and X is the market return.) The sample variance and covariance use N-1 as the denominator, not N. For the error variance from a regression equation, we use N-k, where k is the number of explanatory variables (the independent variables plus the constant term). Jacob: If the population variance uses N as the divisor and the sample variance uses N-1, why do we use the term variance for both? This just confuses the matter. Rachel: The sample variance is an unbiased estimator of the population variance. By sample variance, we mean the estimate of the population variance using sample data. Example 2.1 on page 23 gives the probability distribution for the population, so we use N, not N-1, to derive the variances and the covariance. In Section on page 24-27, we use samples, so we use N-1, not N.

12 Several of the modules have intuition postings, often in a question and answer format with numerical illustrations. Work through the numerical illustrations to make sure you follow the reasoning. It takes a while to grasp why the sample variance is an unbiased estimate of the population variance. DEGREES OF FREEDOM The degrees of freedom is essential for statistical analysis. Jacob: What does degrees of freedom signify? The textbook uses this term but does not give a clear definition. Rachel: Suppose we take N deviations from the population whose mean is known, and we want to determine the sum of squared deviations. {Definition: the deviation is the data point minus the mean.} For example, if the mean is 4, and we take 3 deviations of 3, +1, and +4, the sum of squared deviations is = 26. We need all three deviations to determine the sum of squared deviations. Now suppose we take a sample of N deviations from the population whose mean is not known, and we want to determine the sum of squared deviations. We estimate the population mean from the sample of N points. This implies that the sum of the deviations (not the squared deviations) is zero. For example, if we take 3 deviations, of which the first two are 3 and +1, the third deviation must be +2, and the sum of squared deviations is = 14. We need only two deviations to determine the sum of squared deviations. We restate this in statistical language. When we take three deviations from a population with a known mean, we have the freedom to change any of the three points. No constraint limits the relation among the three points. When we take a sample of three deviations from a population with an unknown mean, we use the sample to estimate the population mean. Once we know two of the deviations, the third deviation is constrained by the relation that the sum of the deviations is zero. We have the freedom to change only two of the three points; once two of the three points are known, the third is determined. Jacob: The terms sample and population confuse me. What is important is whether the mean is known or unknown. Rachel: What is important is whether the points being used for the sum of squared deviations are also used to determine the mean. If they are, they are constrained by a totality constraint, and the degrees of freedom is reduced by one. Note: Mahler s Guide to Regression Analysis discusses this topic in more detail.

13 The central limit theorem in sub-section on page 28 is used throughout actuarial science. The regression analysis and time series final exams do not test this theorem, but you are assumed to know the central limit theorem to understand certain results. Jacob: Where do we use the central limit theorem in regression analysis? Rachel: For statistical testing, we assume the error term has a normal distribution. We assume this (we don t prove this), and it is not true in many situations. Under certain conditions, the central limit theorem says this assumption is true in the limit, as the number of stochastic factors increases. Read section 2.3; understand the meaning of the four properties of estimators:! 2.3.1: Bias! 2.3.2: Efficiency! 2.3.3: Mean squared error! 2.3.4: Consistency We deal most with bias and efficiency in this course. Consistency is an important attribute for large samples, but it is not discussed much in this course. The textbook notes the relation of bias, efficiency, and mean squared error. The relation is used in many actuarial applications, but the final exam does not test this relation. (Note: This relation may be tested on the CAS transition exam; see Mahler s Guide to Regression Analysis.)! Bias: An estimator is unbiased if its expected value is the value we seek.! Efficiency: One estimator is more efficient if it has a smaller variance; see page 29. Jacob: Can you give an example of an unbiased estimator vs a biased estimator? Rachel: Suppose we want to determine the mean and variance of a population. We have no prior knowledge of the mean and variance, so we take a sample of 10 points, {X 1, X 2,, X }. 10 For the mean, we use X j / 10 as our estimate. This estimate is unbiased. If the mean of the population is 4, the sample average may be more than or less than 4, but the expected value of the sample average is 4. For the variance, we start with the sum of squared deviations. The deviation is the data point minus the mean. Since we don t know the mean, we use the sample average, which is an unbiased estimate of the population mean. In this illustration, the population mean is 4. But we don t know the population mean, so we use the sample average, which might be 3.5 or 5.2 or some other number. Suppose we divide the sum of squared deviations by 10, the number of points. If we have a population of exactly these ten points, the proper divisor is 10, and the sum of squared

14 deviations divided by 10 is the variance. But if we don t know the population mean, and we use the sample average as a proxy, dividing by 10 under-estimates the variance. Dividing by N is a biased estimator. In this case, we have an unbiased estimator as well: dividing by N-1. The proof that dividing by N-1 gives an unbiased estimator is Result 9 in Appendix 2.1 on pages The proof is not required for this course, though you must know the fact. Jacob: Do we always divide by N-1 to get an unbiased estimator of the variance? Rachel: We divide by the degrees of freedom. For the variance of the error terms of a twovariable regression model, we divide by N-2. We deal with this topic in depth when we 2 discuss the F statistic and the adjusted R. Jacob: Would we ever use biased estimators in practice? Rachel: In this example, we have an unbiased estimator. Not always does an unbiased estimator exist. Jacob: If we have two estimators, of which one is biased and one is unbiased, would we ever use the biased estimator? Rachel: In many actuarial pricing scenarios, we have two estimators: one is less biased, and the other is more efficient. Illustration: Suppose we are setting Homeowners rates for Iowa, and we must estimate the average severity of fire losses. We are making rates for 20X7, and we have Homeowners experience for 20X1 through 20X5. We have two estimators of average fire severity: A. The average observed fire loss in Iowa from 20X1 through 20X5. B. The average observed countrywide fire loss from 20X1 through 20X5. We adjust the figures for inflation (loss cost trends) and other known or expected changes. Estimator A is unbiased (if we have properly adjusted for inflation and other changes). The observed fire losses in Iowa is a sample of possible fire losses in Iowa, and the sample average is an unbiased estimate of the mean. But fire losses have a high variance, and Iowa is a small state. A few large losses or the absence of large losses may distort the observed average claim severity. Estimator B is biased. Homes in other states are different from homes in Iowa, in size, construction, and the fire protection facilities in their towns. (Fire protection facilities are fire departments and fire hydrants.) We may not know if the average fire loss in other states is higher or lower than the average loss in Iowa; that is, we may not know if the estimator is biased up or down. But it would be a coincidence if the average fire loss in the country as a whole were the same size as the average fire loss in Iowa.

15 Estimator B is more efficient. The number of fire losses in the country as a whole may be 100 times the number in Iowa. Random loss fluctuations distort the countrywide average must less than the Iowa average. Jacob: Isn t it always better to have an unbiased estimator with a large variance than a biased estimator with a small variance? Rachel: Statisticians tend to use unbiased estimators, and to choose the estimator with the least variance. This is the perspective in the textbook readings, though the authors point out that bias is not always the most important criterion. Actuaries, who often deal with unbiased but highly inefficient estimators, are sensitive to the distorting effects of variance. Illustration: Suppose the expected fire loss by state ranges from $35,000 to $40,000. Fire losses are volatile, and the average observed loss in any five year period in a small state may be $15,000 lower or $30,000 higher than its mean. Iowa s average loss is unbiased, but the high variance makes it an unstable estimator: some years the indicated rate may be twice as high as needed and some years it may be only 50% of adequate. The countrywide average fire loss may be biased up or down by 5%, but it is a stable estimator. For pricing Homeowners insurance in Iowa, we may prefer to use the countrywide average fire loss, which is slightly biased but more stable. Mean Squared Error: Minimizing mean squared error puts together bias and efficiency. A biased estimator has a higher mean squared error than an unbiased estimator (if they have the same efficiency) and a more efficient estimator has a lower mean squared error than a less efficient estimator (if they have the same bias). The mean squared error is the variance plus the square of the bias. The proof is not required for this course. We show an illustration to make the concepts clear. Illustration: Suppose we have several estimators of home prices.! A is biased upward by $1,000, but it has no variance.! B is unbiased, but it is always $1,000 too high or low, with 50% chance of each.! C is biased upward by $1,000, ±$1,000 with 50% chance of each (= A + B). We show the mean squared error of each estimator: 2! Estimator A has a mean squared error of 1,000 = 1,000, ! Estimator B has a mean squared error of ½ (1, ,000 ) = 1,000, ! Estimator C has a mean squared error of ½ (2, ) = 2,000,000. Jacob: If we use average absolute error, are the results similar? Rachel: All three estimators have an average absolute error of 1,000. Jacob: Which makes more sense: average absolute error or mean squared error?

16 Rachel: If one error of $2,000 is twice as bad as two errors of $1,000 each, mean squared error is better. If one error of $2,000 is no better or worse than two errors of $1,000 each, average absolute error is better. Mean squared error is a common test for optimal credibility. Howard Mahler, who wrote one of the credibility readings on Course C and the credibility reading on CAS Exam 9, has developed tools for judging the mean squared error of experience rating credibility. The textbook does not use this formula in the modules for the regression analysis or time series courses, and the final exam does not test this formula. (Note: The CAS transition exam may test this formula; Mahler s Guide to Regression Analysis has practice problems.) Consistency: A consistent estimator is close to the true value if the sample is large enough. Suppose we estimate the standard deviation of a population; the true standard deviation is ; and the estimate is s. This estimate is consistent if s becomes close to as the sample size grows. Jacob: I presume (i) only unbiased estimators are consistent and (ii) as the sample size grows, all unbiased estimators are consistent. Rachel: Neither statement is correct, as the illustration shows. Illustration: Suppose we want the standard deviation of home prices in a population. Two appraisal firms work in the town. Firm A uses N as the divisor instead of N-1. Firm B uses N-1 as the divisor, but it gives the result to the nearest $10,000. ½! Firm A s estimate is biased; it is always too small by a factor of [(N-1)/N]. As the sample size grows, this factor becomes close to one, and the bias becomes immaterial.! Firm B s estimate is unbiased. But the estimate may never get close to the true value. If the true standard deviation is $66,000, Firm B s estimate will be $70,000 for an infinitely large sample size. (In truth, Firm B s estimate may not be unbiased, depending on the distribution of prices.) This course emphasizes bias, not consistency. A branch of regression analysis deals with large samples, for which consistency is more important than bias. Jacob: What should we know about these four attributes for the final exam? Rachel: Know the definitions, and know the examples in the postings and the textbook. Most important, know that bias, efficiency, and consistency are different attributes. Estimator A may be more or less biased or efficient than Estimator B, and either of these estimators may be consistent or not consistent. Read section 2.4 on pages From section 2.4, you must know the normal distribution in subsection You are not tested on the equation of the normal distribution, but you must know its properties. The homework assignment for maximum likelihood (Module 21)

17 uses the equation for the normal distribution. You must know this distribution for Course C, so it does not hurt to learn the equation now. Jacob: What are the properties of the normal distribution that we must know? Rachel: The range of the normal distribution is to +, and the distribution is symmetric about its mean. The distribution is bell-shaped: its value is highest at the mean (its center) and becomes lower the further one moves from the mean. Know how to use the table for the cumulative normal distribution to test hypotheses. Given a cumulative normal distribution table, an ordinary least squares estimator, the variance of the estimator, and a significance level, you should be able to test hypotheses. Jacob: Should we know the commonly used values of the normal distribution, such as for a 90% confidence interval and 1.96 for a 95% confidence interval? Rachel: Any final exam problem that uses a significance level gives you the value. Jacob: So we don t need to practice with these values? Rachel: You should definitely practice. A common mistake is to confuse a one-tailed test with a two-tailed test. Another common mistake is to assume a null hypothesis of zero when it should be something else. A little practice with the procedures avoids these errors. You must know how to use the -squared, t, and F distributions to test hypotheses. The -squared distribution is needed to prove certain theorems in regression analysis, but you will not be tested on the properties of the -squared distribution. You must know certain attributes of each distribution: The t distribution has a thicker tail than the normal distribution; see the first full paragraph on page 36. This implies that the critical t values are greater than the critical z values. We deal with this when we discuss hypothesis testing; see the last paragraph on page 36 and see the comments below about hypothesis testing. Know the shape of the distributions in Figure 2.9; you need not know equation For the F distribution, know the shape in Figure 2.10, as summarized by the last sentence in the first paragraph on page 37: The F distribution has a skewed shape and ranges in value from 0 to infinity. Also important is the two parameters of the F distribution: the first being associated with the number of estimated parameters and the second being associated with the number of degrees of freedom (middle of first paragraph on page 37). Result 13 summarizes the F distribution mathematically. You will not be tested directly on these formulas. But you must know how to use the F distribution to test hypotheses, which uses Result 13. Don t try to learn Result 13 in abstraction. Learn the applications of the

18 F test in the two modules which discuss it. As you study those modules, you can refer back to Chapter 2 of the textbook to see the mathematical under-pinning. CONFIDENCE INTERVALS AND HYPOTHESIS TESTING Read section 2.5 on pages Know the definition of the t statistic on page 40 and the difference between a Z statistic and a t statistic. The Z statistic is used when we know the variance of the random variable; the t statistic is used when we estimate the variance from the observed data; see the paragraph beginning We have assumed on page 40. The textbook shows results assuming first that we know the variance of the population, using the Z statistic. It then shows the corresponding result if we don t know the variance, using the t statistic. In practice, we rarely know the variance. For large samples, the t statistic gives about the same result as the Z statistic. The 95% confidence interval is the interval (a, b) such that 2.5% of the probability lies in (, a) and 2.5% lies in (b, + ). For the normal distribution, the confidence intervals are symmetric and the shortest possible confidence intervals. This is not true for other distributions, which may be skewed, such as a lognormal, gamma, or Pareto distribution. (The distributions used for insurance claim severity are skewed distributions.) Page 41 discusses one-tailed tests and two-tailed tests. The authors don t discuss this topic much, since a 95% one-tailed test has the same t value as a 90% two tailed test. Jacob: Are we testing whether the hypothesized result is correct? Rachel: We do not test if a particular value is correct. We test whether a null hypothesis could co-exist with the empirical data. We reject a null hypothesis if the probability that we observe the empirical experience is less than z%, given that the null hypothesis is true. Subsections deals with Type 1 and Type 2 errors. Illustration: Suppose we measure a stock s CAPM beta as If we had no data, we would assume the stock has a CAPM beta of ! The null hypothesis is that the CAPM beta is ! The alternative hypothesis is that the CAPM beta is The statistical tests are done on the null hypothesis, not the alternative hypothesis. Four scenarios are possible: 1. The null hypothesis is true, and we do not discard it. 2. The null hypothesis is true, and we discard it. 3. The null hypothesis is false, and we do not discard it. 4. The null hypothesis is false, and we discard it.

19 Jacob: If the null hypothesis is true, why would we discard it? Rachel: Suppose the null hypothesis is that the average height of North American men is less than 6 feet (about 1.82 meters). To test the hypothesis, we observe the heights of ten men walking along a city street. If the ten men have an average height of 6½ feet (about 2 meters), we would reject the hypothesis. Jacob: Let me see if I understand this. If we observe ten members of a basketball team who are visiting the city, we might reject the null hypothesis even though it is true. Rachel: That s close, but not quite correct. If a basketball team is walking down the street, the assumptions of the regression analysis do not hold, since the heights are correlated. The proper scenario is that the ten men are unrelated, but by happenstance they are all tall. This might happen, though its probability is small. We use the words discard and do not discard. We could replace these terms with discard = do not accept and do not discard = accept. If we have no other information, we presume the null hypothesis is true, so do not discard might be replaced by accept. But statistical testing is like scientific testing. We can not prove that a scientific hypothesis is true. Illustration: Newtonian mechanics explains certain facts and not others. For centuries, it was the best explanation of physical events. But contradictory evidence eventually led us to replace Newtonian mechanics with quantum mechanics. Similarly, we can not prove that a regression coefficient is correct. The empirical data may suggest that the slope parameter is 1.250, but we do not prove that this hypothesis is true. Rather, we show that it is unlikely to get these empirical data if the slope parameter is actually Scenarios 1 and 4 are proper inferences. If the null hypothesis is true, we should not discard it and if it is false, we should discard it. Scenarios 2 and 3 are errors. Scenario 2 is a Type 1 error, and Scenario 3 is a Type 2 error; see section on page 42. Type 2 errors are common. In many situations, they are almost inevitable, and they do not much concern us. Consider the illustration about the CAPM betas. Suppose the CAPM beta of the stock is actually 1.020, not Since is so close to 1.000, it is unlikely that we will reject the null hypothesis, even though it is false. Hypothesis testing focuses on Type 1 errors. Perhaps the null hypothesis is true, but because of random fluctuations, we discard it. We avoid this by choosing a significance level like 5% or 1%, so there is only a small probability of discarding the null hypothesis when it is true. Section discusses p-values, a better way of stating the conclusion of hypothesis testing. Suppose we reject the null hypothesis at a 10% or 5% significance level, but not at a 2% or 1% significance level. This leaves us wondering: what about at a 3% or 4%

20 significance level? A p-value tells us the exact cut-off, such as 2.2% or 4.7%. A p-value of 2.2% makes us more confident that the null hypothesis is false than a p-value of 4.7%. Section 2.5.3, on page 43-45, shows the inter-relation of sample size and hypothesis testing; see Example 2.4. A larger sample makes a Type 1 error less common. Jacob: Does a larger sample size change the expected value of the regression coefficient? Rachel: No; the regression coefficient is unbiased regardless of the sample size. A larger sample reduces the variance of the ordinary least squares estimator of the coefficient. Jacob: Suppose that with a sample of 100, the result is significant at a 10% level. With a sample of 400, do we expect the result to be significant at a 2.5% level or a 5% level? Rachel: The question is incomplete. With a sample of 100, random fluctuations are more likely than with a sample of 400, so we can t compare expected significance. Both the number of observations and the size of the result affect the significance. Jacob: Let s change the question: If we have the same ordinary least squares estimator with a sample of 400, what is the expected significance of the result? Rachel: There is no simple relation; we must look up the answer in a cumulative normal distribution table. A larger sample size makes the confidence interval narrower. The final exam may give a similar scenario as the exercise in the textbook, using pass ratios of male and female candidates on Course C. Section 2.6 is not tested on the final exam, and it has no homework assignments. The material is useful for actuaries, who must present actuarial results to company officers. Historgrams and other graphics are useful, since they help non-actuaries interpret our results. But it is hard to test histograms on a final exam. Appendix 2.1 is simple. We assume you know this material. It is not tested per se on the final exam, and it is not used in the homework assignments. But these results are used throughout the course. We use results 1 through 8 all the time; you are expected to know them on any actuarial exam you take. Result 9 says that the sample variance is an unbiased estimate of the population variance. You are expected to know this, though you do not have to know the proof. Appendix 2.2 deals with maximum likelihood estimation. You learn this Module 21. It is not worth reading now; read it 6 weeks from now. This module has the most pages of any module in this course, but it is background. Most items that you must know are repeated in later modules.

21 Regression Analysis and Time Series, Module 2: Statistical Properties Intuition: Population Variance and Sample Variance (The attached PDF file has better formatting.) Updated: May 25, 2005 Throughout this course, we use sample variances as estimates of population variances. The population variance is divided by N, the number of equally likely scenarios; the sample variance is divided by N-1, or the number of data points minus one. 2 2 The sample variance (s ) is an unbiased estimator of the population variance ( ). Jacob and Rachel are discussing the relation of the sample variance to the population variance. Jacob: Suppose we have a sample of two points: 1 and +1. The mean is zero, and the 2 2 variation is ( 1 0) + (+1 0) = 2. The variation, or total sum of squares (TSS), is the sum of the squares of the deviations of each point from the mean. The sample variance is TSS / (N 1) = 2 / (2 1) = 2. The simplest hypothesis is that these two points come from a population of two points, 1 and +1, each with a 50% chance of occurring. The population variance of this distribution is TSS / N = 2 / 2 = 1. How can we say that the sample variance is an unbiased estimator of the population variance? Rachel: If the population has a distribution of two values, 1 and +1, each with a 50% chance of occurring, there are four equally like samples: ( 1, +1), (+1, 1), ( 1, 1), and (+1, +1). We compute the sample variances of each sample as well as the (incorrect) variances using N as the denominator instead of N-1. Pt A Pt B Mean Deviations Total Variation Sample Variance Population Variance , = , +1 (-1) + 1 = , (-1) = , = Average 1 ½ The true variance of the distribution is 1, not ½. By using the sample variance of the four samples, the estimated population variance is 1, not ½. Jacob: Can you show this for other distributions?

22 Rachel: When the distributions have more points, it is harder to give illustrations. We can prove this result, though the proof is not required for the regression analysis course. (The proof is in the textbook.) Let us look at a distribution with three points. Suppose a distribution has three possible values, 3, 0, +3, each with a probability of a. The mean of the distribution is zero, and the population variance of the distribution is { (3 0) + (0 0) + (3 0 ) } / 3 = 18 / 3 = 6. 2 If we draw a sample of three points and they are ( 3, 0, +3), the sample variance is ( ) / 2 = 18 / 2 = 9. But there are 27 equally likely samples of three values: Points Deviations A B C Mean A B C Total Variation Sample Variance Pop Variance Average

23 The table shows the 27 possible samples, means, deviations, total variation, sample variance, the incorrect sample variance using N as the denominator instead of (N-1), and the averages. The sample variance is an unbiased estimator of the population variance. {Recommendation: If you are not comfortable with sample and population variances, redo these examples with other numbers. Later in the course, you should redo the example using a regression equation with simple data points. Simple illustrations help you master the theory of this course.}

24 Regression Analysis and Time Series, Module 2: Means and Variances (The attached PDF file has better formatting.) Practice Problems Updated: December 11, 2006 Exercise 2.1: Means Given the sample below, what are the estimated means of X and Y? i Y X i i Solution 2.1: = = ( ) / 4 = 13.5 = = ( ) / 4 = 2.05 Exercise 2.2: Deviations What are x i and y, i the deviations of X i and Y? i Solution 2.2: Table 1.1 i y x i i Exercise 2.3: Sample Variances

25 What are the sample variances of X and Y? Solution 2.3: s (X) = = [( 0.5) ( 0.1) ] / 3 = s (Y) = = [( 0.15) ( 0.05) ] / 3 = An alternative method is = = s (X) = = ( ) / 3 = The same alternative formula can be used for the variance of Y. Jacob: Are the two formulas equivalent? 2 2 Rachel: The deviation x i is X i. The square of the deviation is X i 2 X i +. We do this for each of the N values are sum them to get X 2 X + N. 2 2 i i X = N, so this expression simplifies to X N. i We divide by N 1, since this is a sample. Jacob: Which method is easier to use? 2 2 i Rachel: If the problem gives you the mean and the sum of the squares, use the alternative method. Exercise 2.4: Covariance What is estimated covariance between the two random variables? Solution 2.4: = = The estimated covariance of (x,y) is

26 = 0.16 / 3 = Exercise 2.5: Correlation What is the correlation between the two random variables? Solution 2.5: ½ = = 0.16 / ( ) =

27 Regression Analysis and Time Series, Module 2: Means and Variances (The attached PDF file has better formatting.) Homework Assignment Updated: May 25, 2005 We use the sample in the table below. i Y X i i A. What are the estimated means of X and Y? B. What are x i and y, i the deviations of X i and Y? i C. What are the sample variances of X and Y? D. What is estimated covariance between the two random variables? E. What is the correlation between the two random variables? {The homework assignment reviews the material in the practice problems.}

28 Time Series, Module 3, Simple Extrapolation Models Required Reading (The attached PDF file has better formatting.) Updated: May 27, 2005 Read section 15.1 on pages ; this is an introduction. Focus on the difference between deterministic and stochastic models. The authors say at the bottom of page 467: These models are deterministic in that no reference is made to the sources or nature of the underlying randomness. Jacob: This definition seems convoluted. Why not say that a deterministic model gives a point estimate and a stochastic model gives a range of values? Rachel: Stochastic models also give point estimates, as the expected values of the ranges. Deterministic models also give ranges, as distributions about a point estimate. Jacob: What do the authors mean by the sources or nature of the underlying randomness? Rachel: Suppose auto insurance average claim severity is $10,000 in 20X0. We contrast a deterministic exponential trend model with a stochastic autoregressive model to predict future average claim severities. The deterministic exponential trend model says that the expected average claim severity 0.08t is $10,000 e, where t is the number of years after 20X0 and 8% is the continuously compounded loss cost trend. For example, the expected average claim severity for 20X is $10,000 e = $17,507. This is the best estimate of the average claim severity. The actual average claim severity will not be exactly this amount, but the linear trend model does not suggest a probability distribution for the claim severity. Jacob: If we add a probability distribution about the estimate, such as a normal distribution with a standard deviation of $1,000, is the trend model stochastic? Rachel: This distribution is ad hoc; it does not explain what the sources or nature of the uncertainty (randomness). Contrast a stochastic autoregressive model, where Y t+1 = e 0.08 Y t + t. If we know the distribution of the error term t, we can derive the probability distribution of Y t+k for any projection distance k. We vary the form of the model in several ways, each with its own explanation of the source of the uncertainty in the projection.! moving average vs autoregressive models (MA vs AR)! one period lags vs higher order lags! stationary vs homogeneous non-stationary models! combined ARMA or ARIMA models

29 Later modules discuss the sources of stochasticity for moving average vs autoregressive models of different lags and for stationary vs homogeneous non-stationary models. Read section on pages Know equations 15.2 (linear trend), 15.4 and its extension into 15.7 (exponential growth), 15.8 (autoregressive trend), 15.9 (logarithmic autoregressive trend); these models will be tested on the final exam. Jacob: Is the autoregressive trend model stochastic, as you describe above? Rachel: An autoregressive model can be stochastic or deterministic. The authors say autoregressive trend for the deterministic model and AR(p) for the stochastic model. The deterministic model has no error term. An actuary using a deterministic model knows that there is uncertainty, but the model does not quantify it. Skip the material on quadratic trend (15.10), logistic curve (15.11), and sales saturation model (15.12 and 15.13). These models are not used by actuaries. In contrast, the first four deterministic models are commonly used by actuaries to price insurance products. {Note: All the deterministic models, including quadratic trend, logistic curves, and sales saturation models are on the CAS transition exam. If you have questions about these models, post them on the discussion board, though we cannot promise that our faculty will have the time to answer these questions.} Example 15.1 shows the use of the models. Readers of Part 4 are assumed to be familiar with the regression techniques in Parts 1 and 2. You will not be asked to do a regression 2 on the time series final exam, but you must know how to evaluate a regression, using R, 2 adjusted R, and the Durbin-Watson statistic. See the examples in Modules Jacob: I took regression several years ago in college. We did not cover the adjusted R 2 or the Durbin-Watson statistic. What should I do? 2 2 Rachel: Review regression analysis, module 4 for R and adjusted R. Understand degrees of freedom; see especially the postings in Regression, Module 4. When we come to the Durbin-Watson statistic, we specify what you should know. 2 For the four models in Example 15.1, the R shows the percentage of the variance explained by the regression equation.! The autoregressive trend models are better than the linear trend models; actuaries use autoregressive trend models.! The logarithmic trend model is better than the linear trend model; actuaries use 2 logarithmic trend models. For the autoregressive models, the R is not materially different between the linear and logarithmic model. The F statistic, which says whether we should reject the null hypothesis that there is no trend, is significant for all four models. A higher F statistic means more significant (i.e., a

Read Section 1.1, Examples of time series, on pages 1-8. These example introduce the book; you are not tested on them.

Read Section 1.1, Examples of time series, on pages 1-8. These example introduce the book; you are not tested on them. TS Module 1 Time series overview (The attached PDF file has better formatting.)! Model building! Time series plots Read Section 1.1, Examples of time series, on pages 1-8. These example introduce the book;

More information

Regression of Time Series

Regression of Time Series Mahlerʼs Guide to Regression of Time Series CAS Exam S prepared by Howard C. Mahler, FCAS Copyright 2016 by Howard C. Mahler. Study Aid 2016F-S-9Supplement Howard Mahler hmahler@mac.com www.howardmahler.com/teaching

More information

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages: Glossary The ISI glossary of statistical terms provides definitions in a number of different languages: http://isi.cbs.nl/glossary/index.htm Adjusted r 2 Adjusted R squared measures the proportion of the

More information

appstats27.notebook April 06, 2017

appstats27.notebook April 06, 2017 Chapter 27 Objective Students will conduct inference on regression and analyze data to write a conclusion. Inferences for Regression An Example: Body Fat and Waist Size pg 634 Our chapter example revolves

More information

Keller: Stats for Mgmt & Econ, 7th Ed July 17, 2006

Keller: Stats for Mgmt & Econ, 7th Ed July 17, 2006 Chapter 17 Simple Linear Regression and Correlation 17.1 Regression Analysis Our problem objective is to analyze the relationship between interval variables; regression analysis is the first tool we will

More information

The Simple Linear Regression Model

The Simple Linear Regression Model The Simple Linear Regression Model Lesson 3 Ryan Safner 1 1 Department of Economics Hood College ECON 480 - Econometrics Fall 2017 Ryan Safner (Hood College) ECON 480 - Lesson 3 Fall 2017 1 / 77 Bivariate

More information

STOCKHOLM UNIVERSITY Department of Economics Course name: Empirical Methods Course code: EC40 Examiner: Lena Nekby Number of credits: 7,5 credits Date of exam: Saturday, May 9, 008 Examination time: 3

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

Institute of Actuaries of India

Institute of Actuaries of India Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics For 2018 Examinations Subject CT3 Probability and Mathematical Statistics Core Technical Syllabus 1 June 2017 Aim The

More information

Regression, part II. I. What does it all mean? A) Notice that so far all we ve done is math.

Regression, part II. I. What does it all mean? A) Notice that so far all we ve done is math. Regression, part II I. What does it all mean? A) Notice that so far all we ve done is math. 1) One can calculate the Least Squares Regression Line for anything, regardless of any assumptions. 2) But, if

More information

2 Prediction and Analysis of Variance

2 Prediction and Analysis of Variance 2 Prediction and Analysis of Variance Reading: Chapters and 2 of Kennedy A Guide to Econometrics Achen, Christopher H. Interpreting and Using Regression (London: Sage, 982). Chapter 4 of Andy Field, Discovering

More information

Chapter 16. Simple Linear Regression and Correlation

Chapter 16. Simple Linear Regression and Correlation Chapter 16 Simple Linear Regression and Correlation 16.1 Regression Analysis Our problem objective is to analyze the relationship between interval variables; regression analysis is the first tool we will

More information

at least 50 and preferably 100 observations should be available to build a proper model

at least 50 and preferably 100 observations should be available to build a proper model III Box-Jenkins Methods 1. Pros and Cons of ARIMA Forecasting a) need for data at least 50 and preferably 100 observations should be available to build a proper model used most frequently for hourly or

More information

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics A short review of the principles of mathematical statistics (or, what you should have learned in EC 151).

More information

review session gov 2000 gov 2000 () review session 1 / 38

review session gov 2000 gov 2000 () review session 1 / 38 review session gov 2000 gov 2000 () review session 1 / 38 Overview Random Variables and Probability Univariate Statistics Bivariate Statistics Multivariate Statistics Causal Inference gov 2000 () review

More information

Business Statistics. Lecture 10: Course Review

Business Statistics. Lecture 10: Course Review Business Statistics Lecture 10: Course Review 1 Descriptive Statistics for Continuous Data Numerical Summaries Location: mean, median Spread or variability: variance, standard deviation, range, percentiles,

More information

An Analysis of College Algebra Exam Scores December 14, James D Jones Math Section 01

An Analysis of College Algebra Exam Scores December 14, James D Jones Math Section 01 An Analysis of College Algebra Exam s December, 000 James D Jones Math - Section 0 An Analysis of College Algebra Exam s Introduction Students often complain about a test being too difficult. Are there

More information

Chapter 27 Summary Inferences for Regression

Chapter 27 Summary Inferences for Regression Chapter 7 Summary Inferences for Regression What have we learned? We have now applied inference to regression models. Like in all inference situations, there are conditions that we must check. We can test

More information

Machine Learning, Fall 2009: Midterm

Machine Learning, Fall 2009: Midterm 10-601 Machine Learning, Fall 009: Midterm Monday, November nd hours 1. Personal info: Name: Andrew account: E-mail address:. You are permitted two pages of notes and a calculator. Please turn off all

More information

THE ROYAL STATISTICAL SOCIETY 2008 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE (MODULAR FORMAT) MODULE 4 LINEAR MODELS

THE ROYAL STATISTICAL SOCIETY 2008 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE (MODULAR FORMAT) MODULE 4 LINEAR MODELS THE ROYAL STATISTICAL SOCIETY 008 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE (MODULAR FORMAT) MODULE 4 LINEAR MODELS The Society provides these solutions to assist candidates preparing for the examinations

More information

Descriptive Statistics-I. Dr Mahmoud Alhussami

Descriptive Statistics-I. Dr Mahmoud Alhussami Descriptive Statistics-I Dr Mahmoud Alhussami Biostatistics What is the biostatistics? A branch of applied math. that deals with collecting, organizing and interpreting data using well-defined procedures.

More information

Midterm 2 - Solutions

Midterm 2 - Solutions Ecn 102 - Analysis of Economic Data University of California - Davis February 23, 2010 Instructor: John Parman Midterm 2 - Solutions You have until 10:20am to complete this exam. Please remember to put

More information

I used college textbooks because they were the only resource available to evaluate measurement uncertainty calculations.

I used college textbooks because they were the only resource available to evaluate measurement uncertainty calculations. Introduction to Statistics By Rick Hogan Estimating uncertainty in measurement requires a good understanding of Statistics and statistical analysis. While there are many free statistics resources online,

More information

The SOA requires independent student projects for the regression analysis and time series courses.

The SOA requires independent student projects for the regression analysis and time series courses. TIME SERIES STUDENT PROJECTS: TIME SERIES TECHNIQUES (The attached PDF file has better formatting.) Updated: May 1, 2008 The SOA requires independent student projects for the regression analysis and time

More information

Chapter 16. Simple Linear Regression and dcorrelation

Chapter 16. Simple Linear Regression and dcorrelation Chapter 16 Simple Linear Regression and dcorrelation 16.1 Regression Analysis Our problem objective is to analyze the relationship between interval variables; regression analysis is the first tool we will

More information

Chapter 1 Statistical Inference

Chapter 1 Statistical Inference Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations

More information

Chapter 3. Introduction to Linear Correlation and Regression Part 3

Chapter 3. Introduction to Linear Correlation and Regression Part 3 Tuesday, December 12, 2000 Ch3 Intro Correlation Pt 3 Page: 1 Richard Lowry, 1999-2000 All rights reserved. Chapter 3. Introduction to Linear Correlation and Regression Part 3 Regression The appearance

More information

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis. 401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis

More information

Last two weeks: Sample, population and sampling distributions finished with estimation & confidence intervals

Last two weeks: Sample, population and sampling distributions finished with estimation & confidence intervals Past weeks: Measures of central tendency (mean, mode, median) Measures of dispersion (standard deviation, variance, range, etc). Working with the normal curve Last two weeks: Sample, population and sampling

More information

Warm-up Using the given data Create a scatterplot Find the regression line

Warm-up Using the given data Create a scatterplot Find the regression line Time at the lunch table Caloric intake 21.4 472 30.8 498 37.7 335 32.8 423 39.5 437 22.8 508 34.1 431 33.9 479 43.8 454 42.4 450 43.1 410 29.2 504 31.3 437 28.6 489 32.9 436 30.6 480 35.1 439 33.0 444

More information

Introduction to Regression Analysis. Dr. Devlina Chatterjee 11 th August, 2017

Introduction to Regression Analysis. Dr. Devlina Chatterjee 11 th August, 2017 Introduction to Regression Analysis Dr. Devlina Chatterjee 11 th August, 2017 What is regression analysis? Regression analysis is a statistical technique for studying linear relationships. One dependent

More information

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals Past weeks: Measures of central tendency (mean, mode, median) Measures of dispersion (standard deviation, variance, range, etc). Working with the normal curve Last week: Sample, population and sampling

More information

16.400/453J Human Factors Engineering. Design of Experiments II

16.400/453J Human Factors Engineering. Design of Experiments II J Human Factors Engineering Design of Experiments II Review Experiment Design and Descriptive Statistics Research question, independent and dependent variables, histograms, box plots, etc. Inferential

More information

where Female = 0 for males, = 1 for females Age is measured in years (22, 23, ) GPA is measured in units on a four-point scale (0, 1.22, 3.45, etc.

where Female = 0 for males, = 1 for females Age is measured in years (22, 23, ) GPA is measured in units on a four-point scale (0, 1.22, 3.45, etc. Notes on regression analysis 1. Basics in regression analysis key concepts (actual implementation is more complicated) A. Collect data B. Plot data on graph, draw a line through the middle of the scatter

More information

Review of Statistics 101

Review of Statistics 101 Review of Statistics 101 We review some important themes from the course 1. Introduction Statistics- Set of methods for collecting/analyzing data (the art and science of learning from data). Provides methods

More information

Do not copy, post, or distribute

Do not copy, post, or distribute 14 CORRELATION ANALYSIS AND LINEAR REGRESSION Assessing the Covariability of Two Quantitative Properties 14.0 LEARNING OBJECTIVES In this chapter, we discuss two related techniques for assessing a possible

More information

GRE Quantitative Reasoning Practice Questions

GRE Quantitative Reasoning Practice Questions GRE Quantitative Reasoning Practice Questions y O x 7. The figure above shows the graph of the function f in the xy-plane. What is the value of f (f( ))? A B C 0 D E Explanation Note that to find f (f(

More information

1 Random walks and data

1 Random walks and data Inference, Models and Simulation for Complex Systems CSCI 7-1 Lecture 7 15 September 11 Prof. Aaron Clauset 1 Random walks and data Supposeyou have some time-series data x 1,x,x 3,...,x T and you want

More information

Econometrics Part Three

Econometrics Part Three !1 I. Heteroskedasticity A. Definition 1. The variance of the error term is correlated with one of the explanatory variables 2. Example -- the variance of actual spending around the consumption line increases

More information

Plotting data is one method for selecting a probability distribution. The following

Plotting data is one method for selecting a probability distribution. The following Advanced Analytical Models: Over 800 Models and 300 Applications from the Basel II Accord to Wall Street and Beyond By Johnathan Mun Copyright 008 by Johnathan Mun APPENDIX C Understanding and Choosing

More information

EC4051 Project and Introductory Econometrics

EC4051 Project and Introductory Econometrics EC4051 Project and Introductory Econometrics Dudley Cooke Trinity College Dublin Dudley Cooke (Trinity College Dublin) Intro to Econometrics 1 / 23 Project Guidelines Each student is required to undertake

More information

Inference with Simple Regression

Inference with Simple Regression 1 Introduction Inference with Simple Regression Alan B. Gelder 06E:071, The University of Iowa 1 Moving to infinite means: In this course we have seen one-mean problems, twomean problems, and problems

More information

Statistics 251: Statistical Methods

Statistics 251: Statistical Methods Statistics 251: Statistical Methods 1-sample Hypothesis Tests Module 9 2018 Introduction We have learned about estimating parameters by point estimation and interval estimation (specifically confidence

More information

t-test for b Copyright 2000 Tom Malloy. All rights reserved. Regression

t-test for b Copyright 2000 Tom Malloy. All rights reserved. Regression t-test for b Copyright 2000 Tom Malloy. All rights reserved. Regression Recall, back some time ago, we used a descriptive statistic which allowed us to draw the best fit line through a scatter plot. We

More information

Quantitative Methods for Economics, Finance and Management (A86050 F86050)

Quantitative Methods for Economics, Finance and Management (A86050 F86050) Quantitative Methods for Economics, Finance and Management (A86050 F86050) Matteo Manera matteo.manera@unimib.it Marzio Galeotti marzio.galeotti@unimi.it 1 This material is taken and adapted from Guy Judge

More information

ECON 4230 Intermediate Econometric Theory Exam

ECON 4230 Intermediate Econometric Theory Exam ECON 4230 Intermediate Econometric Theory Exam Multiple Choice (20 pts). Circle the best answer. 1. The Classical assumption of mean zero errors is satisfied if the regression model a) is linear in the

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

Harvard University. Rigorous Research in Engineering Education

Harvard University. Rigorous Research in Engineering Education Statistical Inference Kari Lock Harvard University Department of Statistics Rigorous Research in Engineering Education 12/3/09 Statistical Inference You have a sample and want to use the data collected

More information

Statistics for Managers using Microsoft Excel 6 th Edition

Statistics for Managers using Microsoft Excel 6 th Edition Statistics for Managers using Microsoft Excel 6 th Edition Chapter 13 Simple Linear Regression 13-1 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value of

More information

Volatility. Gerald P. Dwyer. February Clemson University

Volatility. Gerald P. Dwyer. February Clemson University Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use

More information

SOME BASICS OF TIME-SERIES ANALYSIS

SOME BASICS OF TIME-SERIES ANALYSIS SOME BASICS OF TIME-SERIES ANALYSIS John E. Floyd University of Toronto December 8, 26 An excellent place to learn about time series analysis is from Walter Enders textbook. For a basic understanding of

More information

Error Analysis in Experimental Physical Science Mini-Version

Error Analysis in Experimental Physical Science Mini-Version Error Analysis in Experimental Physical Science Mini-Version by David Harrison and Jason Harlow Last updated July 13, 2012 by Jason Harlow. Original version written by David M. Harrison, Department of

More information

Midterm 2 - Solutions

Midterm 2 - Solutions Ecn 102 - Analysis of Economic Data University of California - Davis February 24, 2010 Instructor: John Parman Midterm 2 - Solutions You have until 10:20am to complete this exam. Please remember to put

More information

y response variable x 1, x 2,, x k -- a set of explanatory variables

y response variable x 1, x 2,, x k -- a set of explanatory variables 11. Multiple Regression and Correlation y response variable x 1, x 2,, x k -- a set of explanatory variables In this chapter, all variables are assumed to be quantitative. Chapters 12-14 show how to incorporate

More information

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n = Hypothesis testing I I. What is hypothesis testing? [Note we re temporarily bouncing around in the book a lot! Things will settle down again in a week or so] - Exactly what it says. We develop a hypothesis,

More information

Algebra Year 10. Language

Algebra Year 10. Language Algebra Year 10 Introduction In Algebra we do Maths with numbers, but some of those numbers are not known. They are represented with letters, and called unknowns, variables or, most formally, literals.

More information

Probability Distributions

Probability Distributions CONDENSED LESSON 13.1 Probability Distributions In this lesson, you Sketch the graph of the probability distribution for a continuous random variable Find probabilities by finding or approximating areas

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators

More information

Linear Regression with 1 Regressor. Introduction to Econometrics Spring 2012 Ken Simons

Linear Regression with 1 Regressor. Introduction to Econometrics Spring 2012 Ken Simons Linear Regression with 1 Regressor Introduction to Econometrics Spring 2012 Ken Simons Linear Regression with 1 Regressor 1. The regression equation 2. Estimating the equation 3. Assumptions required for

More information

1 Least Squares Estimation - multiple regression.

1 Least Squares Estimation - multiple regression. Introduction to multiple regression. Fall 2010 1 Least Squares Estimation - multiple regression. Let y = {y 1,, y n } be a n 1 vector of dependent variable observations. Let β = {β 0, β 1 } be the 2 1

More information

Solutions to the Spring 2015 CAS Exam ST

Solutions to the Spring 2015 CAS Exam ST Solutions to the Spring 2015 CAS Exam ST (updated to include the CAS Final Answer Key of July 15) There were 25 questions in total, of equal value, on this 2.5 hour exam. There was a 10 minute reading

More information

Practice Problems Section Problems

Practice Problems Section Problems Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,

More information

1 Measurement Uncertainties

1 Measurement Uncertainties 1 Measurement Uncertainties (Adapted stolen, really from work by Amin Jaziri) 1.1 Introduction No measurement can be perfectly certain. No measuring device is infinitely sensitive or infinitely precise.

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)

More information

Confidence intervals

Confidence intervals Confidence intervals We now want to take what we ve learned about sampling distributions and standard errors and construct confidence intervals. What are confidence intervals? Simply an interval for which

More information

Section 3: Simple Linear Regression

Section 3: Simple Linear Regression Section 3: Simple Linear Regression Carlos M. Carvalho The University of Texas at Austin McCombs School of Business http://faculty.mccombs.utexas.edu/carlos.carvalho/teaching/ 1 Regression: General Introduction

More information

AP Final Review II Exploring Data (20% 30%)

AP Final Review II Exploring Data (20% 30%) AP Final Review II Exploring Data (20% 30%) Quantitative vs Categorical Variables Quantitative variables are numerical values for which arithmetic operations such as means make sense. It is usually a measure

More information

Correlation Analysis

Correlation Analysis Simple Regression Correlation Analysis Correlation analysis is used to measure strength of the association (linear relationship) between two variables Correlation is only concerned with strength of the

More information

Chapter 11 Sampling Distribution. Stat 115

Chapter 11 Sampling Distribution. Stat 115 Chapter 11 Sampling Distribution Stat 115 1 Definition 11.1 : Random Sample (finite population) Suppose we select n distinct elements from a population consisting of N elements, using a particular probability

More information

Business Statistics. Lecture 9: Simple Regression

Business Statistics. Lecture 9: Simple Regression Business Statistics Lecture 9: Simple Regression 1 On to Model Building! Up to now, class was about descriptive and inferential statistics Numerical and graphical summaries of data Confidence intervals

More information

Economics 308: Econometrics Professor Moody

Economics 308: Econometrics Professor Moody Economics 308: Econometrics Professor Moody References on reserve: Text Moody, Basic Econometrics with Stata (BES) Pindyck and Rubinfeld, Econometric Models and Economic Forecasts (PR) Wooldridge, Jeffrey

More information

ECON 497 Midterm Spring

ECON 497 Midterm Spring ECON 497 Midterm Spring 2009 1 ECON 497: Economic Research and Forecasting Name: Spring 2009 Bellas Midterm You have three hours and twenty minutes to complete this exam. Answer all questions and explain

More information

Math 5a Reading Assignments for Sections

Math 5a Reading Assignments for Sections Math 5a Reading Assignments for Sections 4.1 4.5 Due Dates for Reading Assignments Note: There will be a very short online reading quiz (WebWork) on each reading assignment due one hour before class on

More information

Marquette University Executive MBA Program Statistics Review Class Notes Summer 2018

Marquette University Executive MBA Program Statistics Review Class Notes Summer 2018 Marquette University Executive MBA Program Statistics Review Class Notes Summer 2018 Chapter One: Data and Statistics Statistics A collection of procedures and principles

More information

Subject CS1 Actuarial Statistics 1 Core Principles

Subject CS1 Actuarial Statistics 1 Core Principles Institute of Actuaries of India Subject CS1 Actuarial Statistics 1 Core Principles For 2019 Examinations Aim The aim of the Actuarial Statistics 1 subject is to provide a grounding in mathematical and

More information

79 Wyner Math Academy I Spring 2016

79 Wyner Math Academy I Spring 2016 79 Wyner Math Academy I Spring 2016 CHAPTER NINE: HYPOTHESIS TESTING Review May 11 Test May 17 Research requires an understanding of underlying mathematical distributions as well as of the research methods

More information

ACTEX CAS EXAM 3 STUDY GUIDE FOR MATHEMATICAL STATISTICS

ACTEX CAS EXAM 3 STUDY GUIDE FOR MATHEMATICAL STATISTICS ACTEX CAS EXAM 3 STUDY GUIDE FOR MATHEMATICAL STATISTICS TABLE OF CONTENTS INTRODUCTORY NOTE NOTES AND PROBLEM SETS Section 1 - Point Estimation 1 Problem Set 1 15 Section 2 - Confidence Intervals and

More information

MATH 341, Section 001 FALL 2014 Introduction to the Language and Practice of Mathematics

MATH 341, Section 001 FALL 2014 Introduction to the Language and Practice of Mathematics MATH 341, Section 001 FALL 2014 Introduction to the Language and Practice of Mathematics Class Meetings: MW 9:30-10:45 am in EMS E424A, September 3 to December 10 [Thanksgiving break November 26 30; final

More information

Data Science for Engineers Department of Computer Science and Engineering Indian Institute of Technology, Madras

Data Science for Engineers Department of Computer Science and Engineering Indian Institute of Technology, Madras Data Science for Engineers Department of Computer Science and Engineering Indian Institute of Technology, Madras Lecture 36 Simple Linear Regression Model Assessment So, welcome to the second lecture on

More information

Basic Probability Reference Sheet

Basic Probability Reference Sheet February 27, 2001 Basic Probability Reference Sheet 17.846, 2001 This is intended to be used in addition to, not as a substitute for, a textbook. X is a random variable. This means that X is a variable

More information

Final Exam - Solutions

Final Exam - Solutions Ecn 102 - Analysis of Economic Data University of California - Davis March 19, 2010 Instructor: John Parman Final Exam - Solutions You have until 5:30pm to complete this exam. Please remember to put your

More information

MATH 1070 Introductory Statistics Lecture notes Relationships: Correlation and Simple Regression

MATH 1070 Introductory Statistics Lecture notes Relationships: Correlation and Simple Regression MATH 1070 Introductory Statistics Lecture notes Relationships: Correlation and Simple Regression Objectives: 1. Learn the concepts of independent and dependent variables 2. Learn the concept of a scatterplot

More information

Solving Equations by Adding and Subtracting

Solving Equations by Adding and Subtracting SECTION 2.1 Solving Equations by Adding and Subtracting 2.1 OBJECTIVES 1. Determine whether a given number is a solution for an equation 2. Use the addition property to solve equations 3. Determine whether

More information

Simple Linear Regression: One Quantitative IV

Simple Linear Regression: One Quantitative IV Simple Linear Regression: One Quantitative IV Linear regression is frequently used to explain variation observed in a dependent variable (DV) with theoretically linked independent variables (IV). For example,

More information

Multiple Regression Theory 2006 Samuel L. Baker

Multiple Regression Theory 2006 Samuel L. Baker MULTIPLE REGRESSION THEORY 1 Multiple Regression Theory 2006 Samuel L. Baker Multiple regression is regression with two or more independent variables on the right-hand side of the equation. Use multiple

More information

Final Exam Bus 320 Spring 2000 Russell

Final Exam Bus 320 Spring 2000 Russell Name Final Exam Bus 320 Spring 2000 Russell Do not turn over this page until you are told to do so. You will have 3 hours minutes to complete this exam. The exam has a total of 100 points and is divided

More information

Lecture 8. Using the CLR Model. Relation between patent applications and R&D spending. Variables

Lecture 8. Using the CLR Model. Relation between patent applications and R&D spending. Variables Lecture 8. Using the CLR Model Relation between patent applications and R&D spending Variables PATENTS = No. of patents (in 000) filed RDEP = Expenditure on research&development (in billions of 99 $) The

More information

FinQuiz Notes

FinQuiz Notes Reading 10 Multiple Regression and Issues in Regression Analysis 2. MULTIPLE LINEAR REGRESSION Multiple linear regression is a method used to model the linear relationship between a dependent variable

More information

Stochastic Processes

Stochastic Processes Stochastic Processes Stochastic Process Non Formal Definition: Non formal: A stochastic process (random process) is the opposite of a deterministic process such as one defined by a differential equation.

More information

Algebra Exam. Solutions and Grading Guide

Algebra Exam. Solutions and Grading Guide Algebra Exam Solutions and Grading Guide You should use this grading guide to carefully grade your own exam, trying to be as objective as possible about what score the TAs would give your responses. Full

More information

Chapter 9: Roots and Irrational Numbers

Chapter 9: Roots and Irrational Numbers Chapter 9: Roots and Irrational Numbers Index: A: Square Roots B: Irrational Numbers C: Square Root Functions & Shifting D: Finding Zeros by Completing the Square E: The Quadratic Formula F: Quadratic

More information

Chapter 11. Correlation and Regression

Chapter 11. Correlation and Regression Chapter 11. Correlation and Regression The word correlation is used in everyday life to denote some form of association. We might say that we have noticed a correlation between foggy days and attacks of

More information

Final Exam. Name: Solution:

Final Exam. Name: Solution: Final Exam. Name: Instructions. Answer all questions on the exam. Open books, open notes, but no electronic devices. The first 13 problems are worth 5 points each. The rest are worth 1 point each. HW1.

More information

df=degrees of freedom = n - 1

df=degrees of freedom = n - 1 One sample t-test test of the mean Assumptions: Independent, random samples Approximately normal distribution (from intro class: σ is unknown, need to calculate and use s (sample standard deviation)) Hypotheses:

More information

Ch3. TRENDS. Time Series Analysis

Ch3. TRENDS. Time Series Analysis 3.1 Deterministic Versus Stochastic Trends The simulated random walk in Exhibit 2.1 shows a upward trend. However, it is caused by a strong correlation between the series at nearby time points. The true

More information

Eco517 Fall 2004 C. Sims MIDTERM EXAM

Eco517 Fall 2004 C. Sims MIDTERM EXAM Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering

More information

CHAPTER 21: TIME SERIES ECONOMETRICS: SOME BASIC CONCEPTS

CHAPTER 21: TIME SERIES ECONOMETRICS: SOME BASIC CONCEPTS CHAPTER 21: TIME SERIES ECONOMETRICS: SOME BASIC CONCEPTS 21.1 A stochastic process is said to be weakly stationary if its mean and variance are constant over time and if the value of the covariance between

More information

Lectures 5 & 6: Hypothesis Testing

Lectures 5 & 6: Hypothesis Testing Lectures 5 & 6: Hypothesis Testing in which you learn to apply the concept of statistical significance to OLS estimates, learn the concept of t values, how to use them in regression work and come across

More information

Purposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions

Purposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions Part 1: Probability Distributions Purposes of Data Analysis True Distributions or Relationships in the Earths System Probability Distribution Normal Distribution Student-t Distribution Chi Square Distribution

More information

Marketing Research Session 10 Hypothesis Testing with Simple Random samples (Chapter 12)

Marketing Research Session 10 Hypothesis Testing with Simple Random samples (Chapter 12) Marketing Research Session 10 Hypothesis Testing with Simple Random samples (Chapter 12) Remember: Z.05 = 1.645, Z.01 = 2.33 We will only cover one-sided hypothesis testing (cases 12.3, 12.4.2, 12.5.2,

More information