Review of Statistics

Similar documents
Statistics for Managers Using Microsoft Excel/SPSS Chapter 4 Basic Probability And Discrete Probability Distributions

Statistics for Managers Using Microsoft Excel (3 rd Edition)

2011 Pearson Education, Inc

Correlation Analysis

Applied Econometrics (QEM)

Correlation and Regression

Econometrics Summary Algebraic and Statistical Preliminaries

Semester 2, 2015/2016

Statistics for Managers using Microsoft Excel 6 th Edition

AIM HIGH SCHOOL. Curriculum Map W. 12 Mile Road Farmington Hills, MI (248)

Bivariate distributions

Chapter 14 Simple Linear Regression (A)

Basics of Experimental Design. Review of Statistics. Basic Study. Experimental Design. When an Experiment is Not Possible. Studying Relations

Ch 2: Simple Linear Regression

Lecture 2: Review of Probability

Can you tell the relationship between students SAT scores and their college grades?

Review of Statistics 101

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

Institute of Actuaries of India

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

Chapter 13. Multiple Regression and Model Building

ECON The Simple Regression Model

Sociology 6Z03 Review II

What is a Hypothesis?

Chapte The McGraw-Hill Companies, Inc. All rights reserved.

Inferences for Regression

Simple Linear Regression

What is Probability? Probability. Sample Spaces and Events. Simple Event

Recitation 2: Probability

Basic Business Statistics 6 th Edition

KDF2C QUANTITATIVE TECHNIQUES FOR BUSINESSDECISION. Unit : I - V

Chapter 16. Simple Linear Regression and dcorrelation

Math Review Sheet, Fall 2008

Hypothesis Testing hypothesis testing approach

5.1 Model Specification and Data 5.2 Estimating the Parameters of the Multiple Regression Model 5.3 Sampling Properties of the Least Squares

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

Chapter 2. Probability

Sample Problems. Note: If you find the following statements true, you should briefly prove them. If you find them false, you should correct them.

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

Statistics Introductory Correlation

Probability Theory and Statistics. Peter Jochumzen

Regression Analysis. y t = β 1 x t1 + β 2 x t2 + β k x tk + ϵ t, t = 1,..., T,

Topic 2: Probability & Distributions. Road Map Probability & Distributions. ECO220Y5Y: Quantitative Methods in Economics. Dr.

Test 3 Practice Test A. NOTE: Ignore Q10 (not covered)

Regression Analysis II

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

Chapter 4: Regression Models

LECTURE 5 HYPOTHESIS TESTING

Business Statistics. Lecture 10: Correlation and Linear Regression

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model

Correlation. A statistics method to measure the relationship between two variables. Three characteristics

Ch. 1: Data and Distributions

REVIEW 8/2/2017 陈芳华东师大英语系

4.1 Least Squares Prediction 4.2 Measuring Goodness-of-Fit. 4.3 Modeling Issues. 4.4 Log-Linear Models

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

Economics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects

Multiple Regression Analysis

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES

Econ 325: Introduction to Empirical Economics

Business Statistics. Lecture 9: Simple Regression

LECTURE 5. Introduction to Econometrics. Hypothesis testing

Lecture 14 Simple Linear Regression

Formal Statement of Simple Linear Regression Model

Bias Variance Trade-off

Chapter 3 Multiple Regression Complete Example

Linear Regression. Simple linear regression model determines the relationship between one dependent variable (y) and one independent variable (x).

STA301- Statistics and Probability Solved Subjective From Final term Papers. STA301- Statistics and Probability Final Term Examination - Spring 2012

Origins of Probability Theory

Review of Econometrics

Psychology 282 Lecture #4 Outline Inferences in SLR

Chapter 16. Simple Linear Regression and Correlation

Inference for Regression

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

f X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx

Probability Dr. Manjula Gunarathna 1

Simple Linear Regression: The Model

Scatter plot of data from the study. Linear Regression

ECON3150/4150 Spring 2015

Lecture 2 Simple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: Chapter 1

Lecture 9: Linear Regression

Spatial Regression. 3. Review - OLS and 2SLS. Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Correlation and Linear Regression

Statistical Distribution Assumptions of General Linear Models

Statistics Primer. A Brief Overview of Basic Statistical and Probability Principles. Essential Statistics for Data Analysts Using Excel

Sample Space: Specify all possible outcomes from an experiment. Event: Specify a particular outcome or combination of outcomes.

Mathematics for Economics MA course

Inference for the Regression Coefficient

Keller: Stats for Mgmt & Econ, 7th Ed July 17, 2006

Econometrics. 4) Statistical inference

1 INFO Sep 05

Summary of Chapters 7-9

Course 4 Solutions November 2001 Exams

Appendix A : Introduction to Probability and stochastic processes

02 Background Minimum background on probability. Random process

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber

Subject CS1 Actuarial Statistics 1 Core Principles

Regression and Statistical Inference

Multiple Regression Analysis

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47

Transcription:

Review of Statistics

Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and correlation Central Limit Theorem Hypothesis testing z-test, p-value Simple Linear Regression

Statistical Methods Statistical Methods Descriptive Statistics Inferential Statistics

Descriptive Statistics Involves Collecting Data Presenting Data Characterizing Data Purpose Describe Data 90 80 70 60 50 40 30 20 10 0 1st Qtr 2nd Qtr 3rd Qtr 4th Qtr East West North

Inferential Statistics Involves Estimation Hypothesis Testing Population? Purpose Make Decisions About Population Characteristics

Descriptive Statistics

Mean Measure of central tendency Acts as Balance Point Affected by extreme values ( outliers ) Formula: X n Xi i = = = n X + X + + X 1 1 2 n n

Median Measure of central tendency Middle value in ordered sequence If odd n, Middle Value of Sequence If even n, Average of 2 Middle Values Value that splits the distribution into two halves Not Affected by Extreme Values

Median (Example) Raw Data: 17 16 21 18 13 16 12 11 Ordered: 11 12 13 16 16 17 18 21 Position: 1 2 3 4 5 6 7 8 Median = 16 + 16 2 = 16

Mode Measure of Central Tendency Value That Occurs Most Often Not Affected by Extreme Values There May Be Several Modes Raw Data: 17 16 21 18 13 16 12 11 Ordered: 11 12 13 16 16 17 18 21

Sample Variance S 2 = n i = 1 (X X) i n 1 2 n - 1 in denominator! (Use n if population variance) = (X X) + (X X) + + (X X) 1 2 2 n 1 2 2 n

Sample Standard Deviation S = S 2 = n i = 1 (X X ) i n 1 2 = (X X ) + (X X ) + + (X X ) 1 2 2 n 1 2 2 n

Probability

Event, Sample Space Event: one possible outcome Sample space: collection of all the possible events S = { } Probability of an outcome: proportion of times that the outcome occurs in the long run The complement of event A: includes all the events that are not part of the event A: Symbol A Event A { } Complement of A A { }

Properties of an Event 1. Mutually Exclusive Two outcomes that cannot occur at the same time Experiment: Observe gender of one person 2. Collectively Exhaustive One outcome in sample space must occur

Joint Events Joint event: Event that has two or more characteristics means intersection of event (set) A and event (set) B Example: A and B, (A B): Female, Under age 20

Compound Events Union of event A and event B ( A B ): Total area of the two circles A B contains all the outcomes which are part of event (set) A, part of event (set) B or part of both A and B means union of event A and event B

Compound Probability Addition Rule Used to Get Compound Probabilities for Unions of Events P(A OR B) = P(A B) = P(A) + P(B) - P(A B) For Mutually Exclusive Events: P(A OR B) = P(A B) = P(A) + P(B) Mutually Exclusive Events A B

Random variables Random variable numerical summary of a random outcome a function that assigns a numerical value to each simple event in a sample space Discrete or continuous random variables Discrete: only a discrete set of possible values => summarized by probability distribution: list of all possible values of the variables and the probability that each value will occur. Continuous: continuum of possible values => summarized by the probability density function (pdf)

Discrete Probability Distribution 1. List of pairs [ X i, P(X i ) ] X i = Value of Random Variable (Outcome) P(X i ) = Probability Associated with Value 2. Mutually exclusive (no overlap) 3. Collectively exhaustive (nothing left out) 4. 0 P(X i ) 1 5. Σ P(X i ) = 1

Joint Probability Using Contingency Table Event Event B 1 B 2 Total A 1 P(A1 B1) P(A1 B2) P(A 1 ) A 2 P(A2 B1) P(A2 B2) P(A2) Total P(B 1 ) P(B 2 ) 1 Conditional probability: P( AB1) P( A1 B1) PB ( 1) PA ( 2 B1) PB ( ) 1 1 Joint Probability Joint distribution: Marginal Probability Marginal distributions: Conditional distribution:

Contingency Table Example Joint Event: Draw 1 Card. Note Kind, Color Color Type Red Black Total Ace 2/52 2/52 4/52 Non-Ace 24/52 24/52 48/52 Total 26/52 26/52 52/52 P(Ace) P(Red) P(Ace AND Red)

Moments Discrete Case Moment: Summary of a certain aspect of a distribution Mean, Expected Value Mean of Probability Distribution Weighted Average of All Possible Values μ = E(X) = ΣX i P(X i ) Variance Weighted Average Squared Deviation about Mean σ 2 = E[ (X i μ) 2 ] = Σ (X i μ) 2 P(X i )

Statistical Independence When the outcome of one event (B) does not affect the probability of occurrence of another event (A), the events A and B are said to be statistically independent. Example: Toss a coin twice => no causality Condition for independence: Two events A and B are statistically independent if and only if (iff) P(A B) = P(A)

Bayes Theorem and Multiplication Rule Bayes Theorem P(A B) = P(A B) P(B) The difficult part is P(A B) Use above equation to derive P(A B) P(A and B) = P(A B) = P(A)P(B A) = P(B)P(A B) For independent events: P(A and B) = P(A B) = P(A)P(B)

Covariance Measures the joint variability of two random variables N σ XY = (X i μ X )(Y i μ Y )P(X i, Y i ) i=1 Can take any value in the real numbers Depends on units of measurement (e.g., dollars, cents, billions of dollars) Example: positive covariance = y and x are positively related; when y is above its mean, x tends to be above its mean; when y is below its mean, x tends to be below its mean.

Correlation Standardized covariance, takes values in [-1, 1] Does not depend on unit of measurement Correlation coefficient (ρ) formula: ρ = cov( XY ) σ σ X Y = σ σ X XY σ Covariance and correlation measure only linear dependence! Example: Cov(X,Y)=0 Does not necessarily imply that y and x are independent. They may be non-linearly related. But if X and Y are jointly normally distributed, then they are independent. Y

Sum of Two Random Variables Expected Value of the Sum of Two Random Variables E(X + Y) = E(X) +E(Y) Variance of the Sum of Two Random Variables Var (X + Y) = σ 2 = + σ 2 Y + 2σ XY σ 2 X X+Y

Continuous Probability Distributions - Normal Distribution Bell-Shaped, symmetrical Mean, median, mode are equal Infinite range 68% of the data are within 1 standard deviation of the mean 95% of the data are within 2 standard deviations of the mean In early 1800's, German mathematician and physicist Karl Gauss used it to analyze astronomical data, therefore known as Gaussian distribution. f(x) Mean, Median, Mode X

Normal Distribution Probability Density Function f( X) = 1 2 e π σ 1 2 X μ σ 2 f(x) = frequency of random variable X π = 3.14159; e = 2.71828 σ = population standard deviation X = value of random variable (- < X < ) μ = population mean

Effect of Varying Parameters (μ & σ) f(x) B A C X

Normal Distribution Probability Probability is the area under the curve! d Pc ( X d) = f ( x) dx? c f(x) c d X

Infinite Number of Normal Distribution Tables Normal distributions differ by mean & standard deviation. Each distribution would require its own table. f(x) X That s an infinite number!

Standardize the Normal Distribution Normal Distribution σ Z = X μ σ Standardized Normal Distribution σ z = 1 μ X μ Z = 0 Z One table!

Standardizing Example Normal Distribution σ = 10 Z = X μ = 6.2 5 = 0.12 σ 10 Standardized Normal Distribution σ Z = 1 μ = 5 6.2 X μ Z = 0.12 Z

Moments: Mean, Variance (Continuous Case) Mean, Expected Value Mean of probability distribution Weighted average of all possible values Variance μ = E(X) = X f(x) dx - Weighted average squared deviation about mean σ 2 = E[ (X μ) 2 ] = (X- μ) 2 f(x) dx -

Moments: Skewness, Kurtosis ( ) 3 Skewness: E X μ S = 3 Measures asymmetry in distribution σ The larger the absolute size of the skewness, the more asymmetric is distribution. A large positive value indicates a long right tail, and a large negative value indicates a long left tail. A zero value indicates symmetry around the mean. ( μ ) 4 E X Kurtosis: 4 σ Measures thickness of tails of a distribution A kurtosis above three indicates fat tails or leptokurtosis, relative to the normal, i.e. extreme events are more likely to occur. K =

Central Limit Theorem: Basic Idea As sample size gets large (n 30)... sample mean will have a normal distribution. X

Important Continuous Distributions All derived from normal distribution 2 χ distribution: arises from squared normal random variables, t distribution: arises from ratios of normal 2 and χ variables F distribution: arises 2 from ratios of χ variables. 2 χ t distribution (red), normal distribution (blue) distribution F distribution

Fundamentals of Hypothesis Testing

Identifying Hypotheses 1. Question, e.g. test that the population mean is equal to 3 2. State the question statistically (H 0 : μ = 3) 3. State its opposite statistically (H 1 : μ 3) Hypotheses are mutually exclusive & exhaustive Sometimes it is easier to form the alternative hypothesis first. 4. Choose level of significance α Typical values are 0.01, 0.05, 0.10 Rejection region of sampling distribution: the unlikely values of sample statistic if null hypothesis is true

Identifying Hypotheses: Examples 1. Is the population average amount of TV viewing 12 hours? μ = 12 μ 12 H 0 : μ = 12 H 1 : μ 12 2. Is the population average amount of TV viewing different from 12 hours? μ 12 μ = 12 H 0 : μ = 12 H 1 : μ 12

Hypothesis Testing: Basic Idea Sampling Distribution It is unlikely that we would get a sample mean of this value...... if in fact this were the population mean.... Therefore, we reject the null hypothesis that μ = 50. 20 μ = 50 H 0 sample mean

Example: Z-test statistic (σ known) 1. Convert Sample Statistic (e.g., ) to Standardized Z Variable X μx X μ Z = = σ σ x n 2. Compare to Critical Z Values If Z-test statistic falls in critical region, reject H 0 ; Otherwise do not reject H 0 X

p-value Probability of obtaining a test statistic more extreme ( or ) than actual sample value given H 0 is true Smallest value of α for which H 0 can be rejected Used to make rejection decision If p value α, do not reject H 0 If p value < α, reject H 0

One-Tailed Test: Rejection Region H 0 : μ 0 H 1 : μ < 0 H 0 : μ 0 H 1 : μ > 0 Reject H 0 α Reject H 0 α 0 Must be significantly below μ. Z 0 Z Here: Small values don t contradict H 0.

One-Tailed Z Test: Finding Critical Z Values What Is Z Given α = 0.025?.500 -.025.475 0 σ Z = 1 1.96 Z α /2 =.025 Standardized Normal Probability Table (Portion).06 Z.05.07 1.6.4505.4515.4525 1.7.4599.4608.4616 1.8.4678.4686.4693 1.9.4744.4750.4756

Two-Tailed Test: Rejection Regions Sampling Distribution Rejection Region 1/2 α 1 - α Nonrejection Region H 0 : μ = 0 H 1 : μ 0 Level of Confidence Rejection Region 1/2 α Critical Value H 0 Value Critical Value Sample Statistic

t-test, F-test Test statistic may not be normally distributed => z-test not applicable Examples: Variance unknown, but estimated. Hypothesis that the slope of a regression line differs significantly from zero. => t-test Hypothesis that the standard deviations of two normally distributed populations are equal. => F-test

Jarque-Bera test Assesses whether a given sample of data is normally distributed. Aggregates information in the data about both skewness and kurtosis. Test of the hypothesis that S = 0 and K = 3, based on Ŝ and ˆK. Test statistic: T ˆ 2 1 ( ˆ ) 2 JB = S + K 3 6 4 (here T is the number of observations) Under the null hypothesis of independent normallydistributed observations, the Jarque-Bera statistic is 2 distributed in large samples as a χ random variable with 2 degrees of freedom.

Simple Linear Regression

Simple Linear Regression Model y-intercept slope random iid 2 error ε (0, σ ) t y = + x + t 0 1 t t β β ε dependent (response) variable independent (explanatory) variable

Linear Regression Assumptions 1. x is exogenously determined 2. ε t are iid(0,σ 2 ) (iid = independently and identically distributed ) Zero mean Independence of errors (no autocorrelation) Constant variance (homoscedasticity) More things to think about: Normality of ε t (if not satisfied, inference procedures only asymptotically valid) Model specification (e.g. linearity, β 1 constant over time?)

Simple Linear Regression Model y y = β + β x + ε t 0 1 t t ε t = disturbance observed value observed value ( ) E yx = β + β x * * 0 1 x

-- Sample Linear Regression Model y y = b + b x + e i 0 1 i i e i = Random Error y = b + b x i 0 1 i Unsampled Observation Observed Value x

Ordinary Least Squares OLS minimizes sum of squared residuals min ˆ β, ˆ β 0 1 y T t= 1 ( y ˆ ˆ ) t β0 β1xt 2 y = β + β x + ε t 0 1 t t T ( ) 2 2 y yˆ = e t t t t= 1 t= 1 T predicted value e 2 e 4 e 1 e 3 fitted value (in-sample forecast) y = ˆ β + ˆ β x ˆt 0 1 x t

On Thursday: Evaluating the Model 1. Examine variation measures coefficient of determination ( goodness of fit ) standard error of the estimate 2. Analyze residuals e serial correlation 3. Test coefficients for significance β y ˆt = ˆ β + ˆ β x 0 1 t

Random Error Variation 1.Variation of Actual Y from Predicted Y 2. Measured by Standard Error of Estimate Sample Standard Deviation of e Denoted S YX 3. Affects Several Factors Parameter Significance Prediction Accuracy

Measures of Variation in Regression 1.Total Sum of Squares (SST) Measures variation of observed Y i around the mean, Y 2.Explained Variation (SSR) Variation due to relationship between X & Y 3.Unexplained Variation (SSE) Variation due to other factors

Variation Measures Y Y i Unexplained Sum of Squares (Y i - Y ^ i ) 2 Total Sum of Squares (Y i -Y) 2 X i Explained Sum of Squares (Y ^ i -Y) 2 Y = b + b X i Y 0 1 X i

Coefficient of Determination Proportion of Variation Explained by Relationship Between X & Y r 2 Explained Variation SSR = = Total Variation SST = n n b Y + b XY n(y) 0 i 1 i i i = 1 i = 1 n 2 Yi i = 1 n(y) 2 ˆ 2 0 r 2 1

Coefficients of Determination (r2) and Correlation (r) Yr 2 = 1, r = +1 Y r 2 = 1, r = -1 ^ Y^ i = b 0 + b 1 X i X Y i = b 0 + b 1 X i Yr 2 =.8, r = +0.9 Y r 2 = 0, r = 0 X Y^ i = b 0 + b 1 X i X Y^ i = b 0 + b 1 X i X

Standard Error of Estimate 2 2 ) ˆ ( 1 1 1 1 0 2 1 2 = = = = = = n Y X b Y b Y S n Y Y S n i n i n i i i i i YX n i i i YX

Residual Analysis 1.Graphical Analysis of Residuals Plot residuals vs. X i values Residuals mean errors Difference between actual Y i & predicted Y i 2.Purposes Examine functional form (linear vs. non-linear Model) Evaluate violations of assumptions

Test of Slope Coefficient for Significance 1.Tests If There Is a Linear Relationship Between X & Y 2.Hypotheses H 0 : β 1 = 0 (No Linear Relationship) H 1 : β 1 0 (Linear Relationship) 3.Test Statistic b = 1 t β 1 n 2 S b 1 where S = b 1 S YX n X 2 n( X ) 2 i i = 1