Introduction to Statistical Inference

Similar documents
Introduction to probability theory

6.4 Type I and Type II Errors

Chapter 7: Hypothesis Testing

z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests

Statistical Inference

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

Dover- Sherborn High School Mathematics Curriculum Probability and Statistics

Chapter 5: HYPOTHESIS TESTING

Null Hypothesis Significance Testing p-values, significance level, power, t-tests Spring 2017

Math Review Sheet, Fall 2008

Introductory Econometrics. Review of statistics (Part II: Inference)

Background to Statistics

Review. December 4 th, Review

PHP2510: Principles of Biostatistics & Data Analysis. Lecture X: Hypothesis testing. PHP 2510 Lec 10: Hypothesis testing 1

18.05 Practice Final Exam

Ling 289 Contingency Table Statistics

Introductory Econometrics

Summary: the confidence interval for the mean (σ 2 known) with gaussian assumption

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary

The Chi-Square Distributions

Non-parametric Inference and Resampling

Lecture 1: Probability Fundamentals

Quantitative Methods for Economics, Finance and Management (A86050 F86050)

Preliminary Statistics Lecture 5: Hypothesis Testing (Outline)

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

Frequentist Statistics and Hypothesis Testing Spring

Central Limit Theorem ( 5.3)

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages

This does not cover everything on the final. Look at the posted practice problems for other topics.

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

Statistical Inference

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

Space Telescope Science Institute statistics mini-course. October Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses

Statistical Preliminaries. Stony Brook University CSE545, Fall 2016

F79SM STATISTICAL METHODS

An introduction to biostatistics: part 1

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Design of Engineering Experiments

Math 494: Mathematical Statistics

EXAMINERS REPORT & SOLUTIONS STATISTICS 1 (MATH 11400) May-June 2009

Probability & Statistics - FALL 2008 FINAL EXAM

Probability theory basics

VTU Edusat Programme 16

Outline. Unit 3: Inferential Statistics for Continuous Data. Outline. Inferential statistics for continuous data. Inferential statistics Preliminaries

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

Lecture 12: Small Sample Intervals Based on a Normal Population Distribution

CENTRAL LIMIT THEOREM (CLT)

Practice Problems Section Problems

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur

Statistical Data Analysis Stat 3: p-values, parameter estimation

Business Statistics: Lecture 8: Introduction to Estimation & Hypothesis Testing

Probability Theory and Statistics. Peter Jochumzen

INTERVAL ESTIMATION AND HYPOTHESES TESTING

Chapter 8 of Devore , H 1 :

14.30 Introduction to Statistical Methods in Economics Spring 2009

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

The Central Limit Theorem

Problem 1 (20) Log-normal. f(x) Cauchy

MTMS Mathematical Statistics

simple if it completely specifies the density of x

If we want to analyze experimental or simulated data we might encounter the following tasks:

Class 26: review for final exam 18.05, Spring 2014

Introduction to statistics

MAT 271E Probability and Statistics

Null Hypothesis Significance Testing p-values, significance level, power, t-tests

Section 9.4. Notation. Requirements. Definition. Inferences About Two Means (Matched Pairs) Examples

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

Sociology 6Z03 Review II

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics

Review for Final. Chapter 1 Type of studies: anecdotal, observational, experimental Random sampling

Linear Models: Comparing Variables. Stony Brook University CSE545, Fall 2017

Probability and Statistics

CONTINUOUS RANDOM VARIABLES

Lecture 8 Sampling Theory

7.1 Basic Properties of Confidence Intervals

Sampling Distributions of Statistics Corresponds to Chapter 5 of Tamhane and Dunlop

Normal (Gaussian) distribution The normal distribution is often relevant because of the Central Limit Theorem (CLT):

Hypothesis tests

Introduction and Descriptive Statistics p. 1 Introduction to Statistics p. 3 Statistics, Science, and Observations p. 5 Populations and Samples p.

10/4/2013. Hypothesis Testing & z-test. Hypothesis Testing. Hypothesis Testing

The Chi-Square Distributions

Review of Statistics

Exam 2 Practice Questions, 18.05, Spring 2014

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

Evaluating Hypotheses

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Institute of Actuaries of India

Answer keys for Assignment 10: Measurement of study variables (The correct answer is underlined in bold text)

Review of Basic Probability Theory

What is a random variable

Week 1 Quantitative Analysis of Financial Markets Distributions A

Mathematical Statistics

Hypothesis testing: theory and methods

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Chapters 9. Properties of Point Estimators

Statistical Intervals (One sample) (Chs )

0 0'0 2S ~~ Employment category

LAB 2. HYPOTHESIS TESTING IN THE BIOLOGICAL SCIENCES- Part 2

Chapter 24. Comparing Means

Transcription:

Introduction to Statistical Inference Dr. Fatima Sanchez-Cabo f.sanchezcabo@tugraz.at http://www.genome.tugraz.at Institute for Genomics and Bioinformatics, Graz University of Technology, Austria Introduction to Probability Theory. March 9, 2006 p. 1/29

Summary last session Motivation Algebra of sets Definition of probability space: Sample space Sigma algebra Probability axioms Conditional Probability and independence of events Random variables Continuos vs discrete pdf and mass function Distribution function Introduction to Probability Theory. March 9, 2006 p. 2/29

Part IV: From Probability to Statistics Introduction to Probability Theory. March 9, 2006 p. 3/29

Statistical Inference The target of statistical inference is to provide some information about the probability distribution P defined over the probability space (Ω, F). Differently from the previous examples where an exhaustive observation was possible, this is often difficult. Hence, statistical inference focusses in the analysis and interpretation of the realizations of the random variable in order to draw conclusions about the probability law under study. The conclusions can be relative to: Estimation of a unique value for a parameter or parameters essential in the probability distribution (i.e., p for a Binomial r.v) Estimation of a confidence interval for this parameter/s. Accept or reject a certain hypothesis about the probability distribution of interest. Introduction to Probability Theory. March 9, 2006 p. 4/29

Statistical Inference In general, there is some knowledge about the probability distribution explaining a certain random process The inference process is often involved with deciding (looking at the available data) which is the distribution that best explains the available data among a set of them: {F θ : θǫθ}. That s known as parametric statistics. Example 4.1: We are betting with a friend to "head" or "tail" when tossing a coin. We want to know if the coin is unbiased. The most logic approach will be to toss the coin "enough" times to get an approximate value for the probability of head. Introduction to Probability Theory. March 9, 2006 p. 5/29

Choosing a random sample The first step in this process is to choose a sample representative from the whole population (what is "enough" in the previous example?) We should also take care of the cost/time. Random sampling is one way to obtain samples from a population: all individuals have the same probability to be chosen. This type of sampling is particularly useful because the observations are then independent from each other; hence F(x 1,...,x n ) = i F(x i) Sampling with replacement (some individuals might be repeated in the sample) Sampling without replacement (each individual can be only once in the sample) Introduction to Probability Theory. March 9, 2006 p. 6/29

Combinatorics We can define the probability of an event A as: P(A) = number favorables cases number of possible cases The number of possible samples of r elements chosen without replacement among n possible elements is C(n, r) = n! (n r)!r! The number of possible ways to create a sample of size n using r < n elements, if repetition is allowed, is r n Examples: 1. An urn contains 5 red, 3 green, 2 blue and 4 white balls. A sample of size 8 is selected at random without replacement. Calculate the probability that the sample contains 2 red, 2 green, 1 blue, and 3 white balls. 2. Consider a class with 30 students. Calculate the probability that all 30 birthdays are different. Introduction to Probability Theory. March 9, 2006 p. 7/29

Example 4.2 We are betting with a friend to "head" or "tail" when tossing a coin. We want to know what is the probability of getting "head" so that we decide what to bet on. Hence, we have the random variable X = { 0 Head 1 Tail with P(X = 0) = θ, P(X = 1) = 1 θ. Given a random sample (X 1, X 2, X 3 ), determine the mass function (probability for each possible triplet). Does this sample has a known mass distribution? Introduction to Probability Theory. March 9, 2006 p. 8/29

Descriptive statistics Given a very large or very complex sample it might not be possible to determine the distribution easily. It is often useful to make use of some tools that help to understand the target distribution and the main characteristics of the sample data: Probability Statistic Base Population Sample Central tendency Dispersion Expectation Variance Mean Median Sample variance IQR Mode Introduction to Probability Theory. March 9, 2006 p. 9/29

Descriptive statistics Given a sample (x 1,...,x n ) we can calculate the following expressions to get a rough idea of the properties of the sample. 1. Parametric Mean: x = 1 n i x i Sample variance: s 2 = 1 n n i=1 (x i x) 2 Sample Quasi-variance S 2 = 1 n 1 n i=1 (x i x) 2 2. Non-parametric (order statistics) Median: x [ n+1 2 ] if n is odd Interquantiles range: IQR=q 3 -q 1 Mode: most repeated value x [ n 2 ]+x [ n 2 ]+1 2 if n is even Introduction to Probability Theory. March 9, 2006 p. 10/29

Descriptive statistics Histogram of x Frequency 0 5 10 15 20 25 30 2 4 6 8 2 4 6 8 x (a) Histogram (b) Boxplot Introduction to Probability Theory. March 9, 2006 p. 11/29

Descriptive statistics Given x 1,...,x n a sample of size n, calculate the p th percentile: 1. R = int( p 100 FR = p 100 2. Find x [R], x [R]+1 (N + 1)); p (N + 1) int( 100 3. p = x [R] + (x [R]+1 x [R] ) FR (N + 1)) Introduction to Probability Theory. March 9, 2006 p. 12/29

Estimators and statistics Definition: A statistic is a function T : (Ω, F) (R k, B k ), i.e. T(x 1,...,x n ), over which a probability function can be defined. k is the dimension of the statistic. An estimator is a particular case of a statistic that approximates one of the unknown parameters of the distribution. Examples: ˆµ = x = 1 n n i=1 x i ˆθ in example 4.2. An estimator is a random variable itself: Its moments can be calculated An estimator is said to be unbiased if E[ˆθ] = θ Introduction to Probability Theory. March 9, 2006 p. 13/29

Example 4.3 Given (X 1,...,X n ) a random sample such as E[X i ] = µ and V [X i ] = σ 2, show that for X = 1 n i X i then E[ X] = µ and V [ X] = σ 2 /n. Find an unbiased estimator for σ 2? Note: The variance of an estimator is the sample error (SE). Introduction to Probability Theory. March 9, 2006 p. 14/29

Distribution of the statistics Since the statistics are function of random variables they are theirselves random variables for which a probability distribution can be defined. Particularly important are the distributions of statistics based on normally distributed random variables. We have proved that X N(µ, σ 2 /n). For the Central Limit Theorem X µ σ/ N(0, 1). n If σ is unknown and we substitute it by its natural estimator S 2, then t = X µ S/ n That is the Student-t distribution. Introduction to Probability Theory. March 9, 2006 p. 15/29

Student s t Introduction to Probability Theory. March 9, 2006 p. 16/29

χ 2 and F-Snedecor Other important distributions of statistics from normally distributed random variables are: χ 2 n = i X2 i when X i N(0, 1) F n,m = χ2 n /n χ 2 m /m Chi square (d.f.=10) F(5,10) Density 0.00 0.02 0.04 0.06 0.08 0.10 Density 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0 10 20 30 40 x 0 1 2 3 4 5 x (c) χ 2 density (d) F-Snedecor Introduction to Probability Theory. March 9, 2006 p. 17/29

Part V: Hypothesis testing Introduction to Probability Theory. March 9, 2006 p. 18/29

Motivation There are some situations when we want to test if a certain hypothesis about the parameter of interest is true Is the probability of "head" the same than the probability of "tail" when flipping the coin of the previous experiment? QUESTION???? Formulate an hypothesis EXPERIMENT Results support the hypothesis Results DON T support the hypothesis Figure 1: The Scientific Method Introduction to Probability Theory. March 9, 2006 p. 19/29

Main concepts A statistical null hypothesis H 0 is the one that want to be tested. Statistical tools provide the appropriated framework to guarantee that its rejection is due to real mechanisms and not to chance. Its rejection will provide the called "statistical significance". We never talk about "accepting" H 0 : with the available data we can say if we reject or not the H 0. Introduction to Probability Theory. March 9, 2006 p. 20/29

Error types H 0 is really... Decision True False Error! A false hypothesis has OK!A true hypothesis has Accept H 0 been accepted. This is a Type II error. been accepted The probability of this is β Error! a true hypothesis has been rejected. OK! Reject H 0 This is a Type I error. A false hypothesis has been rejected The probability of this occurring is α Introduction to Probability Theory. March 9, 2006 p. 21/29

Error types The ideal situation would be to minimize the probability of both errors (α,β), where: α = P(Reject H 0 /H 0 is true) = P(ETI) β = P(Accept H 0 /H 0 is false) = P(ETII) However, they are related and cannot be minimized at the same time. Definition: The p-value corresponds to the probability of rejecting the null hypothesis if it is actually true. If the p-value is smaller than the allowed α-level H 0 is rejected (always with the available data!) Introduction to Probability Theory. March 9, 2006 p. 22/29

Protocol for statistically testing a null hypothesis 1. Question that we want to address (aufpassen: Statistical tests can only reject the null hypothesis, H0) 2. Experimental design: How many repetitions we need How are we going to select a sample that represents the population? Which are the sources of variation? How do we remove the systematic errors? 3. Data collection 4. Removal systematic errors 5. Which statistic summarizes best my sample. Does it have a known distribution? (statistic: numerical variable that summarizes my sample data in a meaningful way) Introduction to Probability Theory. March 9, 2006 p. 23/29

Example (5) Introduction to Probability Theory. March 9, 2006 p. 24/29

Simple hypothesis testing 1. We want to test if a particular sample comes from a distribution with a particular mean value, let s say 3. It is quite logical to try to estimate how close the estimator of the population mean is from the proposed value. So we will calculate: ( x 3) 2. To minimize the intrinsic error of the sample mean, we will divide this difference by the standard error (variance of the sample mean). We have then the statistic: t = x 3 s/ n 3. Regarding to the definition of t-student, if H 0 is true (H 0 : µ = 3) then, t t n 1 where n is the number of samples. This makes sense, because in the t student distribution, the probability of getting t = 0 is very high and this is exactly what would happen under H 0, because: x 3 x 3 0 t = 0 Introduction to Probability Theory. March 9, 2006 p. 25/29

Simple hypothesis testing (cont.) α was defined as: 1 α = P( t t ) α = P( t > t ) Where t = t n 1; α 2 and α is as well defined as: α = P(Reject H 0 /H 0 true) = P(ETI) In this particular example, if we are under the H0 the value calculated for the statistic t should be close to 0. Otherwise, we reject H 0 (this would mean that x 3). But, how big must be the difference ( x-3)? For that, we fixed the value α such as we just reject H 0 being true for 5 out of 100 samples that we take. And for this particular, this α determines a unique value, called critical value: t = t n 1; α 2 Introduction to Probability Theory. March 9, 2006 p. 26/29

If t > t, we reject the H 0, because the probability of t = 0 (µ =3) is very small (<0.05). Just in 5 out of 100 samples (what we can consider by chance) the value of the statistic will be very far away from 0 although the H 0 is true. This condition is equivalent to: p = P(t n 1 > t ) < α = 0.05 If t < t the result obtained is the one we would expect for a distribution following a t-distribution with n 1 degrees of freedom, as expected if the H 0 is true. We don t have then enough evidence to reject the H 0. This is equivalent to say that: p = P(t n 1 > t ) > α = 0.05 Introduction to Probability Theory. March 9, 2006 p. 27/29

0.4 t distribution 0.35 0.3 0.25 Density 0.2 0.15 α=0.05 0.1 α/2=0.025 0.05 0 6 4 t* 0 t*=2.22 4 6 Introduction to Probability Theory. March 9, 2006 p. 28/29

References [1] Durbin, R., Eddy, S., Krogh, A. and Mitchison, G. (1996) Biological sequence analysis, Cambridge University Press. [2] Durret, R. (1996) Probability: Theory and examples, Duxbury Press, Second edition. [3] Rohatgi, V.K. and Ehsanes Saleh, A.K.Md. (1988) An introduction to probability and statistics, Wiley, Second Edition. [4] Tuckwell, H.C. (1988) Elementary applications of probability theory, [5] Gonick, L. and Smith, W. (2000) The cartoon guide to statistics. Chapman and Hal [Engineering statistics] http://www.itl.nist.gov/div898/handbook/ Introduction to Probability Theory. March 9, 2006 p. 29/29