Math 494: Mathematical Statistics

Size: px
Start display at page:

Download "Math 494: Mathematical Statistics"

Transcription

1 Math 494: Mathematical Statistics Instructor: Jimin Ding Department of Mathematics Washington University in St. Louis Class materials are available on course website ( jmding/math494/ ) Spring 2018 Jimin Ding, Math WUSTL Math 494 Spring / 44

2 Introduction to Statistical Inferences Jimin Ding, Math WUSTL Math 494 Spring / 44

3 Statistical Problem A typical statistical problem can be summarized in the following flow: random experiments data analysis inferences The random experiments may come from natural sciences, social sciences, or engineered system. Sometimes they are well designed to control the factors of interest (experimental design), and they are from real-world processes (observational studies). Jimin Ding, Math WUSTL Math 494 Spring / 44

4 Statistical Problem A typical statistical problem can be summarized in the following flow: random experiments data analysis inferences The random experiments may come from natural sciences, social sciences, or engineered system. Sometimes they are well designed to control the factors of interest (experimental design), and they are from real-world processes (observational studies). The collected data could be scalars (exam responses), vectors (stock price), matrices (digital image), arrays (contingency table), characters (text mining), or functions (time series, fmri). Jimin Ding, Math WUSTL Math 494 Spring / 44

5 Statistical Analysis There are two ways of analyzing data: Descriptive data analysis (explanatory data analysis) summarizes data into some statistics, such as the mean, median, range, standard deviation... visualize features of data by histogram, pie chart, boxplot... Jimin Ding, Math WUSTL Math 494 Spring / 44

6 Statistical Analysis There are two ways of analyzing data: Descriptive data analysis (explanatory data analysis) summarizes data into some statistics, such as the mean, median, range, standard deviation... visualize features of data by histogram, pie chart, boxplot... Inferential Statistical Analysis assumes a probability model on the collected data (sample) investigates features of the distribution or a family of distributions to infer the features of the entire group of individuals (population) that may be impossible or too expansive to exam. Jimin Ding, Math WUSTL Math 494 Spring / 44

7 Statistical Analysis There are two ways of analyzing data: Descriptive data analysis (explanatory data analysis) summarizes data into some statistics, such as the mean, median, range, standard deviation... visualize features of data by histogram, pie chart, boxplot... Inferential Statistical Analysis assumes a probability model on the collected data (sample) investigates features of the distribution or a family of distributions to infer the features of the entire group of individuals (population) that may be impossible or too expansive to exam. We will focus on the later one in this course. Jimin Ding, Math WUSTL Math 494 Spring / 44

8 Example 1: Quality Control Consider a population of N elements, for instance, a shipment of manufactured items. An unknown number N θ of these elements are defective. It will be expensive to exam all items for large N (or impossible if the inspection is destructive). To learn θ, one may randomly draw a sample of n without replacement and inspect. (Assume all items have the same probability to be defective.) Population: Sample (data): Probability model: Jimin Ding, Math WUSTL Math 494 Spring / 44

9 Example 1: Quality Control Consider a population of N elements, for instance, a shipment of manufactured items. An unknown number N θ of these elements are defective. It will be expensive to exam all items for large N (or impossible if the inspection is destructive). To learn θ, one may randomly draw a sample of n without replacement and inspect. (Assume all items have the same probability to be defective.) Population: N items, N θ defective items Sample (data): Probability model: Jimin Ding, Math WUSTL Math 494 Spring / 44

10 Example 1: Quality Control Consider a population of N elements, for instance, a shipment of manufactured items. An unknown number N θ of these elements are defective. It will be expensive to exam all items for large N (or impossible if the inspection is destructive). To learn θ, one may randomly draw a sample of n without replacement and inspect. (Assume all items have the same probability to be defective.) Population: N items, N θ defective items Sample (data): n sampled items, X defective items in sample Probability model: Jimin Ding, Math WUSTL Math 494 Spring / 44

11 Example 1: Quality Control Consider a population of N elements, for instance, a shipment of manufactured items. An unknown number N θ of these elements are defective. It will be expensive to exam all items for large N (or impossible if the inspection is destructive). To learn θ, one may randomly draw a sample of n without replacement and inspect. (Assume all items have the same probability to be defective.) Population: N items, N θ defective items Sample (data): n sampled items, X defective items in sample Probability model: P r(x = k) = ( Nθ k )( N Nθ n k ) ( N n), k = 0, 1,, min(nθ, n). This is the hypergeometric distribution, H(N θ, N, n), which is a family of distributions indexed by θ. Here θ is the unknown parameter that we want to estimate. Jimin Ding, Math WUSTL Math 494 Spring / 44

12 Example 2: Measurement Problem An experimenter makes n independent determinations of a physical constant µ. Population: Sample (data): Probability model: Jimin Ding, Math WUSTL Math 494 Spring / 44

13 Example 2: Measurement Problem An experimenter makes n independent determinations of a physical constant µ. Population: all measurements for physical experiments. Sample (data): Probability model: Jimin Ding, Math WUSTL Math 494 Spring / 44

14 Example 2: Measurement Problem An experimenter makes n independent determinations of a physical constant µ. Population: all measurements for physical experiments. Sample (data): Observed measurements X 1,, X n from recorded experiments, subject to measurement errors. Probability model: Jimin Ding, Math WUSTL Math 494 Spring / 44

15 Example 2: Measurement Problem An experimenter makes n independent determinations of a physical constant µ. Population: all measurements for physical experiments. Sample (data): Observed measurements X 1,, X n from recorded experiments, subject to measurement errors. Probability model: X i = µ + ɛ i, 1 i n. Jimin Ding, Math WUSTL Math 494 Spring / 44

16 Example 2: Measurement Problem An experimenter makes n independent determinations of a physical constant µ. Population: all measurements for physical experiments. Sample (data): Observed measurements X 1,, X n from recorded experiments, subject to measurement errors. Probability model: X i = µ + ɛ i, 1 i n. Different distribution assumptions can be considered for ɛ. 1. ɛ i are independent. 2. ɛ i are identically distributed 3. the distribution of ɛ i does not depend on µ. 4. (1) ɛ i iid N(0, σ 2 ). Here θ = (µ, σ 2 ) Θ, Θ = R R + R 2. (2) ɛ i iid F, where F is a cdf with expectation 0 and finite variance σ 2. Jimin Ding, Math WUSTL Math 494 Spring / 44

17 Example 3: Treatment Effect To compare the treatment effect of drug A and B for treating a given disease, m subjects were assign to drug A and n were assign to drug B. Let X 1,..., X m be responses of m subjects receiving drug A, and Y 1,..., Y n be responses of n subjects receiving drug B. If drug A is placebo, then X 1,..., X n are referred to as control observations, Y 1,..., Y n are treatment observations. Population: Sample (data): Probability model: Jimin Ding, Math WUSTL Math 494 Spring / 44

18 Example 3: Treatment Effect To compare the treatment effect of drug A and B for treating a given disease, m subjects were assign to drug A and n were assign to drug B. Let X 1,..., X m be responses of m subjects receiving drug A, and Y 1,..., Y n be responses of n subjects receiving drug B. If drug A is placebo, then X 1,..., X n are referred to as control observations, Y 1,..., Y n are treatment observations. Population: Responses of all subjects receiving drug A and B. Sample (data): Probability model: Jimin Ding, Math WUSTL Math 494 Spring / 44

19 Example 3: Treatment Effect To compare the treatment effect of drug A and B for treating a given disease, m subjects were assign to drug A and n were assign to drug B. Let X 1,..., X m be responses of m subjects receiving drug A, and Y 1,..., Y n be responses of n subjects receiving drug B. If drug A is placebo, then X 1,..., X n are referred to as control observations, Y 1,..., Y n are treatment observations. Population: Responses of all subjects receiving drug A and B. Sample (data): X 1,..., X m for A, Y 1,..., Y n for B. Probability model: Jimin Ding, Math WUSTL Math 494 Spring / 44

20 Example 3: Treatment Effect To compare the treatment effect of drug A and B for treating a given disease, m subjects were assign to drug A and n were assign to drug B. Let X 1,..., X m be responses of m subjects receiving drug A, and Y 1,..., Y n be responses of n subjects receiving drug B. If drug A is placebo, then X 1,..., X n are referred to as control observations, Y 1,..., Y n are treatment observations. Population: Responses of all subjects receiving drug A and B. Sample (data): X 1,..., X m for A, Y 1,..., Y n for B. Probability model: Different models can be considered: 1. X i iid F, Yj iid G, then the model is (F, G) 2. F N(µ, σ 2 ) and G N(µ +, σ 2 ), then the parameter is (, µ, σ 2 ). Jimin Ding, Math WUSTL Math 494 Spring / 44

21 Two Types of Questions iid Given X 1,, X n F or f, (referred to as a random sample of size n from the distribution F (or f) ), how do we infer F or f? F = {f(x) :f > 0, f(x)dx = 1} or F = {F (x) :F ( ) = 0, F (+ ) = 1, R F is monotone increasing and right continuous.} Jimin Ding, Math WUSTL Math 494 Spring / 44

22 Two Types of Questions iid Given X 1,, X n F or f, (referred to as a random sample of size n from the distribution F (or f) ), how do we infer F or f? F = {f(x) :f > 0, f(x)dx = 1} or F = {F (x) :F ( ) = 0, F (+ ) = 1, R F is monotone increasing and right continuous.} Given X 1,, X n iid F (x; θ) or f(x; θ), how do we infer θ? F = {f(x; θ) : θ Θ} (where θ is an unknown parameter take can take values in the parameter space Θ.) Jimin Ding, Math WUSTL Math 494 Spring / 44

23 Two Types of Questions iid Given X 1,, X n F or f, (referred to as a random sample of size n from the distribution F (or f) ), how do we infer F or f? F = {f(x) :f > 0, f(x)dx = 1} or F = {F (x) :F ( ) = 0, F (+ ) = 1, Nonparametric model R F is monotone increasing and right continuous.} Given X 1,, X n iid F (x; θ) or f(x; θ), how do we infer θ? F = {f(x; θ) : θ Θ} (where θ is an unknown parameter take can take values in the parameter space Θ.) Parametric model Jimin Ding, Math WUSTL Math 494 Spring / 44

24 Type of Statistical Models Nonparametric model: goal is F or f Parametric model: goal is θ Jimin Ding, Math WUSTL Math 494 Spring / 44

25 Type of Statistical Models Nonparametric model: goal is F or f Parametric model: goal is θ Since the second question has a smaller space of candidates, we first start with question 2 (simpler). Jimin Ding, Math WUSTL Math 494 Spring / 44

26 Type of Statistical Models Nonparametric model: goal is F or f Parametric model: goal is θ Since the second question has a smaller space of candidates, we first start with question 2 (simpler). Sometimes, θ is a vector that we are only interested in one (or some) component of θ. In this case, the remaining parameters are referred to as nuisance parameter. Jimin Ding, Math WUSTL Math 494 Spring / 44

27 Type of Statistical Models Nonparametric model: goal is F or f Parametric model: goal is θ Since the second question has a smaller space of candidates, we first start with question 2 (simpler). Sometimes, θ is a vector that we are only interested in one (or some) component of θ. In this case, the remaining parameters are referred to as nuisance parameter. Semiparametric model is a combination of parametric and nonparametric model and the goal is f(x; θ, g) which is only partially specified by the parameter θ, and g is a unspecified function (infinite-dimension). Jimin Ding, Math WUSTL Math 494 Spring / 44

28 What is Statistical Inference? Statistical inference is the process of using data to infer the distribution that generated data. Jimin Ding, Math WUSTL Math 494 Spring / 44

29 What is Statistical Inference? Statistical inference is the process of using data to infer the distribution that generated data. Three components of statistical inference: Point Estimation Confidence Interval Hypothesis Testing Jimin Ding, Math WUSTL Math 494 Spring / 44

30 Point Estimation Jimin Ding, Math WUSTL Math 494 Spring / 44

31 Point Estimation Parameter of Interest: A fixed and unknown population parameter, θ, or a function of model parameter, g(θ). Eg: population mean µ, population standard deviation σ,... Point Estimator: A statistic from a sample that is used to estimate the parameter of interest, ˆθ (which is a r.v.). Eg: sample mean X = i X i/n for µ, (X sample standard deviation S = i X) 2 i (n 1) for σ. Estimate: The numerical value of an estimator in an observed sample. Eg: observed sample mean x = i x i/n for µ, (x observed sample standard deviation s = i x) 2 i (n 1) for σ. Point estimation: the process of providing a point estimator. Jimin Ding, Math WUSTL Math 494 Spring / 44

32 Example 1: Poisson Model (Ex 4.1.3) Suppose the number of customers X that enter a store during 9-10am. follows a Poisson distribution with parameter θ. Suppose a random sample of the number of customers that enter the store during 9-10am. for 10 days results in the values: What is a good guess of θ? 9, 7, 9, 15, 10, 13, 11, 7, 2, 12. go back to CI Jimin Ding, Math WUSTL Math 494 Spring / 44

33 Maximum Likelihood Estimation (MLE) The observed values in a random sample, x 1, x 2,, x n, can be called as realizations of the sample. Likelihood function: joint pdf of the realizations (observed data) L(θ) =f 1 (x 1 ; θ)f 2 (x 2 ; θ) f n (x n ; θ) =Π n i=1f(x i ; θ) (if iid) The larger L(θ) is, the more possible that we observe these realizations x 1, x 2,, x n. The θ that maximizes the likelihood function, L(θ), is referred to as maximum likelihood estimator (MLE). ˆθ = arg max θ Θ L(θ) = arg max log L(θ) θ Θ Jimin Ding, Math WUSTL Math 494 Spring / 44

34 Example 2: Normal Model (Example 4.1.3) Jimin Ding, Math WUSTL Math 494 Spring / 44

35 Example 3: Uniform Distribution (Example 4.1.4) Jimin Ding, Math WUSTL Math 494 Spring / 44

36 Properties of MLE Plug-In Theorem: If ˆθ is the MLE of θ, then g(ˆθ) is the MLE of g(θ). The MLE is consistent. The MLE is asymptotically unbiased. The MLE is efficient. The MLE is often asymptotically normal. The MLE is equivalent to LSE in regression under normality assumption. Jimin Ding, Math WUSTL Math 494 Spring / 44

37 Nonparametric MLE: for CDF Empirical CDF for F ˆF X (x) = 1 n n 1(X i x). i=1 This stepwise function is called the empirical CDF. It is an unbiased estimator of CDF. Jimin Ding, Math WUSTL Math 494 Spring / 44

38 Nonparametric MLE: for CDF Empirical CDF for F ˆF X (x) = 1 n n 1(X i x). i=1 This stepwise function is called the empirical CDF. It is an unbiased estimator of CDF. Because it can be viewed as an average of Bernoulli random variables, it follows LLN as n. It can be proved to be a n-consistent estimator of CDF, and it is asymptotically normal. It can be further proved it is a uniformly consistent estimator of CDF. Jimin Ding, Math WUSTL Math 494 Spring / 44

39 Nonparametric MLE: for PDF or PMF Discrete case: If the sample space (possible outcome of X) is finite, one may list all possible values {a 1, a 2,, a m }, then define ˆf(a j ) = 1 n n 1(X i = a j ), for j = 1,, m i=1 If the sample space is infinite, one may select some higher possible values and group the others {a 1, a 2,, a m, ã m+1 }, where ã m+1 = {a m+1, a m+2, } Rule of thumb: select m s.t. ˆf(am ) > ˆf(ã m+1 ) Continuous case: For a given x, consider (x h, x + h) for some h > 0, Jimin Ding, Math WUSTL Math 494 Spring / 44

40 Nonparametric MLE: for PDF or PMF Discrete case: If the sample space (possible outcome of X) is finite, one may list all possible values {a 1, a 2,, a m }, then define ˆf(a j ) = 1 n n 1(X i = a j ), for j = 1,, m i=1 If the sample space is infinite, one may select some higher possible values and group the others {a 1, a 2,, a m, ã m+1 }, where ã m+1 = {a m+1, a m+2, } Rule of thumb: select m s.t. ˆf(am ) > ˆf(ã m+1 ) Continuous case: For a given x, consider (x h, x + h) for some h > 0, ˆf(x) = n i=1 1{X i (x h, x + h)} 2hn Jimin Ding, Math WUSTL Math 494 Spring / 44

41 Confidence Interval Jimin Ding, Math WUSTL Math 494 Spring / 44

42 Motivation for Confidence Interval In the previous Normal Model example, ˆµ ML = X maximizes the likelihood of observing the recorded data. But P (ˆµ = µ) =? Jimin Ding, Math WUSTL Math 494 Spring / 44

43 Motivation for Confidence Interval In the previous Normal Model example, ˆµ ML = X maximizes the likelihood of observing the recorded data. But P (ˆµ = µ) =? So we need an estimate of estimation error. In this example, we want to estimate SD(ˆµ) or V ar(ˆµ), to understand how much did ˆµ miss µ. In general, let θ be the parameter of interest and ˆθ be an estimator of θ. We call an estimator of the stardard deviation of ˆθ as standard error of ˆθ, and denoted by se(ˆθ). se(ˆθ) := ˆ SD(ˆθ). Jimin Ding, Math WUSTL Math 494 Spring / 44

44 Confidence Interval Interval Estimate: interval bounded by two values and used to estimate the population parameter of interest. (The bound could be infinity.) Some intervals are more useful than others: Jimin Ding, Math WUSTL Math 494 Spring / 44

45 Confidence Interval Interval Estimate: interval bounded by two values and used to estimate the population parameter of interest. (The bound could be infinity.) Some intervals are more useful than others: we want intervals to 1. have small length Jimin Ding, Math WUSTL Math 494 Spring / 44

46 Confidence Interval Interval Estimate: interval bounded by two values and used to estimate the population parameter of interest. (The bound could be infinity.) Some intervals are more useful than others: we want intervals to 1. have small length and 2.have a high chance to capture the true parameter. Jimin Ding, Math WUSTL Math 494 Spring / 44

47 Confidence Interval Interval Estimate: interval bounded by two values and used to estimate the population parameter of interest. (The bound could be infinity.) Some intervals are more useful than others: we want intervals to 1. have small length and 2.have a high chance to capture the true parameter. When we increase the interval we are always more confident that true parameter is inside it because we ve include more values. Jimin Ding, Math WUSTL Math 494 Spring / 44

48 Confidence Interval Interval Estimate: interval bounded by two values and used to estimate the population parameter of interest. (The bound could be infinity.) Some intervals are more useful than others: we want intervals to 1. have small length and 2.have a high chance to capture the true parameter. When we increase the interval we are always more confident that true parameter is inside it because we ve include more values. Confidence Level/Coefficient: the likelihood of an interval estimate to capture the population parameter of interest, often denoted by 1 α, for some α (0, 1). Jimin Ding, Math WUSTL Math 494 Spring / 44

49 Confidence Interval Interval Estimate: interval bounded by two values and used to estimate the population parameter of interest. (The bound could be infinity.) Some intervals are more useful than others: we want intervals to 1. have small length and 2.have a high chance to capture the true parameter. When we increase the interval we are always more confident that true parameter is inside it because we ve include more values. Confidence Level/Coefficient: the likelihood of an interval estimate to capture the population parameter of interest, often denoted by 1 α, for some α (0, 1). Confidence Interval (CI) Estimate: an interval estimate (L, U) that has a specified (and justified) level of confidence, say 100(1 α)%, where α [0, 1]. 1 α = P (θ (L, U)) Jimin Ding, Math WUSTL Math 494 Spring / 44

50 How to Interpret Confidence Intervals Since sample is random, so are confidence intervals. Just like an estimator. If I collect many different samples and so have many different 100(1 α)% CIs, then I expect 100(1 α)% of these intervals to capture the true parameter. Practical interpretation: I am 100(1 α)% confident that the interval contains/captures the population parameter (mean, variance,median...). The interpretation rests on the idea of repeated samples. Just like the sampling distributions. Note: once the sample is drawn, the realized value of the confidence interval is (l, u), an interval of real numbers, which is not random any longer. Jimin Ding, Math WUSTL Math 494 Spring / 44

51 How to Interpret Confidence Intervals Since sample is random, so are confidence intervals. Just like an estimator. If I collect many different samples and so have many different 100(1 α)% CIs, then I expect 100(1 α)% of these intervals to capture the true parameter. Practical interpretation: I am 100(1 α)% confident that the interval contains/captures the population parameter (mean, variance,median...). The interpretation rests on the idea of repeated samples. Just like the sampling distributions. Note: once the sample is drawn, the realized value of the confidence interval is (l, u), an interval of real numbers, which is not random any longer. It either contains θ or not. Jimin Ding, Math WUSTL Math 494 Spring / 44

52 Example: CI for µ Under normality Case 1: σ 2 is known. Case 2: σ 2 is unknown. (Example 4.2.1) Jimin Ding, Math WUSTL Math 494 Spring / 44

53 Pivotal Statistics In the construction of above CIs, we have used the statistics Z = X µ σ/ X µ and T = n S/. These statistics are called pivotal n statistics. Jimin Ding, Math WUSTL Math 494 Spring / 44

54 Pivotal Statistics In the construction of above CIs, we have used the statistics Z = X µ σ/ X µ and T = n S/. These statistics are called pivotal n statistics. They are random variables with the following two properties I. A function of the unknown parameter of interest, µ, but contains no other unknown parameters. II. The distribution of a pivotal statistic is known and free of any unknown parameters. Jimin Ding, Math WUSTL Math 494 Spring / 44

55 Pivotal Statistics In the construction of above CIs, we have used the statistics Z = X µ σ/ X µ and T = n S/. These statistics are called pivotal n statistics. They are random variables with the following two properties I. A function of the unknown parameter of interest, µ, but contains no other unknown parameters. II. The distribution of a pivotal statistic is known and free of any unknown parameters. It is a common technique to use pivotal statistics to construct confidence intervals and other statistical inference procedures in statistics. Jimin Ding, Math WUSTL Math 494 Spring / 44

56 Example: CI for µ Without normality Recall CLT: Let X 1, X 2,, X n iid with µ and σ 2 <. Then X µ σ/ n D N(0, 1), n. The result still hold if σ is replaced by a consistent estimator ˆσ (to be proved in Chapter 5), such as sample standard deviation ˆσ = S. Case 1: large sample (Example 4.2.2) Jimin Ding, Math WUSTL Math 494 Spring / 44

57 Example: CI for µ Without normality Case 2: large sample of Bernoulli rv (Example 4.2.3) Jimin Ding, Math WUSTL Math 494 Spring / 44

58 Example: CI for µ Without normality Case 3: small sample (Poisson Model) Jimin Ding, Math WUSTL Math 494 Spring / 44

59 Example: CI for µ 1 µ 2 Case 1: large sample Jimin Ding, Math WUSTL Math 494 Spring / 44

60 Example: CI for µ 1 µ 2 Case 2: under normality (Example 4.2.4) Jimin Ding, Math WUSTL Math 494 Spring / 44

61 Example: CI for p 1 p 2 in Bernoulli/Binomial Jimin Ding, Math WUSTL Math 494 Spring / 44

62 Hypothesis Testing Jimin Ding, Math WUSTL Math 494 Spring / 44

63 Motivation Estimation was concerned with taking the data and making guesses about what the parameter could be. What if we first guess what the parameter is, and decide whether the data support that guess? Our initial guess is called the null hypothesis, denoted by H 0. The alternative hypothesis, denoted by H 1, is the opposite of the null hypothesis. Hypothesis testing: a process to make a decision based data between two opposing hypotheses. Statistical hypotheses are certain explicit statements about population parameters. Jimin Ding, Math WUSTL Math 494 Spring / 44

64 Hypothesis Testing Hypothesis testing is an inference tool to use sample data to examine whether the statistical hypotheses are true. It is usually done in the following 5 steps. Jimin Ding, Math WUSTL Math 494 Spring / 44

65 Hypothesis Testing Hypothesis testing is an inference tool to use sample data to examine whether the statistical hypotheses are true. It is usually done in the following 5 steps. I. State null, H 0, and alternative, H 1, hypotheses. H 0 : neutral / no difference / no effect H 1 : a contradictory claim to H 0. It depends on the questions we want to answer. Jimin Ding, Math WUSTL Math 494 Spring / 44

66 Hypothesis Testing Hypothesis testing is an inference tool to use sample data to examine whether the statistical hypotheses are true. It is usually done in the following 5 steps. I. State null, H 0, and alternative, H 1, hypotheses. H 0 : neutral / no difference / no effect H 1 : a contradictory claim to H 0. It depends on the questions we want to answer. II. Specify the significance level: pre-determined a small positive number to bound p-value, denoted by α Jimin Ding, Math WUSTL Math 494 Spring / 44

67 Hypothesis Testing Hypothesis testing is an inference tool to use sample data to examine whether the statistical hypotheses are true. It is usually done in the following 5 steps. I. State null, H 0, and alternative, H 1, hypotheses. H 0 : neutral / no difference / no effect H 1 : a contradictory claim to H 0. It depends on the questions we want to answer. II. Specify the significance level: pre-determined a small positive number to bound p-value, denoted by α III. Computer test statistic: numerical summary of data which measures the deviation between H 0 and H 1, usually a pivotal statistic. Jimin Ding, Math WUSTL Math 494 Spring / 44

68 Hypothesis Testing Hypothesis testing is an inference tool to use sample data to examine whether the statistical hypotheses are true. It is usually done in the following 5 steps. I. State null, H 0, and alternative, H 1, hypotheses. H 0 : neutral / no difference / no effect H 1 : a contradictory claim to H 0. It depends on the questions we want to answer. II. Specify the significance level: pre-determined a small positive number to bound p-value, denoted by α III. Computer test statistic: numerical summary of data which measures the deviation between H 0 and H 1, usually a pivotal statistic. IV. Derive rejection region, or find p-value which is the probability of rejecting H 0 when it is actually true. Jimin Ding, Math WUSTL Math 494 Spring / 44

69 Hypothesis Testing Hypothesis testing is an inference tool to use sample data to examine whether the statistical hypotheses are true. It is usually done in the following 5 steps. I. State null, H 0, and alternative, H 1, hypotheses. H 0 : neutral / no difference / no effect H 1 : a contradictory claim to H 0. It depends on the questions we want to answer. II. Specify the significance level: pre-determined a small positive number to bound p-value, denoted by α III. Computer test statistic: numerical summary of data which measures the deviation between H 0 and H 1, usually a pivotal statistic. IV. Derive rejection region, or find p-value which is the probability of rejecting H 0 when it is actually true. V. Draw conclusions: reject H 0 if p-value is smaller than α or observed data fall in to rejection region. Jimin Ding, Math WUSTL Math 494 Spring / 44

70 Decisions in Hypothesis Testing Let X 1,, X n iid f(x; θ), θ Θ. Consider to test H 0 : θ Θ 0, v.s. H 1 : θ Θ 1, where Θ 0 Θ 1 = Θ and Θ 0 Θ 1 = Ø. There are two types of wrong decisions in hypothesis testing. When H 0 is true, we reject H 0. (False positive, Type I error) When H 1 is true, we fail to reject H 0. (False negative, Type II error) H 0 is true H 1 is true Jimin Ding, Math WUSTL Math 494 Spring / 44

71 Decisions in Hypothesis Testing Let X 1,, X n iid f(x; θ), θ Θ. Consider to test H 0 : θ Θ 0, v.s. H 1 : θ Θ 1, where Θ 0 Θ 1 = Θ and Θ 0 Θ 1 = Ø. There are two types of wrong decisions in hypothesis testing. When H 0 is true, we reject H 0. (False positive, Type I error) When H 1 is true, we fail to reject H 0. (False negative, Type II error) Decision: Reject H 0 Decision: Fail to Reject H 0 (Accept H 0 ) H 0 is true H 1 is true Jimin Ding, Math WUSTL Math 494 Spring / 44

72 Decisions in Hypothesis Testing Let X 1,, X n iid f(x; θ), θ Θ. Consider to test H 0 : θ Θ 0, v.s. H 1 : θ Θ 1, where Θ 0 Θ 1 = Θ and Θ 0 Θ 1 = Ø. There are two types of wrong decisions in hypothesis testing. When H 0 is true, we reject H 0. (False positive, Type I error) When H 1 is true, we fail to reject H 0. (False negative, Type II error) Decision: Reject H 0 H 0 is true H 1 is true Correct Decision Decision: Fail to Reject H 0 (Accept H 0 ) Correct Decision Jimin Ding, Math WUSTL Math 494 Spring / 44

73 Decisions in Hypothesis Testing Let X 1,, X n iid f(x; θ), θ Θ. Consider to test H 0 : θ Θ 0, v.s. H 1 : θ Θ 1, where Θ 0 Θ 1 = Θ and Θ 0 Θ 1 = Ø. There are two types of wrong decisions in hypothesis testing. When H 0 is true, we reject H 0. (False positive, Type I error) When H 1 is true, we fail to reject H 0. (False negative, Type II error) H 0 is true H 1 is true Decision: Type I error Correct Decision Reject H 0 Decision: Correct Decision Type II error Fail to Reject H 0 (Accept H 0 ) Jimin Ding, Math WUSTL Math 494 Spring / 44

74 Decisions in Hypothesis Testing Let X 1,, X n iid f(x; θ), θ Θ. Consider to test H 0 : θ Θ 0, v.s. H 1 : θ Θ 1, where Θ 0 Θ 1 = Θ and Θ 0 Θ 1 = Ø. There are two types of wrong decisions in hypothesis testing. When H 0 is true, we reject H 0. (False positive, Type I error) When H 1 is true, we fail to reject H 0. (False negative, Type II error) H 0 is true H 1 is true Decision: Type I error Correct Decision Reject H 0 Probability=α Decision: Correct Decision Type II error Fail to Reject H 0 Probability=β (Accept H 0 ) Jimin Ding, Math WUSTL Math 494 Spring / 44

75 Decisions in Hypothesis Testing Let X 1,, X n iid f(x; θ), θ Θ. Consider to test H 0 : θ Θ 0, v.s. H 1 : θ Θ 1, where Θ 0 Θ 1 = Θ and Θ 0 Θ 1 = Ø. There are two types of wrong decisions in hypothesis testing. When H 0 is true, we reject H 0. (False positive, Type I error) When H 1 is true, we fail to reject H 0. (False negative, Type II error) H 0 is true H 1 is true Decision: Type I error Correct Decision Reject H 0 Probability=α Power (1 β) Decision: Correct Decision Type II error Fail to Reject H 0 Probability= 1 α Probability=β (Accept H 0 ) Jimin Ding, Math WUSTL Math 494 Spring / 44

76 Decisions in Hypothesis Testing Let X 1,, X n iid f(x; θ), θ Θ. Consider to test H 0 : θ Θ 0, v.s. H 1 : θ Θ 1, where Θ 0 Θ 1 = Θ and Θ 0 Θ 1 = Ø. There are two types of wrong decisions in hypothesis testing. When H 0 is true, we reject H 0. (False positive, Type I error) When H 1 is true, we fail to reject H 0. (False negative, Type II error) H 0 is true H 1 is true Decision: Type I error Correct Decision Reject H 0 Probability=α Power (1 β) (False Positive) Decision: Correct Decision Type II error Fail to Reject H 0 Probability= 1 α Probability=β (Accept H 0 ) (False Negative) Jimin Ding, Math WUSTL Math 494 Spring / 44

77 Decisions in Hypothesis Testing Let X 1,, X n iid f(x; θ), θ Θ. Consider to test H 0 : θ Θ 0, v.s. H 1 : θ Θ 1, where Θ 0 Θ 1 = Θ and Θ 0 Θ 1 = Ø. There are two types of wrong decisions in hypothesis testing. When H 0 is true, we reject H 0. (False positive, Type I error) When H 1 is true, we fail to reject H 0. (False negative, Type II error) H 0 is true H 1 is true Decision: Type I error Correct Decision Reject H 0 Probability=α Power (1 β) (False Positive) (Correct Positive) Decision: Correct Decision Type II error Fail to Reject H 0 Probability= 1 α Probability=β (Accept H 0 ) (Correct Negative) (False Negative) Jimin Ding, Math WUSTL Math 494 Spring / 44

78 Decision Rule If (X 1,, X n ) C, then reject H 0 ; (rejection/critical region) If (X 1,, X n ) C c, then fail to reject H 0. (acceptance region) Goal: select C so that the probability of making errors are minimized. Typically we consider Type I error is more serious. So we first bound the probability of Type I error and then minimize the probability of Type II error. The probability of making type II error is denoted by β. Power: The probability of making a correct rejection, 1 β. Generally, the probability of making Type I error increases as the probability of making Type II error decreases. Jimin Ding, Math WUSTL Math 494 Spring / 44

79 Size of the Test and Power We say a rejection/critical region C is of size α if α = max θ Θ 0 P θ {(X 1,, X n ) C}, which is the upper bound of the probability of false rejection. Furthermore, power is defined as power = P θ {(X 1,, X n ) C c }, θ Θ 1, which is 1 probability of making Type II error. Jimin Ding, Math WUSTL Math 494 Spring / 44

80 Size of the Test and Power We say a rejection/critical region C is of size α if α = max θ Θ 0 P θ {(X 1,, X n ) C}, which is the upper bound of the probability of false rejection. Furthermore, power is defined as power = P θ {(X 1,, X n ) C c }, θ Θ 1, which is 1 probability of making Type II error. We see the power of a test depends on its critical region (rule). Denote the power function by r C (θ) = P θ {(X 1,, X n ) C c }, θ Θ. Jimin Ding, Math WUSTL Math 494 Spring / 44

81 Size of the Test and Power We say a rejection/critical region C is of size α if α = max θ Θ 0 P θ {(X 1,, X n ) C}, which is the upper bound of the probability of false rejection. Furthermore, power is defined as power = P θ {(X 1,, X n ) C c }, θ Θ 1, which is 1 probability of making Type II error. We see the power of a test depends on its critical region (rule). Denote the power function by r C (θ) = P θ {(X 1,, X n ) C c }, θ Θ. Given two critical regions, C 1 and C 2, which are both of size of α, we claim C 1 is better than C 2 if r C1 (θ) r C2 (θ), θ Θ 1. Jimin Ding, Math WUSTL Math 494 Spring / 44

82 Example 1: Binomial Model Jimin Ding, Math WUSTL Math 494 Spring / 44

83 Jimin Ding, Math WUSTL Math 494 Spring / 44

84 Example 2: Normal Model Jimin Ding, Math WUSTL Math 494 Spring / 44

85 P-value So far, we have derived decision rules and rejection regions using α. We see they depend on the choice of significance level α. But no data information, except sample size, was used in decision rules. Jimin Ding, Math WUSTL Math 494 Spring / 44

86 P-value So far, we have derived decision rules and rejection regions using α. We see they depend on the choice of significance level α. But no data information, except sample size, was used in decision rules. In practice, however, the data are often already collected, and one may not know a good choice of α in advance, or want to know all decision rules for a set values of α. For example, the observed sample mean is given x = 5, should we reject H 0 for α = 0.01, or α = 0.05, or α = Jimin Ding, Math WUSTL Math 494 Spring / 44

87 P-value So far, we have derived decision rules and rejection regions using α. We see they depend on the choice of significance level α. But no data information, except sample size, was used in decision rules. In practice, however, the data are often already collected, and one may not know a good choice of α in advance, or want to know all decision rules for a set values of α. For example, the observed sample mean is given x = 5, should we reject H 0 for α = 0.01, or α = 0.05, or α = In this case, a p-value instead of a rejection region is often used to provide more information. Precisely, p-value is, given that the H 0 is true, the conditional probability of observing more extreme data in the direction of H 1. The p-value can be viewed as an observed significance level. Jimin Ding, Math WUSTL Math 494 Spring / 44

88 Example: Zea mays Growth (Example 4.5.1,4.5.5) In 1878, Darwin recorded the heights of Zea mays plants to see the difference between cross- and self- fertilization. On each of the 15 plots, one cross-fertilized plant and one self-fertilized plant were planted to grow and then measured. The height difference of the cross-fertilized plant and the self-fertilized plant in the same plot is recorded. For 15 plots, the mean is x = 2.62 and standard deviation is s = Assume the difference in height is independent, and normally distributed. Is cross-fertilized plant taller than self-fertilized plant? Population Sample Probability model Jimin Ding, Math WUSTL Math 494 Spring / 44

89 Example: Zea mays Growth (Example 4.5.1,4.5.5) In 1878, Darwin recorded the heights of Zea mays plants to see the difference between cross- and self- fertilization. On each of the 15 plots, one cross-fertilized plant and one self-fertilized plant were planted to grow and then measured. The height difference of the cross-fertilized plant and the self-fertilized plant in the same plot is recorded. For 15 plots, the mean is x = 2.62 and standard deviation is s = Assume the difference in height is independent, and normally distributed. Is cross-fertilized plant taller than self-fertilized plant? Population Sample Probability model Parameter of interest Null and alternative hypotheses Test statistics P-value Conclusion Jimin Ding, Math WUSTL Math 494 Spring / 44

90 Example 3: F-test for Variances Recall in testing two sample means, to improve power, we assumed the variances of two normal samples were same. Is this assumption supported by data? Can we test the assumption formally? Jimin Ding, Math WUSTL Math 494 Spring / 44

91 Example 3: F-test for Variances Recall in testing two sample means, to improve power, we assumed the variances of two normal samples were same. Is this assumption supported by data? Can we test the assumption formally? Simpson, Olsen, and Eden (1975) describe an experiment, in which a random sample of 26 clouds were seeded with silver nitrate to see if they produced more rain than unseeded clouds. Suppose that on a log scale, the rainfall in both samples is approximated normal. And the sample means and sample variances are x = 5.13, ȳ = 3.99, s 2 x = 63.96, s 2 y = Jimin Ding, Math WUSTL Math 494 Spring / 44

92 Level, Power, P-value of F-test in One-sided Test Consider H 0 : σ 2 1 σ 2 2 v.s. H 1 : σ 2 1 > σ 2 2. Decision rule: reject H 0 in favor of H 1 if S 2 x/s 2 y c s.t. α = max P (Sx/S 2 σ1 2 y 2 c σ1, 2 σ2). 2,σ2 2 Jimin Ding, Math WUSTL Math 494 Spring / 44

93 Level, Power, P-value of F-test in One-sided Test Consider H 0 : σ 2 1 σ 2 2 v.s. H 1 : σ 2 1 > σ 2 2. Decision rule: reject H 0 in favor of H 1 if S 2 x/s 2 y c s.t. α = max P (Sx/S 2 σ1 2 y 2 c σ1, 2 σ2). 2,σ2 2 Proposition: Let V = S 2 x/s 2 y, c be the α upper tail quantile of F n 1,m 1, and G n 1,m 1 be the c.d.f. of F n 1,m 1 distribution. (That is, 1 α = G(c)). Then the power function of the test P (V c) satisfies: 1. P (V c) = 1 G n 1,m 1 (σ 2 2 /σ2 1 c). 2. P (V c) = α, if σ 2 1 = σ P (V c) < α, if σ 2 1 < σ P (V c) > α, if σ 2 1 > σ P (V c) 0, if σ 2 1 /σ P (V c) 1, if σ 2 1 /σ2 2. Jimin Ding, Math WUSTL Math 494 Spring / 44

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Review. December 4 th, Review

Review. December 4 th, Review December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

Statistical Inference

Statistical Inference Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will

More information

Smoking Habits. Moderate Smokers Heavy Smokers Total. Hypertension No Hypertension Total

Smoking Habits. Moderate Smokers Heavy Smokers Total. Hypertension No Hypertension Total Math 3070. Treibergs Final Exam Name: December 7, 00. In an experiment to see how hypertension is related to smoking habits, the following data was taken on individuals. Test the hypothesis that the proportions

More information

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis

More information

Business Statistics: Lecture 8: Introduction to Estimation & Hypothesis Testing

Business Statistics: Lecture 8: Introduction to Estimation & Hypothesis Testing Business Statistics: Lecture 8: Introduction to Estimation & Hypothesis Testing Agenda Introduction to Estimation Point estimation Interval estimation Introduction to Hypothesis Testing Concepts en terminology

More information

Probability and Statistics Notes

Probability and Statistics Notes Probability and Statistics Notes Chapter Seven Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Seven Notes Spring 2011 1 / 42 Outline

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Probability Theory and Statistics. Peter Jochumzen

Probability Theory and Statistics. Peter Jochumzen Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................

More information

BTRY 4090: Spring 2009 Theory of Statistics

BTRY 4090: Spring 2009 Theory of Statistics BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)

More information

Introductory Econometrics

Introductory Econometrics Session 4 - Testing hypotheses Roland Sciences Po July 2011 Motivation After estimation, delivering information involves testing hypotheses Did this drug had any effect on the survival rate? Is this drug

More information

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,

More information

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Summary: the confidence interval for the mean (σ 2 known) with gaussian assumption

Summary: the confidence interval for the mean (σ 2 known) with gaussian assumption Summary: the confidence interval for the mean (σ known) with gaussian assumption on X Let X be a Gaussian r.v. with mean µ and variance σ. If X 1, X,..., X n is a random sample drawn from X then the confidence

More information

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Hypothesis Testing Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Suppose the family of population distributions is indexed

More information

Basic Concepts of Inference

Basic Concepts of Inference Basic Concepts of Inference Corresponds to Chapter 6 of Tamhane and Dunlop Slides prepared by Elizabeth Newton (MIT) with some slides by Jacqueline Telford (Johns Hopkins University) and Roy Welsch (MIT).

More information

Practice Problems Section Problems

Practice Problems Section Problems Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,

More information

Dr. Maddah ENMG 617 EM Statistics 10/15/12. Nonparametric Statistics (2) (Goodness of fit tests)

Dr. Maddah ENMG 617 EM Statistics 10/15/12. Nonparametric Statistics (2) (Goodness of fit tests) Dr. Maddah ENMG 617 EM Statistics 10/15/12 Nonparametric Statistics (2) (Goodness of fit tests) Introduction Probability models used in decision making (Operations Research) and other fields require fitting

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

One-Sample Numerical Data

One-Sample Numerical Data One-Sample Numerical Data quantiles, boxplot, histogram, bootstrap confidence intervals, goodness-of-fit tests University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html

More information

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Fall, 2013 Page 1 Random Variable and Probability Distribution Discrete random variable Y : Finite possible values {y

More information

Lecture 8 Inequality Testing and Moment Inequality Models

Lecture 8 Inequality Testing and Moment Inequality Models Lecture 8 Inequality Testing and Moment Inequality Models Inequality Testing In the previous lecture, we discussed how to test the nonlinear hypothesis H 0 : h(θ 0 ) 0 when the sample information comes

More information

Math Review Sheet, Fall 2008

Math Review Sheet, Fall 2008 1 Descriptive Statistics Math 3070-5 Review Sheet, Fall 2008 First we need to know about the relationship among Population Samples Objects The distribution of the population can be given in one of the

More information

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis. 401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis

More information

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics A short review of the principles of mathematical statistics (or, what you should have learned in EC 151).

More information

Statistics. Statistics

Statistics. Statistics The main aims of statistics 1 1 Choosing a model 2 Estimating its parameter(s) 1 point estimates 2 interval estimates 3 Testing hypotheses Distributions used in statistics: χ 2 n-distribution 2 Let X 1,

More information

Statistical Inference

Statistical Inference Statistical Inference Bernhard Klingenberg Institute of Statistics Graz University of Technology Steyrergasse 17/IV, 8010 Graz www.statistics.tugraz.at February 12, 2008 Outline Estimation: Review of concepts

More information

Lecture 10: Generalized likelihood ratio test

Lecture 10: Generalized likelihood ratio test Stat 200: Introduction to Statistical Inference Autumn 2018/19 Lecture 10: Generalized likelihood ratio test Lecturer: Art B. Owen October 25 Disclaimer: These notes have not been subjected to the usual

More information

Tests about a population mean

Tests about a population mean October 2 nd, 2017 Overview Week 1 Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 1: Descriptive statistics Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter 8: Confidence

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Dover- Sherborn High School Mathematics Curriculum Probability and Statistics

Dover- Sherborn High School Mathematics Curriculum Probability and Statistics Mathematics Curriculum A. DESCRIPTION This is a full year courses designed to introduce students to the basic elements of statistics and probability. Emphasis is placed on understanding terminology and

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

HT Introduction. P(X i = x i ) = e λ λ x i

HT Introduction. P(X i = x i ) = e λ λ x i MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework

More information

Economics 583: Econometric Theory I A Primer on Asymptotics: Hypothesis Testing

Economics 583: Econometric Theory I A Primer on Asymptotics: Hypothesis Testing Economics 583: Econometric Theory I A Primer on Asymptotics: Hypothesis Testing Eric Zivot October 12, 2011 Hypothesis Testing 1. Specify hypothesis to be tested H 0 : null hypothesis versus. H 1 : alternative

More information

Lecture 4: Parameter Es/ma/on and Confidence Intervals. GENOME 560, Spring 2015 Doug Fowler, GS

Lecture 4: Parameter Es/ma/on and Confidence Intervals. GENOME 560, Spring 2015 Doug Fowler, GS Lecture 4: Parameter Es/ma/on and Confidence Intervals GENOME 560, Spring 2015 Doug Fowler, GS (dfowler@uw.edu) 1 Review: Probability DistribuIons Discrete: Binomial distribuion Hypergeometric distribuion

More information

Math 475. Jimin Ding. August 29, Department of Mathematics Washington University in St. Louis jmding/math475/index.

Math 475. Jimin Ding. August 29, Department of Mathematics Washington University in St. Louis   jmding/math475/index. istical A istic istics : istical Department of Mathematics Washington University in St. Louis www.math.wustl.edu/ jmding/math475/index.html August 29, 2013 istical August 29, 2013 1 / 18 istical A istic

More information

STAT 461/561- Assignments, Year 2015

STAT 461/561- Assignments, Year 2015 STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and

More information

Summary of Chapters 7-9

Summary of Chapters 7-9 Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two

More information

Masters Comprehensive Examination Department of Statistics, University of Florida

Masters Comprehensive Examination Department of Statistics, University of Florida Masters Comprehensive Examination Department of Statistics, University of Florida May 6, 003, 8:00 am - :00 noon Instructions: You have four hours to answer questions in this examination You must show

More information

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1 Math 66/566 - Midterm Solutions NOTE: These solutions are for both the 66 and 566 exam. The problems are the same until questions and 5. 1. The moment generating function of a random variable X is M(t)

More information

Hypothesis testing: theory and methods

Hypothesis testing: theory and methods Statistical Methods Warsaw School of Economics November 3, 2017 Statistical hypothesis is the name of any conjecture about unknown parameters of a population distribution. The hypothesis should be verifiable

More information

Political Science 236 Hypothesis Testing: Review and Bootstrapping

Political Science 236 Hypothesis Testing: Review and Bootstrapping Political Science 236 Hypothesis Testing: Review and Bootstrapping Rocío Titiunik Fall 2007 1 Hypothesis Testing Definition 1.1 Hypothesis. A hypothesis is a statement about a population parameter The

More information

Parametric Techniques Lecture 3

Parametric Techniques Lecture 3 Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to

More information

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section

More information

1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ).

1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Estimation February 3, 206 Debdeep Pati General problem Model: {P θ : θ Θ}. Observe X P θ, θ Θ unknown. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Examples: θ = (µ,

More information

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain 0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher

More information

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or

More information

Chapter 7. Hypothesis Testing

Chapter 7. Hypothesis Testing Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function

More information

Interval estimation. October 3, Basic ideas CLT and CI CI for a population mean CI for a population proportion CI for a Normal mean

Interval estimation. October 3, Basic ideas CLT and CI CI for a population mean CI for a population proportion CI for a Normal mean Interval estimation October 3, 2018 STAT 151 Class 7 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0

More information

MTH U481 : SPRING 2009: PRACTICE PROBLEMS FOR FINAL

MTH U481 : SPRING 2009: PRACTICE PROBLEMS FOR FINAL MTH U481 : SPRING 2009: PRACTICE PROBLEMS FOR FINAL 1). Two urns are provided as follows: urn 1 contains 2 white chips and 4 red chips, while urn 2 contains 5 white chips and 3 red chips. One chip is chosen

More information

Parametric Techniques

Parametric Techniques Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

Post-exam 2 practice questions 18.05, Spring 2014

Post-exam 2 practice questions 18.05, Spring 2014 Post-exam 2 practice questions 18.05, Spring 2014 Note: This is a set of practice problems for the material that came after exam 2. In preparing for the final you should use the previous review materials,

More information

STAT 4385 Topic 01: Introduction & Review

STAT 4385 Topic 01: Introduction & Review STAT 4385 Topic 01: Introduction & Review Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso xsu@utep.edu Spring, 2016 Outline Welcome What is Regression Analysis? Basics

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science

UNIVERSITY OF TORONTO Faculty of Arts and Science UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator

More information

Class 24. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Class 24. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700 Class 4 Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science Copyright 013 by D.B. Rowe 1 Agenda: Recap Chapter 9. and 9.3 Lecture Chapter 10.1-10.3 Review Exam 6 Problem Solving

More information

This does not cover everything on the final. Look at the posted practice problems for other topics.

This does not cover everything on the final. Look at the posted practice problems for other topics. Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

Glossary for the Triola Statistics Series

Glossary for the Triola Statistics Series Glossary for the Triola Statistics Series Absolute deviation The measure of variation equal to the sum of the deviations of each value from the mean, divided by the number of values Acceptance sampling

More information

ME3620. Theory of Engineering Experimentation. Spring Chapter IV. Decision Making for a Single Sample. Chapter IV

ME3620. Theory of Engineering Experimentation. Spring Chapter IV. Decision Making for a Single Sample. Chapter IV Theory of Engineering Experimentation Chapter IV. Decision Making for a Single Sample Chapter IV 1 4 1 Statistical Inference The field of statistical inference consists of those methods used to make decisions

More information

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1 4 Hypothesis testing 4. Simple hypotheses A computer tries to distinguish between two sources of signals. Both sources emit independent signals with normally distributed intensity, the signals of the first

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary Patrick Breheny October 13 Patrick Breheny Biostatistical Methods I (BIOS 5710) 1/25 Introduction Introduction What s wrong with z-tests? So far we ve (thoroughly!) discussed how to carry out hypothesis

More information

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30 MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)

More information

Classical and Bayesian inference

Classical and Bayesian inference Classical and Bayesian inference AMS 132 January 18, 2018 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 1 / 9 Sampling from a Bernoulli Distribution Theorem (Beta-Bernoulli

More information

2014/2015 Smester II ST5224 Final Exam Solution

2014/2015 Smester II ST5224 Final Exam Solution 014/015 Smester II ST54 Final Exam Solution 1 Suppose that (X 1,, X n ) is a random sample from a distribution with probability density function f(x; θ) = e (x θ) I [θ, ) (x) (i) Show that the family of

More information

Performance Evaluation and Comparison

Performance Evaluation and Comparison Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Cross Validation and Resampling 3 Interval Estimation

More information

Statistical Inference. Hypothesis Testing

Statistical Inference. Hypothesis Testing Statistical Inference Hypothesis Testing Previously, we introduced the point and interval estimation of an unknown parameter(s), say µ and σ 2. However, in practice, the problem confronting the scientist

More information

Review and continuation from last week Properties of MLEs

Review and continuation from last week Properties of MLEs Review and continuation from last week Properties of MLEs As we have mentioned, MLEs have a nice intuitive property, and as we have seen, they have a certain equivariance property. We will see later that

More information

Better Bootstrap Confidence Intervals

Better Bootstrap Confidence Intervals by Bradley Efron University of Washington, Department of Statistics April 12, 2012 An example Suppose we wish to make inference on some parameter θ T (F ) (e.g. θ = E F X ), based on data We might suppose

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Chapter 8 of Devore , H 1 :

Chapter 8 of Devore , H 1 : Chapter 8 of Devore TESTING A STATISTICAL HYPOTHESIS Maghsoodloo A statistical hypothesis is an assumption about the frequency function(s) (i.e., PDF or pdf) of one or more random variables. Stated in

More information

ST 371 (IX): Theories of Sampling Distributions

ST 371 (IX): Theories of Sampling Distributions ST 371 (IX): Theories of Sampling Distributions 1 Sample, Population, Parameter and Statistic The major use of inferential statistics is to use information from a sample to infer characteristics about

More information

Introduction to Statistical Inference

Introduction to Statistical Inference Introduction to Statistical Inference Dr. Fatima Sanchez-Cabo f.sanchezcabo@tugraz.at http://www.genome.tugraz.at Institute for Genomics and Bioinformatics, Graz University of Technology, Austria Introduction

More information

Some Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2

Some Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2 STA 248 H1S MIDTERM TEST February 26, 2008 SURNAME: SOLUTIONS GIVEN NAME: STUDENT NUMBER: INSTRUCTIONS: Time: 1 hour and 50 minutes Aids allowed: calculator Tables of the standard normal, t and chi-square

More information

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45 Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS 21 June 2010 9:45 11:45 Answer any FOUR of the questions. University-approved

More information

Primer on statistics:

Primer on statistics: Primer on statistics: MLE, Confidence Intervals, and Hypothesis Testing ryan.reece@gmail.com http://rreece.github.io/ Insight Data Science - AI Fellows Workshop Feb 16, 018 Outline 1. Maximum likelihood

More information

Parametric Models: from data to models

Parametric Models: from data to models Parametric Models: from data to models Pradeep Ravikumar Co-instructor: Manuela Veloso Machine Learning 10-701 Jan 22, 2018 Recall: Model-based ML DATA MODEL LEARNING MODEL MODEL INFERENCE KNOWLEDGE Learning:

More information

Statistics 135 Fall 2008 Final Exam

Statistics 135 Fall 2008 Final Exam Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations

More information

Robustness and Distribution Assumptions

Robustness and Distribution Assumptions Chapter 1 Robustness and Distribution Assumptions 1.1 Introduction In statistics, one often works with model assumptions, i.e., one assumes that data follow a certain model. Then one makes use of methodology

More information

Statistical inference

Statistical inference Statistical inference Contents 1. Main definitions 2. Estimation 3. Testing L. Trapani MSc Induction - Statistical inference 1 1 Introduction: definition and preliminary theory In this chapter, we shall

More information

Psychology 282 Lecture #4 Outline Inferences in SLR

Psychology 282 Lecture #4 Outline Inferences in SLR Psychology 282 Lecture #4 Outline Inferences in SLR Assumptions To this point we have not had to make any distributional assumptions. Principle of least squares requires no assumptions. Can use correlations

More information

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology Singapore University of Design and Technology Lecture 9: Hypothesis testing, uniformly most powerful tests. The Neyman-Pearson framework Let P be the family of distributions of concern. The Neyman-Pearson

More information

Probability and Statistics

Probability and Statistics Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be CHAPTER 4: IT IS ALL ABOUT DATA 4a - 1 CHAPTER 4: IT

More information

An interval estimator of a parameter θ is of the form θl < θ < θu at a

An interval estimator of a parameter θ is of the form θl < θ < θu at a Chapter 7 of Devore CONFIDENCE INTERVAL ESTIMATORS An interval estimator of a parameter θ is of the form θl < θ < θu at a confidence pr (or a confidence coefficient) of 1 α. When θl =, < θ < θu is called

More information

The Components of a Statistical Hypothesis Testing Problem

The Components of a Statistical Hypothesis Testing Problem Statistical Inference: Recall from chapter 5 that statistical inference is the use of a subset of a population (the sample) to draw conclusions about the entire population. In chapter 5 we studied one

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

TUTORIAL 8 SOLUTIONS #

TUTORIAL 8 SOLUTIONS # TUTORIAL 8 SOLUTIONS #9.11.21 Suppose that a single observation X is taken from a uniform density on [0,θ], and consider testing H 0 : θ = 1 versus H 1 : θ =2. (a) Find a test that has significance level

More information

Ch. 1: Data and Distributions

Ch. 1: Data and Distributions Ch. 1: Data and Distributions Populations vs. Samples How to graphically display data Histograms, dot plots, stem plots, etc Helps to show how samples are distributed Distributions of both continuous and

More information

Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants

Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants 18.650 Statistics for Applications Chapter 5: Parametric hypothesis testing 1/37 Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009

More information

New Bayesian methods for model comparison

New Bayesian methods for model comparison Back to the future New Bayesian methods for model comparison Murray Aitkin murray.aitkin@unimelb.edu.au Department of Mathematics and Statistics The University of Melbourne Australia Bayesian Model Comparison

More information

Business Statistics. Lecture 10: Course Review

Business Statistics. Lecture 10: Course Review Business Statistics Lecture 10: Course Review 1 Descriptive Statistics for Continuous Data Numerical Summaries Location: mean, median Spread or variability: variance, standard deviation, range, percentiles,

More information

Estimation of Quantiles

Estimation of Quantiles 9 Estimation of Quantiles The notion of quantiles was introduced in Section 3.2: recall that a quantile x α for an r.v. X is a constant such that P(X x α )=1 α. (9.1) In this chapter we examine quantiles

More information

F79SM STATISTICAL METHODS

F79SM STATISTICAL METHODS F79SM STATISTICAL METHODS SUMMARY NOTES 9 Hypothesis testing 9.1 Introduction As before we have a random sample x of size n of a population r.v. X with pdf/pf f(x;θ). The distribution we assign to X is

More information