Central Limit Theorem ( 5.3)

Size: px
Start display at page:

Download "Central Limit Theorem ( 5.3)"

Transcription

1 Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately normal with mean nµ and variance nσ 2 as n, that is, ( ) Sn nµ P σ a P(Z a) as n, for < a <, n where Z N(0, 1). Similarly, the distribution of the sample mean X n = 1 Sn becomes n approximately N(µ, σ 2 /n) as n, that is, P ( ) X µ σ/ n a P(Z a) as n, for < a <. Related homework: 1/10, 1/13

2 χ 2, t, and F distributions ( 6.2) Let Z 1, Z 2,..., Z n be independent standard normal random variables and defined X = Z Z Z 2 n. Then the distribution of X is called the chi-square distribution with degrees of freedom n, and is denoted by χ 2 n. Let Z and U be two independent random variables with Z N(0, 1) and U χ 2 n. Then the distribution of the random variable T = Z U/n is called the t distribution with degrees of freedom n, and is denoted by t n. Let U and V be two independent random variables with U χ 2 n and U χ2 n. Then the distribution of the random variable F = U/m V /n is called the F distribution with degrees of freedom m and n, and is denoted by F m,n. Related homework: 1/15, 1/17

3 Sample mean and sample variance ( 6.3) Let X 1, X 2,..., X n be a sequence of i.i.d. random variables (a random sample), each having mean µ and variance σ 2. The sample mean and sample variance are defined as n X = 1 n i=1 respectively. Properties of X and S 2 : X i and S 2 = 1 n 1 n (X i X ) 2, E[X ] = µ, Var[X ] = σ2 n, and E[S2 ] = σ 2. X and S 2 are independent. If the random sample is from a normal distribution, then i=1 X µ S/ n t n 1, and (n 1)S 2 σ 2 χ 2 n 1. Related homework: 1/17, 1/22

4 MME (Method of Moments Estimate) ( 8.4) Let X 1, X 2,..., X n be a random sample from a probability distribution with parameter θ. The method of moments estimate is based on the law of large numbers: n ˆµ k = 1 n i=1 Xi k µ k = E[X1 k ] as n, that is, the kth sample moment ˆµ k converges to the kth moment µ k. Thus we can use ˆµ k to estimate µ k. If the parameter θ can be determined by the moments: θ = g(µ 1,... ), then the MME for θ is ˆθ = g(ˆµ 1,... ). Whenever lower moments are sufficient to determine θ, we do not use higher moments. Related homework: 1/22, 1/24

5 MLE (Maximum Likelihood Estimate) ( 8.5) Let X 1, X 2,..., X n be a random sample from a probability distribution with parameter θ. The maximum likelihood estimate is based on the principle of maximizing the likelihood function (joint pdf/pmf) of the observed sample: lik(θ) = f (X 1, X 2,..., X n θ) = n f (X i θ). (Treating X 1,..., X n as constants and θ as the variable.) Thus the MLE ˆθ for θ satisfies lik(ˆθ) = max {lik(θ)}. θ Usually, it is easier to maximize the log-likelihood function (via calculus) i=1 l(θ) = log[lik(θ)] = log f (X 1, X 2,..., X n θ) = n log f (X i θ). In some cases (when the support of the pdf/pmf depending on θ), we must maximize the likelihood function directly. E.g. Unif(0, θ). Related homework: 1/22, 1/24 i=1

6 Properties of MLE ( 8.5) Let X be a random variable from a probability distribution with parameter θ. The Fisher information for θ is [ ] [ 2 ] I (θ) = E log f (X θ) = E log f (X θ). θ θ2 Let X 1, X 2,..., X n be a random sample from a probability distribution with parameter θ, and ˆθ be the MLE for θ. Then the asymptotic variance of ˆθ is 1 ni (θ). Moreover, ni (θ)(ˆθ θ) becomes approximately N(0, 1) as n, that is, ( ) ˆθ θ P a P(Z a) as n, for < a <, 1/(nI (θ)) where Z N(0, 1). Related homework: 1/29

7 Properties of point estimates ( 8.7) Let X 1, X 2,..., X n be a random sample from a probability distribution with parameter θ, and ˆθ be a point estimate (e.g. MLE, MME) for θ. The bias of the point estimate ˆθ is b(ˆθ) = E[ˆθ] θ. The point estimate ˆθ is said to be unbiased if b(ˆθ) = 0. The mean sqaured error of the point estimate ˆθ is Cramér-Rao lower bound: where I (θ) is the Fisher information. Related homework: 1/27 MSE(ˆθ) = E[(ˆθ θ) 2 ] = Var[ˆθ] + b(ˆθ) 2. Var[ˆθ] 1 ni (θ),

8 Interval estimation confidence intervals ( 8.5) Let X 1, X 2,..., X n be a random sample from a probability distribution with parameter θ, and ˆθ be the MLE for θ. An approximate confidence interval for θ with confidence level 100p% is ( ) 1 1 ˆθ z 1+p 2 ni (ˆθ), ˆθ + z 1+p, 2 ni (ˆθ) where I (θ) is the Fisher information, and P(Z z p) = p for Z N(0, 1). Let X 1, X 2,..., X n be a random sample from a normal distribution N(µ, σ). A confidence interval for µ with confidence level 100p% is X t 1+p 2,n 1 S 2 n, X + t S 2 1+p 2,n 1, n where X and S 2 are the sample mean and sample variance, respectively, and P(T t p,m) = p for T t m. A confidence interval for σ 2 with confidence level 100p% is (n 1)S2 χ 2, 1+p 2,n 1 where P(U χ 2 p,m) = p for U χ 2 m. Related homework: 2/3, 2/5 (n 1)S 2, χ 2 1 p,n 1 2

9 Sufficient statistics ( 8.8) Let X 1, X 2,..., X n be a random sample from a probability distribution with parameter θ, and T = T (X 1,..., X n) be a statistic. The statistic T is said to be sufficient for the parameter θ if the conditional joint distribution of X 1,..., X n given T = t no longer depends on θ for all possible t. Factorization Theorem A statistic T is sufficient for the parameter θ if and only if f (x 1,..., x n θ) = g(t (x 1,..., x n), θ)h(x 1,..., x n), for some function g(t, θ) and h, where f (x 1,..., x n) is the joint pdf/pmf for X 1,..., X n. A probability distribution with parameter θ is said to belonging to the exponential family if its pdf/pmf is of the form f (x θ) = Then T = { e c(θ)t (x)+d(θ)+s(x), x A 0, x / A A does not depend on θ. n T (X i ) is a sufficient statistic for θ, where X 1,..., X n is a random i=1 sample. Related homework: 2/7, 2/10

10 General hypothesis testing ( ) A hypothesis is a statement about the population distribution. If a hypothesis completely specifies the distribution, it is called a simple hypothesis. (e.g. µ = µ 0.) If a hypothesis partially specifies the distribution, it is called a composite hypothesis. (e.g. µ > µ 0.) Typically, H 0 is chosen to be the more specific hypothesis. A test for the hypotheses H 0 and H A consists of a test statistic T and a rejection region R: if T R, then we reject H 0 ; if T / R, then we do not reject H 0. Consequence of a test decision Fact Related homework: 2/17, 2/19 Decision Reject H 0 Do not reject H 0 H 0 is true Type I error Correct decision H 0 is false Correct decision Type II error

11 General hypothesis testing ( ) The significance level α of a test is the probability of making Type I error of the given test: α = P(reject H 0 H 0 is true) = P(T R H 0 is true). The probability of making Type II error of a test is denoted by β: β = P(do not reject H 0 H 0 is false) = P(T / R H 0 is false). The power of a test is the probability (power) of detecting a false H 0, and it equals to 1 β: power = 1 β = P(reject H 0 H 0 is false) = P(T R H 0 is false). Let X 1,..., X n be a random sample and T (X 1,..., X n) = t. The p-value of the sample X 1,..., X n is the smallest significance level α = α(t) corresponding to the rejection region R = R(t) such that t R(t) (meaning rejecting H 0 based on T = t). Related homework: 2/17, 2/19

12 Likelihood ratio rest ( ) For the hypotheses H 0 : θ = θ 0 and H A : θ = θ 1, the likelihood ratio test based on a random sample X 1,... X n has test statistic Λ = lik(x 1,..., X n θ = θ 0 ) lik(x 1,..., X n θ = θ 1 ), and rejection region R = {Λ < c}. For the likelihood ratio test, the significance level is α = P(Λ < c θ = θ 0 ); the probability of making Type II error is β = P(Λ c θ = θ 1 ); power = 1 β = P(Λ < c θ = θ 1 ); and if Λ(X 1,..., X n) = λ, then the p-value of the sample is p = P(Λ < λ θ = θ 0 ). Related homework: 2/19, 2/21

13 Generalized likelihood ratio test ( 9.4) For the hypotheses H 0 : θ ω 0 and H A : θ ω 1, the generalized likelihood ratio test based on a random sample X 1,... X n has test statistic Λ = max θ ω 0 lik(x 1,..., X n θ) max θ Ω lik(x 1,..., X n θ), where Ω = ω 0 ω 1, and rejection region R = {Λ < c}. For the generalized likelihood ratio test, max θ ω0 lik(x 1,..., X n θ) = lik( θ), where θ is the mle of θ under the restriction of θ ω 0 ; and max θ Ω lik(x 1,..., X n θ) = lik(ˆθ), where ˆθ is the mle of θ under the restriction of θ Ω. Under certain conditions, when the sample size n is large, the distribution of 2 log Λ under H 0 is approximately χ 2 df, where df = dimω dimω 0 and dim refers to the number of free parameters. Related homework: 2/19, 2/21

14 Inference for µ based on normal model known σ 2 Let X 1, X 2,..., X n be a random sample from a normal distribution with mean µ and variance σ 2 (σ 2 is known). A 100(1 α)% confidence interval for µ is ( X z 1 α 2 σ n, X + z 1 α 2 ) σ. n The test for H 0 : µ = µ 0 v.s. H A has test statistic T = X µ 0 σ/ n and H A : µ > µ 0 H A : µ < µ 0 H A : µ µ 0 Rejection Region {T > c} {T < c} { T > c} α (given c) 1 Φ(c) Φ(c) 2[1 Φ(c)] c (given α) z 1 α z α z 1 α 2 p-value (given T = t) 1 Φ(t) Φ(t) 2[1 Φ( t )] β (given µ = µ 1 and α) Ψ (z 1 α ) 1 Ψ (z α) Ψ ( where Ψ(z) = Φ z + µ 0 µ 1 σ/ n Related homework: 2/24 (z 1 α2 ) Ψ ), and Φ is the cdf of N(0, 1). (z α2 )

15 Inference for µ based on normal model unknown σ 2 Let X 1, X 2,..., X n be a random sample from a normal distribution with mean µ and variance σ 2 (σ 2 is unknown). A 100(1 α)% confidence interval for µ is ( X t 1 α 2,n 1 S n, X + t 1 α 2,n 1 ) S. n The test for H 0 : µ = µ 0 v.s. H A has test statistic T = X µ 0 S/ n and H A : µ > µ 0 H A : µ < µ 0 H A : µ µ 0 Rejection Region {T > c} {T < c} { T > c} α (given c) 1 F n 1 (c) F n 1 (c) 2[1 F n 1 (c)] c (given α) t 1 α,n 1 t α,n 1 t 1 α 2,n 1 p-value (given T = t) 1 F n 1 (t) F n 1 (t) 2[1 F n 1 ( t )] where F n 1 is the cdf of the t-distribution with degrees of freedom n 1. Related homework: 2/26

16 Test for goodness-of-fit ( 9.5) Setting: Assume that the population contains m categories and the probability that a random observation being category i is p i. A random sample of size n contains X i observations of category i. (Thus the X i s follow a multinomial distribution with parameters n and p i s.) Hypotheses: H 0 : p i = p i (θ), and H A : H 0 is not true. In words, the null hypothesis H 0 specifies a model for the p i s. The generalized likelihood ratio test: test statistic Λ = max p= p(θ) lik(x 1,..., X m p) max p 1 =1 lik(x 1,..., X m p) = m i=1 ( ) Xi p i (ˆθ), where ˆp i p = (p 1,..., p m), ˆθ is the mle for θ, and ˆp i = X i /n is the mle for p i subject to p i = 1; rejection region R = {Λ < c} and 2 log Λ is approximately χ 2 df with df = (m 1) dimθ. An equivalent test (Pearson s χ 2 test) m test statistic X 2 (O i E i ) 2 =, where O i = X i represents the observed E i=1 i counts and E i = n p i (ˆθ) represents the expected counts; rejection region R = {X 2 > c} and X 2 is approximately χ 2 df with df = (m 1) dimθ. Related homework: 3/10

17 Inference for µ X µ Y based on normal model with two independent samples known σ 2 ( 11.2) Let X 1,..., X n be a random sample from N(µ X, σ 2 ); Y 1,..., Y m be a random sample from N(µ Y, σ 2 ). X s and Y s are independent. (Assume σ 2 is known.) A 100(1 α)% confidence interval for µ X µ Y is ( 1 (X Y ) z 1 α σ 2 n + 1 ) 1 m, (X Y ) + z 1 α σ 2 n + 1, m Let = µ X µ Y. Test for H 0 : = 0 v.s. H A has test statistic T = (X Y ) 0 and 1 σ n + 1 m H A : > 0 H A : < 0 H A : 0 Rejection Region {T > c} {T < c} { T > c} α (given c) 1 Φ(c) Φ(c) 2[1 Φ(c)] c (given α) z 1 α z α z 1 α 2 p-value (given T = t) 1 Φ(t) Φ(t) 2[1 Φ( t )] β (given = 1 and α) Ψ (z 1 α ) 1 Ψ (z α) Ψ ( where Ψ(z) = Φ z σ/ n Related homework: 3/14 (z 1 α2 ) Ψ ), and Φ is the cdf of N(0, 1). (z α2 )

18 Inference for µ X µ Y based on normal model with two independent samples unknown σ 2 ( 11.2) Let X 1,..., X n be a random sample from N(µ X, σ 2 ); Y 1,..., Y m be a random sample from N(µ Y, σ 2 ). X s and Y s are independent. (Assume σ 2 is unknown.) A 100(1 α)% confidence interval for µ X µ Y is ( (X Y ) t 1 α 2,n+m 2 S p 1 n + 1 m, (X Y ) + t 1 α 2,n+m 2 S p 1 n + 1 m ), where S p is the pooled standard error: Sp 2 = (n 1)S2 X + (m 1)S2 Y. n + m 2 Let = µ X µ Y. Test for H 0 : = 0 v.s. H A has test statistic T = (X Y ) 0 and 1 S p n + 1 m H A : > 0 H A : < 0 H A : 0 Rejection Region {T > c} {T < c} { T > c} α (given c) 1 F n+m 2 (c) F n+m 2 (c) 2[1 F n+m 2 (c)] c (given α) t 1 α,n+m 2 t α,n+m 2 t 1 α 2,n+m 2 p-value (given T = t) 1 F n+m 2 (t) F n+m 2 (t) 2[1 F n+m 2 ( t )] where F n+m 2 is the cdf of the t-distribution with degrees of freedom n + m 2. Related homework: 3/14

19 Test for comparing two populations Wilcoxon rank-sum test ( 11.2) Let X 1,..., X n be a random sample from a population with cdf F and Y 1,..., Y m be a random sample from a population with cdf G. The hypotheses are H 0 : F = G and H A : F G. Wilcoxon rank-sum test (also called Mann-Whiteny test) Order the observations X i and Y j, and assign ranks (1 through n + m) to each observation according to their order. Let R(Z) denote the rank of the observation Z. Assume that m < n. The test statistic is T Y = m j=1 R(Y j ). The rejection region is {T Y < c 1 or T Y > c 2 }. The distribution of T Y under H 0 can be determined from combinatorics. For example, the pmf of T Y for n = m = 2 is p(3) = p(4) = p(6) = p(7) = 1 6 and p(5) = 1 5. In practice, we apply symmetry and use test statistic R = min(r, R ), where R = T Y and R = m(n + m + 1) R (assuming m < n); and rejection region {R < c} (TABLE 8 of textbook). Related homework: 3/17

20 Inference for µ X µ Y based on normal model with matched pair design ( 11.3) Let X 1,..., X n be a random sample from a population with mean µ X and Y 1,..., Y n be a random sample from a population with mean µ Y. X i and Y i are paired for each 1 i n. The differences D i = X i Y i can be regarded as a random sample from a population with mean µ X µ Y. Furthermore, D i s are assumed to be a random sample from N(d, σ 2 ), where d = µ X µ Y. The inference methods for d is exactly the same as Inference for µ based on normal model. Related homework: 3/21

21 Test for comparing two populations with matched pair design signed rank test ( 11.3) Let X 1,..., X n be a random sample from a population with cdf F and Y 1,..., Y n be a random sample from a population with cdf G. X i and Y i are paired for each 1 i n. Let D i = X i Y i be the differences. The hypotheses are H 0 : D i s are symmetric about 0, and H A : D i s are not symmetric about 0. Signed rank test Order the magnitude of the differences D i, and assign ranks (1 through n) to each one according to their order. Let R(D i ) denote the rank of D i. The test statistic is W + = n i=1 1 (0, )(D i ) R(D i ), where 1 (0, ) (x) = 1 if x > 0 and 0 otherwise. The rejection region is {W + < c 1 or W + > c 2 }. The distribution of W + under H 0 can be determined from combinatorics. For example, the pmf of W + for n = 2 is p(0) = p(1) = p(2) = p(3) = 1 4. In practice, we apply symmetry and use test statistic W = min(w +, W ), where W = n(n + 1)/2 W +; and rejection region {W < c} (TABLE 9 of textbook). Related homework: 3/24

22 One-way ANOVA ( 12.2) setting Consider I groups (populations). For each group, a random sample of size J is drawn. Let Y ij denote the jth observation in the ith sample. The statistical model is Y ij = µ + α i + ε ij, where µ is the overall average of the I groups, α i is the correction of the ith group, and ε ij s are i.i.d. N(0, σ 2 ) random variables (errors). The sum of squares between groups measures the variation between the I samples: SS B = J I (Y i Y ) 2, and i=1 SS B σ 2 χ 2 I 1 if α i = 0 for all 1 i I, where Y i = 1 J Y ij and Y = 1 I J Y ij. J IJ j=1 i=1 j=1 The sum of squares within groups measures the overall variation inside the I samples: I J SS W = (Y ij Y i ) 2 SS W, and σ 2 χ 2 I (J 1). i=1 j=1 The total sum of squares measures the overall variation of the I samples: SS T = I i=1 j=1 J (Y ij Y ) 2 = SS B + SS W.

23 One-way ANOVA ( 12.2) F test The hypotheses of one-way ANOVA: H 0 : α i = 0 for all 1 i I H A : H 0 is false The F test: Intuition: if the variation between groups (SS B ) is large relative to the variation within groups (SS W ), then H 0 can not be true. Test statistic: F = SS B/(I 1) SS W /(I (J 1)), and F F (I 1, I (J 1)) under H 0. Rejection region: R = {F > c} with c = F 1 α (I 1, I (J 1)), where α is the significance level. The ANOVA table: Sum of Source df Squares Mean Square F Between Groups I 1 SS B MS B = SS B I 1 Within Groups I (J 1) S W MS W = SS W I (J 1) Total IJ 1 SS T Related homework: 4/2, 3/31 F = MS B MS W

24 Application of χ 2 test ( 13.3) test of homogeneity Consider I populations, each containing J categories. A random sample of size N is drawn from these populations: Population 1 Population I Total Category 1 n 11 n I 1 n Category J n 1J n IJ n J Total n 1 n I n where n ij is the number of observations of category j from the ith population, n i = j n ij, n j = i n ij, and N = n = i j n ij. Let p ij be the proportion of category j in population i. The hypotheses of test of homogeneity are: H 0 : p 1j = = p Ij for all 1 j J H A : H 0 is false The χ 2 test (test of goodness-of-fit): Test statistic: I J X 2 (O ij E ij ) 2 =, E i=1 j=1 ij and X 2 χ 2 (I 1)(J 1) under H 0. where O ij = n ij and E ij = n i n j. n Rejection region: R = {X 2 > c} with c = χ 2, where α is the 1 α,(i 1)(J 1) significance level. Related homework: 4/9

25 Application of χ 2 test ( 13.4) test of independence Consider two discrete random variables U and V. U has I possible values, with marginal pmf P(U = u i ) = p i ; and V has J possible values, with marginal pmf P(V = v j ) = q j. A random sample of size N is drawn from the population: u 1 u I Total v 1 n 11 n I 1 n v J n 1J n IJ n J Total n 1 n I n where n ij is the number of observations of the pair (u i, v j ), n i = j n ij, n j = i n ij, and N = n = i j n ij. Let the joint pmf be P(U = u i, V = v j ) = π ij. The hypotheses of test of independence are: H 0 : π ij = p i q j for all 1 i I, 1 j J H A : H 0 is false The χ 2 test (test of goodness-of-fit): I J Test statistic is X 2 (O ij E ij ) 2 =, and X 2 χ 2 E (I 1)(J 1) under i=1 j=1 ij H 0, where O ij = n ij and E ij = n i n j. n Rejection region: R = {X 2 > c} with c = χ 2, where α is the 1 α,(i 1)(J 1) significance level. Related homework: 4/9

26 Simple linear regression ( 14.1) The statistical model: y i = β 0 + β 1 x i + ε i, where ε i s are i.i.d. N(0, σ 2 ) random variables (error), β 0, β 1, and x i s are nonrandom constants. Given sample data (x 1, y 1 ),..., (x n, y n), we use the least square principle to find estimates ˆβ 0 and ˆβ 1 for β 0 and β 1, respectively, that is, we minimize the Residual Sum of Squares (RSS): Consequently, we have RSS = Related homework: 4/11, 4/9 n (y i ˆβ 0 ˆβ 1 x i ) 2. i=1 ( n ) ( n i=1 ˆβ 0 = x2 i i=1 y ) ( i n i=1 x ) ( n i i=1 x ) i y i n n i=1 x2 i ( n i=1 x ) 2, i ˆβ 1 = n n i=1 x i y i ( n i=1 x ) ( n i i=1 y ) i n n i=1 x2 i ( n i=1 x ) 2. i

27 Simple linear regression ( 14.1) continued The least square estimates ˆβ 0, ˆβ 1 are unbiased estimates, that is, Further more, Var[ ˆβ 0 ] = E[ ˆβ 0 ] = β 0, and E[ ˆβ 1 ] = β 1. σ 2 n i=1 x2 i n n i=1 x2 i ( n i=1 x ) 2, Var[ ˆβ 1 ] = i The error variance σ 2 can be estimated by nσ 2 n n i=1 x2 i ( n i=1 x i s 2 = RSS [, where RSS = y 2 i 1 ( ) ] 2 yi [n x i y i ( x i ) ( y i )] 2 n 2 n n 2 xi 2 n ( x i ) 2 Consequently, the estimated variance of ˆβ 0 and ˆβ 1 are s 2ˆβ 0 = s 2 n i=1 x2 i n n i=1 x2 i ( n i=1 x ) 2, s 2ˆβ = 1 i ns 2 n n i=1 x2 i ( n i=1 x i ) 2. ) 2. Moreover, ˆβ 0 β 0 t n 2, and ˆβ 1 β 1 t n 2 s ˆβ s 0 ˆβ 1 Related homework: 4/14

Review. December 4 th, Review

Review. December 4 th, Review December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter

More information

Mathematical statistics

Mathematical statistics October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Institute of Actuaries of India

Institute of Actuaries of India Institute of Actuaries of India Subject CT3 Probability & Mathematical Statistics May 2011 Examinations INDICATIVE SOLUTION Introduction The indicative solution has been written by the Examiners with the

More information

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood Regression Estimation - Least Squares and Maximum Likelihood Dr. Frank Wood Least Squares Max(min)imization Function to minimize w.r.t. β 0, β 1 Q = n (Y i (β 0 + β 1 X i )) 2 i=1 Minimize this by maximizing

More information

Summary of Chapters 7-9

Summary of Chapters 7-9 Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two

More information

Statistics. Statistics

Statistics. Statistics The main aims of statistics 1 1 Choosing a model 2 Estimating its parameter(s) 1 point estimates 2 interval estimates 3 Testing hypotheses Distributions used in statistics: χ 2 n-distribution 2 Let X 1,

More information

Theory of Statistics.

Theory of Statistics. Theory of Statistics. Homework V February 5, 00. MT 8.7.c When σ is known, ˆµ = X is an unbiased estimator for µ. If you can show that its variance attains the Cramer-Rao lower bound, then no other unbiased

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015 Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

BTRY 4090: Spring 2009 Theory of Statistics

BTRY 4090: Spring 2009 Theory of Statistics BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)

More information

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015 STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis

More information

Bias Variance Trade-off

Bias Variance Trade-off Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Fall, 2013 Page 1 Random Variable and Probability Distribution Discrete random variable Y : Finite possible values {y

More information

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1) Summary of Chapter 7 (Sections 7.2-7.5) and Chapter 8 (Section 8.1) Chapter 7. Tests of Statistical Hypotheses 7.2. Tests about One Mean (1) Test about One Mean Case 1: σ is known. Assume that X N(µ, σ

More information

Chapter 8.8.1: A factorization theorem

Chapter 8.8.1: A factorization theorem LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.

More information

Hypothesis testing: theory and methods

Hypothesis testing: theory and methods Statistical Methods Warsaw School of Economics November 3, 2017 Statistical hypothesis is the name of any conjecture about unknown parameters of a population distribution. The hypothesis should be verifiable

More information

EXAMINERS REPORT & SOLUTIONS STATISTICS 1 (MATH 11400) May-June 2009

EXAMINERS REPORT & SOLUTIONS STATISTICS 1 (MATH 11400) May-June 2009 EAMINERS REPORT & SOLUTIONS STATISTICS (MATH 400) May-June 2009 Examiners Report A. Most plots were well done. Some candidates muddled hinges and quartiles and gave the wrong one. Generally candidates

More information

STAT 730 Chapter 4: Estimation

STAT 730 Chapter 4: Estimation STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum

More information

Problem Selected Scores

Problem Selected Scores Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected

More information

Lecture 15. Hypothesis testing in the linear model

Lecture 15. Hypothesis testing in the linear model 14. Lecture 15. Hypothesis testing in the linear model Lecture 15. Hypothesis testing in the linear model 1 (1 1) Preliminary lemma 15. Hypothesis testing in the linear model 15.1. Preliminary lemma Lemma

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

HT Introduction. P(X i = x i ) = e λ λ x i

HT Introduction. P(X i = x i ) = e λ λ x i MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Statistical Inference

Statistical Inference Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)

More information

TUTORIAL 8 SOLUTIONS #

TUTORIAL 8 SOLUTIONS # TUTORIAL 8 SOLUTIONS #9.11.21 Suppose that a single observation X is taken from a uniform density on [0,θ], and consider testing H 0 : θ = 1 versus H 1 : θ =2. (a) Find a test that has significance level

More information

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically

More information

MATH4427 Notebook 2 Fall Semester 2017/2018

MATH4427 Notebook 2 Fall Semester 2017/2018 MATH4427 Notebook 2 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................

More information

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

STAT 135 Lab 2 Confidence Intervals, MLE and the Delta Method

STAT 135 Lab 2 Confidence Intervals, MLE and the Delta Method STAT 135 Lab 2 Confidence Intervals, MLE and the Delta Method Rebecca Barter February 2, 2015 Confidence Intervals Confidence intervals What is a confidence interval? A confidence interval is calculated

More information

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis

More information

Chapters 9. Properties of Point Estimators

Chapters 9. Properties of Point Estimators Chapters 9. Properties of Point Estimators Recap Target parameter, or population parameter θ. Population distribution f(x; θ). { probability function, discrete case f(x; θ) = density, continuous case The

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

Confidence Intervals, Testing and ANOVA Summary

Confidence Intervals, Testing and ANOVA Summary Confidence Intervals, Testing and ANOVA Summary 1 One Sample Tests 1.1 One Sample z test: Mean (σ known) Let X 1,, X n a r.s. from N(µ, σ) or n > 30. Let The test statistic is H 0 : µ = µ 0. z = x µ 0

More information

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire

More information

Chapter 7. Hypothesis Testing

Chapter 7. Hypothesis Testing Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

Stat 5102 Final Exam May 14, 2015

Stat 5102 Final Exam May 14, 2015 Stat 5102 Final Exam May 14, 2015 Name Student ID The exam is closed book and closed notes. You may use three 8 1 11 2 sheets of paper with formulas, etc. You may also use the handouts on brand name distributions

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

ML Testing (Likelihood Ratio Testing) for non-gaussian models

ML Testing (Likelihood Ratio Testing) for non-gaussian models ML Testing (Likelihood Ratio Testing) for non-gaussian models Surya Tokdar ML test in a slightly different form Model X f (x θ), θ Θ. Hypothesist H 0 : θ Θ 0 Good set: B c (x) = {θ : l x (θ) max θ Θ l

More information

Asymptotic Statistics-VI. Changliang Zou

Asymptotic Statistics-VI. Changliang Zou Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.

More information

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently

More information

2017 Financial Mathematics Orientation - Statistics

2017 Financial Mathematics Orientation - Statistics 2017 Financial Mathematics Orientation - Statistics Written by Long Wang Edited by Joshua Agterberg August 21, 2018 Contents 1 Preliminaries 5 1.1 Samples and Population............................. 5

More information

Statistics and Econometrics I

Statistics and Econometrics I Statistics and Econometrics I Point Estimation Shiu-Sheng Chen Department of Economics National Taiwan University September 13, 2016 Shiu-Sheng Chen (NTU Econ) Statistics and Econometrics I September 13,

More information

Mathematical statistics

Mathematical statistics October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation

More information

STAT 135 Lab 7 Distributions derived from the normal distribution, and comparing independent samples.

STAT 135 Lab 7 Distributions derived from the normal distribution, and comparing independent samples. STAT 135 Lab 7 Distributions derived from the normal distribution, and comparing independent samples. Rebecca Barter March 16, 2015 The χ 2 distribution The χ 2 distribution We have seen several instances

More information

Space Telescope Science Institute statistics mini-course. October Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses

Space Telescope Science Institute statistics mini-course. October Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses Space Telescope Science Institute statistics mini-course October 2011 Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses James L Rosenberger Acknowledgements: Donald Richards, William

More information

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30 MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)

More information

Final Examination Statistics 200C. T. Ferguson June 11, 2009

Final Examination Statistics 200C. T. Ferguson June 11, 2009 Final Examination Statistics 00C T. Ferguson June, 009. (a) Define: X n converges in probability to X. (b) Define: X m converges in quadratic mean to X. (c) Show that if X n converges in quadratic mean

More information

1 One-way analysis of variance

1 One-way analysis of variance LIST OF FORMULAS (Version from 21. November 2014) STK2120 1 One-way analysis of variance Assume X ij = µ+α i +ɛ ij ; j = 1, 2,..., J i ; i = 1, 2,..., I ; where ɛ ij -s are independent and N(0, σ 2 ) distributed.

More information

Regression Estimation Least Squares and Maximum Likelihood

Regression Estimation Least Squares and Maximum Likelihood Regression Estimation Least Squares and Maximum Likelihood Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 3, Slide 1 Least Squares Max(min)imization Function to minimize

More information

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or

More information

Sociology 6Z03 Review II

Sociology 6Z03 Review II Sociology 6Z03 Review II John Fox McMaster University Fall 2016 John Fox (McMaster University) Sociology 6Z03 Review II Fall 2016 1 / 35 Outline: Review II Probability Part I Sampling Distributions Probability

More information

I i=1 1 I(J 1) j=1 (Y ij Ȳi ) 2. j=1 (Y j Ȳ )2 ] = 2n( is the two-sample t-test statistic.

I i=1 1 I(J 1) j=1 (Y ij Ȳi ) 2. j=1 (Y j Ȳ )2 ] = 2n( is the two-sample t-test statistic. Serik Sagitov, Chalmers and GU, February, 08 Solutions chapter Matlab commands: x = data matrix boxplot(x) anova(x) anova(x) Problem.3 Consider one-way ANOVA test statistic For I = and = n, put F = MS

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Ch. 5 Hypothesis Testing

Ch. 5 Hypothesis Testing Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,

More information

Chapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments

Chapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments Chapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments We consider two kinds of random variables: discrete and continuous random variables. For discrete random

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

Elements of statistics (MATH0487-1)

Elements of statistics (MATH0487-1) Elements of statistics (MATH0487-1) Prof. Dr. Dr. K. Van Steen University of Liège, Belgium November 12, 2012 Introduction to Statistics Basic Probability Revisited Sampling Exploratory Data Analysis -

More information

Statistics 135 Fall 2008 Final Exam

Statistics 135 Fall 2008 Final Exam Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations

More information

Statistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm

Statistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm Statistics GIDP Ph.D. Qualifying Exam Theory Jan, 06, 9:00am-:00pm Instructions: Provide answers on the supplied pads of paper; write on only one side of each sheet. Complete exactly 5 of the 6 problems.

More information

Lecture 32: Asymptotic confidence sets and likelihoods

Lecture 32: Asymptotic confidence sets and likelihoods Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence

More information

Analysis of Variance

Analysis of Variance Statistical Techniques II EXST7015 Analysis of Variance 15a_ANOVA_Introduction 1 Design The simplest model for Analysis of Variance (ANOVA) is the CRD, the Completely Randomized Design This model is also

More information

Session 3 The proportional odds model and the Mann-Whitney test

Session 3 The proportional odds model and the Mann-Whitney test Session 3 The proportional odds model and the Mann-Whitney test 3.1 A unified approach to inference 3.2 Analysis via dichotomisation 3.3 Proportional odds 3.4 Relationship with the Mann-Whitney test Session

More information

Section 4.6 Simple Linear Regression

Section 4.6 Simple Linear Regression Section 4.6 Simple Linear Regression Objectives ˆ Basic philosophy of SLR and the regression assumptions ˆ Point & interval estimation of the model parameters, and how to make predictions ˆ Point and interval

More information

Introduction to Simple Linear Regression

Introduction to Simple Linear Regression Introduction to Simple Linear Regression Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Introduction to Simple Linear Regression 1 / 68 About me Faculty in the Department

More information

Application of Variance Homogeneity Tests Under Violation of Normality Assumption

Application of Variance Homogeneity Tests Under Violation of Normality Assumption Application of Variance Homogeneity Tests Under Violation of Normality Assumption Alisa A. Gorbunova, Boris Yu. Lemeshko Novosibirsk State Technical University Novosibirsk, Russia e-mail: gorbunova.alisa@gmail.com

More information

Statistics 3858 : Contingency Tables

Statistics 3858 : Contingency Tables Statistics 3858 : Contingency Tables 1 Introduction Before proceeding with this topic the student should review generalized likelihood ratios ΛX) for multinomial distributions, its relation to Pearson

More information

This paper is not to be removed from the Examination Halls

This paper is not to be removed from the Examination Halls ~~ST104B ZA d0 This paper is not to be removed from the Examination Halls UNIVERSITY OF LONDON ST104B ZB BSc degrees and Diplomas for Graduates in Economics, Management, Finance and the Social Sciences,

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida First Year Examination Department of Statistics, University of Florida August 19, 010, 8:00 am - 1:00 noon Instructions: 1. You have four hours to answer questions in this examination.. You must show your

More information

Probability and Statistics Notes

Probability and Statistics Notes Probability and Statistics Notes Chapter Seven Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Seven Notes Spring 2011 1 / 42 Outline

More information

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics Exploring Data: Distributions Look for overall pattern (shape, center, spread) and deviations (outliers). Mean (use a calculator): x = x 1 + x

More information

Some General Types of Tests

Some General Types of Tests Some General Types of Tests We may not be able to find a UMP or UMPU test in a given situation. In that case, we may use test of some general class of tests that often have good asymptotic properties.

More information

CHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007)

CHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007) FROM: PAGANO, R. R. (007) I. INTRODUCTION: DISTINCTION BETWEEN PARAMETRIC AND NON-PARAMETRIC TESTS Statistical inference tests are often classified as to whether they are parametric or nonparametric Parameter

More information

Probability Theory and Statistics. Peter Jochumzen

Probability Theory and Statistics. Peter Jochumzen Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

(a) (3 points) Construct a 95% confidence interval for β 2 in Equation 1.

(a) (3 points) Construct a 95% confidence interval for β 2 in Equation 1. Problem 1 (21 points) An economist runs the regression y i = β 0 + x 1i β 1 + x 2i β 2 + x 3i β 3 + ε i (1) The results are summarized in the following table: Equation 1. Variable Coefficient Std. Error

More information

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:

More information

Problem 1 (20) Log-normal. f(x) Cauchy

Problem 1 (20) Log-normal. f(x) Cauchy ORF 245. Rigollet Date: 11/21/2008 Problem 1 (20) f(x) f(x) 0.0 0.1 0.2 0.3 0.4 0.0 0.2 0.4 0.6 0.8 4 2 0 2 4 Normal (with mean -1) 4 2 0 2 4 Negative-exponential x x f(x) f(x) 0.0 0.1 0.2 0.3 0.4 0.5

More information

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics The candidates for the research course in Statistics will have to take two shortanswer type tests

More information

STAT 135 Lab 3 Asymptotic MLE and the Method of Moments

STAT 135 Lab 3 Asymptotic MLE and the Method of Moments STAT 135 Lab 3 Asymptotic MLE and the Method of Moments Rebecca Barter February 9, 2015 Maximum likelihood estimation (a reminder) Maximum likelihood estimation Suppose that we have a sample, X 1, X 2,...,

More information

Master s Written Examination - Solution

Master s Written Examination - Solution Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2

More information

8. Hypothesis Testing

8. Hypothesis Testing FE661 - Statistical Methods for Financial Engineering 8. Hypothesis Testing Jitkomut Songsiri introduction Wald test likelihood-based tests significance test for linear regression 8-1 Introduction elements

More information

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1, Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem

More information

1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ).

1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Estimation February 3, 206 Debdeep Pati General problem Model: {P θ : θ Θ}. Observe X P θ, θ Θ unknown. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Examples: θ = (µ,

More information

F & B Approaches to a simple model

F & B Approaches to a simple model A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys

More information

Topic 19 Extensions on the Likelihood Ratio

Topic 19 Extensions on the Likelihood Ratio Topic 19 Extensions on the Likelihood Ratio Two-Sided Tests 1 / 12 Outline Overview Normal Observations Power Analysis 2 / 12 Overview The likelihood ratio test is a popular choice for composite hypothesis

More information

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

Performance Evaluation and Comparison

Performance Evaluation and Comparison Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Cross Validation and Resampling 3 Interval Estimation

More information