Math 181B Homework 1 Solution
|
|
- Camilla Howard
- 5 years ago
- Views:
Transcription
1 Math 181B Homework 1 Solution 1. Write down the likelihood: L(λ = n λ X i e λ X i! (a One-sided test: H 0 : λ = 1 vs H 1 : λ = 0.1 The likelihood ratio: where LR = L(1 L(0.1 = 1 X i e n 1 = λ n X i e nλ n X i! = e n X i e n X i = g( X i g(t = en 0.1 t = en 10 t, g (t = e n 10 t log 10 > 0, when t 0 So g(t is strictly increasing and invertible g 1 (g(t = t. Therefore, we can simplify the rejecting region for likelihood-ratio test LR = g( X i k as n X i C = g 1 (k. Remark: You may state without proof: LR is a monotone increasing function of X i. g(t argument serves to convince you. Often, you first come up with the conclusion. Then, you figure out the way to prove it (not necessary. Regular solution: Apply Central Limit Theorem under H 0, Z = n 1/ n (X i 1 H 0 N(0, 1. Obviously, X i = h(z = nz. h(z is a strictly increasing function. The rejecting region is then equivalent to Z C = h 1 (C = h 1 (g 1 (k. To achieve level α, we want P H0 (Z C = α. This is the definition of C being the α quantile of N(0, 1, usually written as Z 1 α = Z α. Extra credit: X H 0 i P oisson(n. To achieve level α, we want PH0 ( X i C = α. This is the definition of C being the α quantile of P oisson(n. (b Two-sided test: H 0 : λ = 1 vs H 1 : λ 1 MLE: ˆλ = X. The likelihood ratio: LR = L(1 L( X = 1 X i e n 1 X X i e n X = en( X 1 X n X X = eg( Regular solution: Apply the asymptotic theory for Generalized Likelihood- Ratio test, log LR = n X log X n( X 1 H 0 χ 1 The rejecting region is log LR C. To achieve level α, we want P H0 ( log LR C = α. This is the definition of C being the 1 α quantile of χ 1, usually written as χ 1,α. 1
2 Extra credit: { > 0 t < 1 g (t = n log t < 0 t > 1 LR is first increasing then decreasing w.r.t. X. Therefore, the rejecting region {LR k} is equivalent to { X a} { X b} for some pair a < 1 < b satisfying g(a = g(b. The distribution of X under H1 is studied in part (a.. Write down the likelihood: MLE: ˆp = X/n. The likelihood ratio: LR = L(0.5 L(X/n = L(p = p X (1 p n X 0.5 n (X/n X (1 X/n n X = en log 0.5 X log X (n X log(n X n log n = e g(x g (x = log x + log(n x { > 0 x < n/ < 0 x > n/ LR is first increasing then decreasing w.r.t. X. Therefore, the rejecting region {LR k} is equivalent to {X a} {X b} for some pair a < n/ < b satisfying g(a = g(b. Extra credit: Notice g(n x = g(x, so we may set b = n a. The rejecting region is then {X a} {X n a} = { X n C = n a}. Finally, figure out a from the distribution of X under H 0. Regular solution: Normal approximation Z = n 1/ (X n/ H 0 N(0, 1. Rejecting region is equivalent to Z > C. To achieve level α, we want P H0 ( Z > C = α. N(0, 1 is symmetric. Therefore, P H0 (Z > C = α/. This is the definition of C being the 1 α/ quantile of N(0, 1, usually written as Z α/. Extra Credit: X H 0 Binom(n, 0.5. To achieve level α, we want P H0 ( X n n a = α. Binom(n, 0.5 is also symmetric. Therefore, P H0 (X a = α/. This is the definition of a being the α/ quantile of Binom(n, Write down the likelihood: Log-likelihood: L(µ, σ = n 1 σ 1 π e (X i µ = (σ π n/ e 1 (X i µ l(µ, σ = n log(σ n log(π 1 (X i µ (a Regular solution: First simplify the null as H 0 : σ = σ 0. The full parameter space is now σ σ 0.
3 MLE: First take partial derivative w.r.t. the unconstrained parameter µ: l µ = n σ ( X µ The solution of the equation is always ˆµ = X for any σ > 0. Therefore, ˆµ = X is the MLE in both Θ 0 and Θ 0 Θ 1. The MLE for σ is then the maximum of the partial likelihood: l (σ = l( X, σ = n log(σ n log(π 1 (X i X Under H 0, σ =. Under H 0 + H 1, compute the derivative of l : { } dσ = n σ + 1 (X σ 4 i X = n σ n 1 (X σ 4 i X Denote the unconstrained MLE ˆσ = n 1 (X i X ˆσ depends on the observed data. It is random whether ˆσ belongs to the full parameter space [, +. We have to discuss the MLE under two cases. Case 1: If ˆσ, then the MLE is the unconstrained MLE ˆσ. The likelihood ratio is LR = L( X, σ 0 L( X, ˆσ = (σ 0 ˆσ n/ exp where ( 1 0 = ( σ 0 ˆσ n/ exp ( nˆσ + n = e (X i X + 1 ˆσ ˆσ g( σ 0 g(t = n log t n t + n, t 1 g (t = n t n 0, t 1 (X i X, ˆσ σ 0 ˆσ σ 0 The rejecting region {LR k} in case 1 is thus equivalent to {ˆσ } { ˆσ C}. Case : ˆσ <, by the derivative of partial log-likelihood dσ < 0, σ > σ 0 > ˆσ 1 3
4 l (σ is strictly decreasing over the parameter space [σ 0, +. The maximum is then the left boundary σ 0. The likelihood ratio is LR = L( X, σ 0 L( X, σ 0 = 1 If {LR = 1} belongs to the rejecting region defined by {LR k}, we must have k 1. However, LR is always no greater than 1, so the type I error is P H0 (LR k P H0 (LR 1 = 1 > α. Therefore, H 0 is always accepted in case to achieve any level α < 1. As a summary, the rejecting region is { nˆσ max(n, C }. Under H 0, nˆσ σ 0 = 1 0 To achieve level α test, we want (X i X χ n 1 P H0 ( nˆσ σ 0 max(n, C = α This is the definition of max(n, C being the 1 α/ quantile of χ n 1, usually written as χ n 1,α. Note the mean of χ n 1 is n 1. thus, its 1 α quantile is greater than n when α is small. (b Extra credit solution: Do not do the simplification. MLE: First take partial derivative w.r.t. the unconstrained parameter µ: l µ = n σ ( X µ The solution of the equation is always ˆµ = X for any σ > 0. Therefore, ˆµ = X is the MLE in both Θ 0 and Θ 0 Θ 1. The MLE for σ is then the maximum of the partial likelihood: l (σ = l( X, σ = n log(σ n log(π 1 (X i X Under H 0 + H 1, ˆσ = n 1 n (X i X is the unconstrained MLE. Under H 0, compute the derivative of l : dσ = n + 1 σ 4 (X i X = n σ 4 { σ n 1 } (X i X ˆσ depends on the observed data. It is random whether ˆσ belongs to the null parameter space (0, σ 0]. We have to discuss the MLE under two cases. 4
5 Case 1: If ˆσ σ 0, by the derivative of partial log-likelihood dσ > 0, σ σ 0 < ˆσ l (σ is strictly increasing over the parameter space (0, ]. The maximum is then the right boundary. The likelihood ratio is LR = L( X, ( L( X, ˆσ = (σ 0 ˆσ n/ exp 1 (X i X + 1 (X ˆσ i X where = ( σ 0 ˆσ n/ exp ( nˆσ + n = e ˆσ g( σ 0 g(t = n log t n t + n, t 1 g (t = n t n 0, t 1, ˆσ σ 0 ˆσ σ 0 The rejecting region {LR k} in case 1 is thus equivalent to {ˆσ } { ˆσ C}. Case : ˆσ <, then the MLE is the unconstrained MLE ˆσ. The likelihood ratio is LR = L( X, L( X, = 1 If {LR = 1} belongs to the rejecting region defined by {LR k}, we must have k 1. However, LR is always no greater than 1, so the type I error is P H0 (LR k P H0 (LR 1 = 1 > α. Therefore, H 0 is always accepted in case to achieve any level α < 1. As a summary, the rejecting region is { nˆσ max(n, C }. Under H 0, nˆσ σ = 1 To achieve level α test, we want ( nˆσ max P σ max(n, C σ σ 0 (X i X χ n 1 1 ( nˆσ = max P σ σ 0 σ σ σ max(n, C = α For any σ, the statistic nˆσ has the same distribution χ σ n 1. The smallest the cutoff σ 0 max(n, C is achieved at σ = σ σ 0. That corresponds to the biggest type I error. max P σ σ ( nˆσ σ 0 ( nˆσ max(n, C = P 5 max(n, C = α
6 This is the definition of max(n, C being the 1 α/ quantile of χ n 1, usually written as χ n 1,α. Note the mean of χ n 1 is n 1. thus, its 1 α quantile is greater than n when α is small. 4. Extra credit: Do the transformation X i Y i = Z i N(µ X µ Y, (1 ρσ. The affine (linear transformation of multivariate normal distribution is normal. Its parameters are given by its mean and variance. EZ = EX EY = µ X µ Y, VarZ = VarX + VarY Cov(X, Y = (1 ρσ Write ν = µ X µ Y, τ = (1 ρσ. Assume ρ ( 1, 1. Then, τ (0, 4σ. Write down the likelihood: n L(ν, τ 1 = τ 1 π e τ (Z i ν = (τ π n/ e 1 τ (Z i ν Log-likelihood: (a MLE: Under H 0, ν = 0. l(ν, τ = n log(τ n log(π 1 τ d dτ l(0, τ = n τ + 1 τ 4 Z i (Z i ν { = n τ n 1 τ 4 We have to discuss the MLE for τ under two cases. constraint τ < 4σ Denote the unconstrained MLE ˆτ 0 = n 1 Z i Z i } Remember there is this ˆτ 0 depends on the observed data. It is random whether ˆτ 0 belongs to the parameter space (0, 4σ. Following the same case-wise argument in previous problem, you will find the MLE under H 0 is min(4σ, ˆτ 0. Under H 0 + H 1, first solve ˆν = Z. Denote the unconstrained MLE ˆτ = n 1 (Z i Z Similarly, the constrained MLE for τ is min(4σ, ˆτ. Therefore, the likelihood-ratio statistic is Λ = L(0, min(4σ, ˆτ 0 L( Z, min(4σ, ˆτ 6
7 (b You may use log LR χ 1. Or follow the large sample argument. Under H 0, both unconstrained MLE ˆτ 0 and ˆτ converge in probability to true τ (0, 4σ by Law of Large Numbers. When sample size is large, we only need to consider the case ˆτ < ˆτ 0 < 4σ. Λ = L(0, ˆτ ( 0 ˆτ n/ L( Z, ˆτ = 0 ˆτ It is a monotone decreasing function of ˆτ 0 /ˆτ. Under H 0, F = ( ˆτ 0 ˆτ 1 (n 1 = To achieve the level α, the rejecting region is n Z (Z i Z /(n 1 F 1,n n Z (Z i Z /(n 1 F 1 α,1,n where F 1 α,1,n is the 1 α quantile of F 1,n. (c It is equivalent to the paired t-test Z T = s /n t α/,n 1 since T = F and t α/,n 1 = F 1,n 1. Read Larsen and Marx section Extra credit: Read Larsen and Marx Appendix 9.A.1 page
Exercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationNotes on the Multivariate Normal and Related Topics
Version: July 10, 2013 Notes on the Multivariate Normal and Related Topics Let me refresh your memory about the distinctions between population and sample; parameters and statistics; population distributions
More informationSTT 843 Key to Homework 1 Spring 2018
STT 843 Key to Homework Spring 208 Due date: Feb 4, 208 42 (a Because σ = 2, σ 22 = and ρ 2 = 05, we have σ 2 = ρ 2 σ σ22 = 2/2 Then, the mean and covariance of the bivariate normal is µ = ( 0 2 and Σ
More informationAnswer Key for STAT 200B HW No. 7
Answer Key for STAT 200B HW No. 7 May 5, 2007 Problem 2.2 p. 649 Assuming binomial 2-sample model ˆπ =.75, ˆπ 2 =.6. a ˆτ = ˆπ 2 ˆπ =.5. From Ex. 2.5a on page 644: ˆπ ˆπ + ˆπ 2 ˆπ 2.75.25.6.4 = + =.087;
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationStat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS
Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS 1a. Under the null hypothesis X has the binomial (100,.5) distribution with E(X) = 50 and SE(X) = 5. So P ( X 50 > 10) is (approximately) two tails
More informationAnswer Key for STAT 200B HW No. 8
Answer Key for STAT 200B HW No. 8 May 8, 2007 Problem 3.42 p. 708 The values of Ȳ for x 00, 0, 20, 30 are 5/40, 0, 20/50, and, respectively. From Corollary 3.5 it follows that MLE exists i G is identiable
More informationExam 2. Jeremy Morris. March 23, 2006
Exam Jeremy Morris March 3, 006 4. Consider a bivariate normal population with µ 0, µ, σ, σ and ρ.5. a Write out the bivariate normal density. The multivariate normal density is defined by the following
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationHypothesis Testing One Sample Tests
STATISTICS Lecture no. 13 Department of Econometrics FEM UO Brno office 69a, tel. 973 442029 email:jiri.neubauer@unob.cz 12. 1. 2010 Tests on Mean of a Normal distribution Tests on Variance of a Normal
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More information2.6.3 Generalized likelihood ratio tests
26 HYPOTHESIS TESTING 113 263 Generalized likelihood ratio tests When a UMP test does not exist, we usually use a generalized likelihood ratio test to verify H 0 : θ Θ against H 1 : θ Θ\Θ It can be used
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More informationTheory of Statistics.
Theory of Statistics. Homework V February 5, 00. MT 8.7.c When σ is known, ˆµ = X is an unbiased estimator for µ. If you can show that its variance attains the Cramer-Rao lower bound, then no other unbiased
More informationLecture 32: Asymptotic confidence sets and likelihoods
Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence
More informationChapter 4: Asymptotic Properties of the MLE (Part 2)
Chapter 4: Asymptotic Properties of the MLE (Part 2) Daniel O. Scharfstein 09/24/13 1 / 1 Example Let {(R i, X i ) : i = 1,..., n} be an i.i.d. sample of n random vectors (R, X ). Here R is a response
More informationMath 152. Rumbos Fall Solutions to Assignment #12
Math 52. umbos Fall 2009 Solutions to Assignment #2. Suppose that you observe n iid Bernoulli(p) random variables, denoted by X, X 2,..., X n. Find the LT rejection region for the test of H o : p p o versus
More informationHypothesis Testing. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA
Hypothesis Testing Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA An Example Mardia et al. (979, p. ) reprint data from Frets (9) giving the length and breadth (in
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationLikelihood Ratio tests
Likelihood Ratio tests For general composite hypotheses optimality theory is not usually successful in producing an optimal test. instead we look for heuristics to guide our choices. The simplest approach
More informationStatistics. Statistics
The main aims of statistics 1 1 Choosing a model 2 Estimating its parameter(s) 1 point estimates 2 interval estimates 3 Testing hypotheses Distributions used in statistics: χ 2 n-distribution 2 Let X 1,
More informationChapter 3: Maximum Likelihood Theory
Chapter 3: Maximum Likelihood Theory Florian Pelgrin HEC September-December, 2010 Florian Pelgrin (HEC) Maximum Likelihood Theory September-December, 2010 1 / 40 1 Introduction Example 2 Maximum likelihood
More information1 Empirical Likelihood
Empirical Likelihood February 3, 2016 Debdeep Pati 1 Empirical Likelihood Empirical likelihood a nonparametric method without having to assume the form of the underlying distribution. It retains some of
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationStatement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.
MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss
More informationStatistical Inference
Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationIntroduction to Normal Distribution
Introduction to Normal Distribution Nathaniel E. Helwig Assistant Professor of Psychology and Statistics University of Minnesota (Twin Cities) Updated 17-Jan-2017 Nathaniel E. Helwig (U of Minnesota) Introduction
More informationMATH5745 Multivariate Methods Lecture 07
MATH5745 Multivariate Methods Lecture 07 Tests of hypothesis on covariance matrix March 16, 2018 MATH5745 Multivariate Methods Lecture 07 March 16, 2018 1 / 39 Test on covariance matrices: Introduction
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More information2.3 Methods of Estimation
96 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE 2.3 Methods of Estimation 2.3. Method of Moments The Method of Moments is a simple technique based on the idea that the sample moments are natural estimators
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationQuick Review on Linear Multiple Regression
Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,
More informationMAS223 Statistical Modelling and Inference Examples
Chapter MAS3 Statistical Modelling and Inference Examples Example : Sample spaces and random variables. Let S be the sample space for the experiment of tossing two coins; i.e. Define the random variables
More informationFinal Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.
1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically
More informationThe Multivariate Normal Distribution. In this case according to our theorem
The Multivariate Normal Distribution Defn: Z R 1 N(0, 1) iff f Z (z) = 1 2π e z2 /2. Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) T with the Z i independent and each Z i N(0, 1). In this
More informationLikelihoods. P (Y = y) = f(y). For example, suppose Y has a geometric distribution on 1, 2,... with parameter p. Then the pmf is
Likelihoods The distribution of a random variable Y with a discrete sample space (e.g. a finite sample space or the integers) can be characterized by its probability mass function (pmf): P (Y = y) = f(y).
More informationEstimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1
Estimation September 24, 2018 STAT 151 Class 6 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0 1 1 1
More informationENGG2430A-Homework 2
ENGG3A-Homework Due on Feb 9th,. Independence vs correlation a For each of the following cases, compute the marginal pmfs from the joint pmfs. Explain whether the random variables X and Y are independent,
More informationHomework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.
Stat 300A Theory of Statistics Homework 7: Solutions Nikos Ignatiadis Due on November 28, 208 Solutions should be complete and concisely written. Please, use a separate sheet or set of sheets for each
More informationCHANGE DETECTION IN TIME SERIES
CHANGE DETECTION IN TIME SERIES Edit Gombay TIES - 2008 University of British Columbia, Kelowna June 8-13, 2008 Outline Introduction Results Examples References Introduction sunspot.year 0 50 100 150 1700
More informationf(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain
0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher
More informationFinal Examination Statistics 200C. T. Ferguson June 11, 2009
Final Examination Statistics 00C T. Ferguson June, 009. (a) Define: X n converges in probability to X. (b) Define: X m converges in quadratic mean to X. (c) Show that if X n converges in quadratic mean
More informationA Study on the Correlation of Bivariate And Trivariate Normal Models
Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 11-1-2013 A Study on the Correlation of Bivariate And Trivariate Normal Models Maria
More informationChapter 11. Hypothesis Testing (II)
Chapter 11. Hypothesis Testing (II) 11.1 Likelihood Ratio Tests one of the most popular ways of constructing tests when both null and alternative hypotheses are composite (i.e. not a single point). Let
More informationProbability Theory and Statistics. Peter Jochumzen
Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................
More informationAsymptotic Statistics-VI. Changliang Zou
Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous
More informationSOLUTION FOR HOMEWORK 6, STAT 6331
SOLUTION FOR HOMEWORK 6, STAT 633. Exerc.7.. It is given that X,...,X n is a sample from N(θ, σ ), and the Bayesian approach is used with Θ N(µ, τ ). The parameters σ, µ and τ are given. (a) Find the joinf
More information1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ).
Estimation February 3, 206 Debdeep Pati General problem Model: {P θ : θ Θ}. Observe X P θ, θ Θ unknown. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Examples: θ = (µ,
More informationLoglikelihood and Confidence Intervals
Stat 504, Lecture 2 1 Loglikelihood and Confidence Intervals The loglikelihood function is defined to be the natural logarithm of the likelihood function, l(θ ; x) = log L(θ ; x). For a variety of reasons,
More informationInference for the Pareto, half normal and related. distributions
Inference for the Pareto, half normal and related distributions Hassan Abuhassan and David J. Olive Department of Mathematics, Southern Illinois University, Mailcode 4408, Carbondale, IL 62901-4408, USA
More informationLecture 25: Review. Statistics 104. April 23, Colin Rundel
Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April
More informationHypothesis Testing. A rule for making the required choice can be described in two ways: called the rejection or critical region of the test.
Hypothesis Testing Hypothesis testing is a statistical problem where you must choose, on the basis of data X, between two alternatives. We formalize this as the problem of choosing between two hypotheses:
More informationCHAPTER 8. Test Procedures is a rule, based on sample data, for deciding whether to reject H 0 and contains:
CHAPTER 8 Test of Hypotheses Based on a Single Sample Hypothesis testing is the method that decide which of two contradictory claims about the parameter is correct. Here the parameters of interest are
More informationNotes on Random Vectors and Multivariate Normal
MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution
More informationMLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22
MLE and GMM Li Zhao, SJTU Spring, 2017 Li Zhao MLE and GMM 1 / 22 Outline 1 MLE 2 GMM 3 Binary Choice Models Li Zhao MLE and GMM 2 / 22 Maximum Likelihood Estimation - Introduction For a linear model y
More informationVariance reduction. Michel Bierlaire. Transport and Mobility Laboratory. Variance reduction p. 1/18
Variance reduction p. 1/18 Variance reduction Michel Bierlaire michel.bierlaire@epfl.ch Transport and Mobility Laboratory Variance reduction p. 2/18 Example Use simulation to compute I = 1 0 e x dx We
More informationComparing two independent samples
In many applications it is necessary to compare two competing methods (for example, to compare treatment effects of a standard drug and an experimental drug). To compare two methods from statistical point
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationAppendix. Proof of Theorem 1. Define. [ ˆΛ 0(D) ˆΛ 0(t) ˆΛ (t) ˆΛ. (0) t. X 0 n(t) = D t. and. 0(t) ˆΛ 0(0) g(t(d t)), 0 < t < D, t.
Appendix Proof of Theorem. Define [ ˆΛ (D) X n (t) = ˆΛ (t) D t ˆΛ (t) ˆΛ () g(t(d t)), t < t < D X n(t) = [ ˆΛ (D) ˆΛ (t) D t ˆΛ (t) ˆΛ () g(t(d t)), < t < D, t where ˆΛ (t) = log[exp( ˆΛ(t)) + ˆp/ˆp,
More informationGeneralized Linear Models Introduction
Generalized Linear Models Introduction Statistics 135 Autumn 2005 Copyright c 2005 by Mark E. Irwin Generalized Linear Models For many problems, standard linear regression approaches don t work. Sometimes,
More informationCanonical Correlation Analysis of Longitudinal Data
Biometrics Section JSM 2008 Canonical Correlation Analysis of Longitudinal Data Jayesh Srivastava Dayanand N Naik Abstract Studying the relationship between two sets of variables is an important multivariate
More informationCherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants
18.650 Statistics for Applications Chapter 5: Parametric hypothesis testing 1/37 Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009
More information2014/2015 Smester II ST5224 Final Exam Solution
014/015 Smester II ST54 Final Exam Solution 1 Suppose that (X 1,, X n ) is a random sample from a distribution with probability density function f(x; θ) = e (x θ) I [θ, ) (x) (i) Show that the family of
More informationi=1 X i/n i=1 (X i X) 2 /(n 1). Find the constant c so that the statistic c(x X n+1 )/S has a t-distribution. If n = 8, determine k such that
Math 47 Homework Assignment 4 Problem 411 Let X 1, X,, X n, X n+1 be a random sample of size n + 1, n > 1, from a distribution that is N(µ, σ ) Let X = n i=1 X i/n and S = n i=1 (X i X) /(n 1) Find the
More informationYou may not use your books/notes on this exam. You may use calculator.
MATH 450 Fall 2018 Review problems 12/03/18 Time Limit: 60 Minutes Name (Print: This exam contains 6 pages (including this cover page and 5 problems. Check to see if any pages are missing. Enter all requested
More informationIEOR 165 Lecture 13 Maximum Likelihood Estimation
IEOR 165 Lecture 13 Maximum Likelihood Estimation 1 Motivating Problem Suppose we are working for a grocery store, and we have decided to model service time of an individual using the express lane (for
More informationChapter 2. Discrete Distributions
Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation
More informationRandom vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.
Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just
More informationSTA 2201/442 Assignment 2
STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution
More informationFinite Singular Multivariate Gaussian Mixture
21/06/2016 Plan 1 Basic definitions Singular Multivariate Normal Distribution 2 3 Plan Singular Multivariate Normal Distribution 1 Basic definitions Singular Multivariate Normal Distribution 2 3 Multivariate
More informationSpring 2012 Math 541A Exam 1. X i, S 2 = 1 n. n 1. X i I(X i < c), T n =
Spring 2012 Math 541A Exam 1 1. (a) Let Z i be independent N(0, 1), i = 1, 2,, n. Are Z = 1 n n Z i and S 2 Z = 1 n 1 n (Z i Z) 2 independent? Prove your claim. (b) Let X 1, X 2,, X n be independent identically
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More information2017 Financial Mathematics Orientation - Statistics
2017 Financial Mathematics Orientation - Statistics Written by Long Wang Edited by Joshua Agterberg August 21, 2018 Contents 1 Preliminaries 5 1.1 Samples and Population............................. 5
More informationIEOR 165: Spring 2019 Problem Set 2
IEOR 65: Spring 209 Problem Set 2 Instructor: Professor Anil Aswani Issued: 2/8/9 Due: 3//9 Problem : Part a: We may first calculate the sample means: ȳ =.6, x = 7. Then: i= ˆβ = y ix i ȳ x i= x2 i x2
More information11 Survival Analysis and Empirical Likelihood
11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with
More informationGaussian Mixture Models
Gaussian Mixture Models David Rosenberg, Brett Bernstein New York University April 26, 2017 David Rosenberg, Brett Bernstein (New York University) DS-GA 1003 April 26, 2017 1 / 42 Intro Question Intro
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 9 for Applied Multivariate Analysis Outline Addressing ourliers 1 Addressing ourliers 2 Outliers in Multivariate samples (1) For
More informationCovariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance
Covariance Lecture 0: Covariance / Correlation & General Bivariate Normal Sta30 / Mth 30 We have previously discussed Covariance in relation to the variance of the sum of two random variables Review Lecture
More informationProblem Selected Scores
Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected
More informationInstitute of Actuaries of India
Institute of Actuaries of India Subject CT3 Probability & Mathematical Statistics May 2011 Examinations INDICATIVE SOLUTION Introduction The indicative solution has been written by the Examiners with the
More informationThe Delta Method and Applications
Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order
More informationLecture 23 Maximum Likelihood Estimation and Bayesian Inference
Lecture 23 Maximum Likelihood Estimation and Bayesian Inference Thais Paiva STA 111 - Summer 2013 Term II August 7, 2013 1 / 31 Thais Paiva STA 111 - Summer 2013 Term II Lecture 23, 08/07/2013 Lecture
More informationHT Introduction. P(X i = x i ) = e λ λ x i
MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework
More informationSpace Telescope Science Institute statistics mini-course. October Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses
Space Telescope Science Institute statistics mini-course October 2011 Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses James L Rosenberger Acknowledgements: Donald Richards, William
More informationEXAMINERS REPORT & SOLUTIONS STATISTICS 1 (MATH 11400) May-June 2009
EAMINERS REPORT & SOLUTIONS STATISTICS (MATH 400) May-June 2009 Examiners Report A. Most plots were well done. Some candidates muddled hinges and quartiles and gave the wrong one. Generally candidates
More informationLecture 17: Likelihood ratio and asymptotic tests
Lecture 17: Likelihood ratio and asymptotic tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationStat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota
Stat 5102 Lecture Slides Deck 3 Charles J. Geyer School of Statistics University of Minnesota 1 Likelihood Inference We have learned one very general method of estimation: method of moments. the Now we
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More information2018 2019 1 9 sei@mistiu-tokyoacjp http://wwwstattu-tokyoacjp/~sei/lec-jhtml 11 552 3 0 1 2 3 4 5 6 7 13 14 33 4 1 4 4 2 1 1 2 2 1 1 12 13 R?boxplot boxplotstats which does the computation?boxplotstats
More informationSAMPLING BIOS 662. Michael G. Hudgens, Ph.D. mhudgens :55. BIOS Sampling
SAMPLIG BIOS 662 Michael G. Hudgens, Ph.D. mhudgens@bios.unc.edu http://www.bios.unc.edu/ mhudgens 2008-11-14 15:55 BIOS 662 1 Sampling Outline Preliminaries Simple random sampling Population mean Population
More informationECE 275B Homework # 1 Solutions Version Winter 2015
ECE 275B Homework # 1 Solutions Version Winter 2015 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2
More informationCh. 5 Hypothesis Testing
Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationQualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama
Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Instructions This exam has 7 pages in total, numbered 1 to 7. Make sure your exam has all the pages. This exam will be 2 hours
More informationLinear Models and Estimation by Least Squares
Linear Models and Estimation by Least Squares Jin-Lung Lin 1 Introduction Causal relation investigation lies in the heart of economics. Effect (Dependent variable) cause (Independent variable) Example:
More information