Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Size: px
Start display at page:

Download "Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A."

Transcription

1 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n N, let G n be a set in A such that G n+1 G n. Show that P (G n ) monotonically converges to P ( k=1 G k) as n. If you can do this problem, you will be OK in a measure theory class. 2. Show formally that for any CDF defined by F (y) = Pr(Y y), we have lim y F (y) = 0, lim y F (y) = 1, and that F is right continuous. 3. Show how you you can use the CDF F of a random variable Y to compute (a) Pr(Y (a, b]); (b) Pr(Y (a, b)); (c) Pr(Y [a, b]). 4. Let Z N(0, 1). Derive the density of X where (a) X = e Z. (b) X = Z 2 ; 5. Let Y be a random variable with a continuous strictly increasing CDF, so in particular F 1 exists and F 1 (F (y)) = y. (a) Find the CDF of U where U = F (Y ); (b) Find the CDF of X, where X = F 1 (U); (c) Explain how these results can be used to simulate a normal distribution in R using only the runif command and the qnorm command. Write out your computer code. 1

2 6. Let F be a discrete CDF with jumps at Y {0, 1, 2, 3, 4}, with (F (0), F (1), F (2), F (3), F (4)) = (.1,.15,.3,.6, 1.0). Describe how to simulate from this distribution using a random variable U uniform(0, 1). 7. Let Y θ Binomial(n, θ) and let θ beta(a, b). Derive the marginal density of Y and the conditional density of θ given Y. 8. Let Y θ N(θ, σ 2 ) and let θ N(µ, τ 2 ). Derive the marginal density of Y and the conditional density of θ. 9. Normal distribution properties via change of variables: (a) Let X N(θ, τ 2 ). Using the univariate change of variables formula, obtain the distribution of Y = µ + σx. (b) Let W N(θ 1, τ 2 1 ) and X N(θ 2, τ 2 2 ) be independent. Find the distribution of W + X using the multivariate change of variables method. (c) Let Y 1,..., Y n i.i.d. N(µ, σ 2 ). Use the above two results to obtain the distribution of Ȳ. (d) Let Y 1,..., Y n be independent with Y i N(µ i, σ 2 i ). Use the first two results to obtain the distribution of Ȳ. 10. Let Y 1,..., Y n i.i.d. N(µ, σ 2 ). Using the multivariate change of variables formula, show that Ȳ = Y i /n is independent of S 2 = (Yi Ȳ )2 /(n 1). 11. Show the following: (a) E[a + by ] = a + be[y ]. (b) V [a + by ] = b 2 V [Y ]. 12. Let Y be a positive random variable. Use Jensen s inequality to relate 2

3 (a) E[Y p ] 1/p to E[Y q ] 1/q for p > q 1; (b) E[1/Y ] to 1/E[Y ]; (c) log E[Y ] to E[log Y ]. 13. Let {Y t : t N} be i.i.d. random variables on Y = { 1, 0, +1}, with Pr(Y t = 1) = p and Pr(Y t = +1) = p +. These random variables represent jumps of a particle along a one-dimensional grid. Let S T = T t=1 Y t be the position of the particle at time T. Compute the mean and variance of S T as a function of p and p +. Describe qualitatively the behavior of the particle as T increases, as a function of p and p Let (w 1, w 2, w 3 ) Dirichlet(α 1, α 2, α 3 ). (a) Compute the expected value and variance of w j for j {1,..., 3}. (b) Compute the variance of w 1 + w 2 + w 3. (c) Compute the covariance of w 1 and w 2, and explain intuitively the sign of the result. (d) Obtain the distribution of θ = w 1 + w Let X and Y be random variables. Show that E[f(X)g(X, Y ) X] = f(x)e[g(x, Y ) X]. 16. Highly skewed data (y 1,..., y n ) are often analyzed on a log scale, i.e. we analyze (x 1,..., x n ) = (ln y 1,..., ln y n ). (a) Show that e x ȳ. (*) Bonus question: Compare f 1 ( f(y i )/n) for f(x) = 1/x, f(x) = ln x and f(x) = x. 17. Variance of correlated sums: (a) Derive the variance of ax + by for possibly correlated random variables X and Y. 3

4 (b) Let Y = (Y 1,..., Y n ) T be a vector of real-valued random variables. Compute the variance of Y i /n when Var[Y i ] = σ 2 for all i and i. Cor[Y i, Y j ] = 0 for all i.j; ii. Cor[Y i, Y j ] = ρ for all i, j; iii. Cor[Y i, Y j ] = ρ if i j = 1 and is zero if i j > 1. Describe in words how correlation affects the variance of the sample mean. 18. Suppose E[Y ] = µ and Var[Y ] = σ 2. Consider the estimator ˆµ = (1 w)µ 0 + wy. (a) Find the expectation and variance of ˆµ. (b) Find the bias and MSE of ˆµ (as functions of µ). (c) For what values of µ does ˆµ have lower MSE than Y? 19. Let C k (Y 1,..., Y n ) = (Ȳ a kσ/ n, Ȳ + a kσ/ n) for k {1, 2}, where a 1 = z.975 and a 2 = 1/.05. Via simulation, find the coverage rates of C 1 and C 2 for n {1, 10} when (a) Y 1,..., Y n i.i.d. N(µ, σ 2 ); (b) Y 1,..., Y n i.i.d. double exponential with variance 2. (c) Y 1,..., Y n i.i.d. beta(.1,.5). Include your code as an appendix to your homework. Discuss your results, and your thoughts on the robustness of the z-interval when the data are not normal (importantly, in this exercise, we are using the true variance of the population instead of an estimate). 20. Interval for a proportion: Let Y binomial(n, θ). (a) Find the approximate (large n) distribution of ˆθ = Y/n. Find a function of ˆθ that is approximately standard normal. 4

5 (b) Based on the normal approximation, obtain the form of an approximate 1 α CI for θ. Roughly how wide to you expect this to be for a given value of θ? (c) Obtain a CI for θ using Hoeffding s inequality. Compare the width of this CI to the approximate normal CI. 21. Convergence of correlated sums: Let {Y i : i N} be a vector of realvalued random variables with E[Y i ] = µ and Var[Y i ] = σ 2. obtain a WLLN for Ȳ = Y i /n in the cases (a) Cor[Y i, Y j ] = ρ for all i, j; (b) Cor[Y i, Y j ] = ρ if i j = 1 and is zero if i j > 1. Try to Discuss how correlation affects the asymptotic concentration of Ȳ around µ. 22. Weighted estimates: Sometimes our measurements of a quantity of interest have differing levels of precision. Let {Y i : i N} be a vector of independent real-valued random variables with E[Y i ] = µ and Var[Y i ] = σ 2 i. (a) Find the mean and variance of Ȳw = n i=1 w iy i, where the w i s are constants that sum to one. (b) Find the values of the w i s that minimize the variance of Ȳw. (c) Obtain a WLLN for Ȳw. 23. Moment generating functions: (a) Obtain the MGFs for the Poisson, exponential, and Gamma distributions. (b) Find the distributions of n i=1 Y i, where Y 1,..., Y n are i.i.d. Poisson, exponential, or gamma random variables. 5

6 24. Normal tail behavior: Show that if Z N(0, 1), then Pr(Z > t) φ(t)/t for t > 0. (Hint: Recall from the proof of Markov s inequality that zp(z) dz tp(z) dz). t t 25. Sketch a proof of the multivariate delta method. 26. Let Y 1,..., Y n be a sample from a bivariate population P where E[Y i ] = (µ A, µ B ), Var[Y i,a ] = Var[Y i,b ] = 1 and Cor[Y i,a, Y i,b ] = ρ. The purpose of this problem is to derive an estimator and standard error for ρ. (a) For this problem, what moments of P does ρ depend on? (b) Find CAN estimators of the moments from (a). (c) Find a CAN estimator ˆρ of ρ and give its limiting distribution. (d) Find a large-n estimate of the standard deviation of ρ, that is, its standard error. 27. Suppose you obtain a random sample from a population and the numerical values of the sample are (y 1,..., y n ), where each y i is a real valued number. Let ˆF be the empirical CDF based on these numbers, that is, ˆF (y) equals the fraction of yi s at or below the value y. (a) ˆF is discrete - where are the jumps? (b) How big are the jumps if there are no ties in the sample? What if there are ties? (c) ˆF is a valid CDF, and so it corresponds to a probability distribution, say ˆP, called the empirical distribution. distribution, that is, describe ˆP ((a, b]). 28. Confidence intervals and bands for CDFs. Describe this (a) Write down the formula for a 95% pointwise confidence interval for F (y), using a plug-in estimate of F (y) (hint: this is just the usual 6

7 normal interval for a binomial proportion, with F (y) replacing p, the population proportion). Also write down the formula for the 95% interval based on Hoeffding/DKW. (b) Simulate at least S = 10, 000 datasets consisting of samples of size n = 10 from the uniform distribution on [0, 1]. For each simulated dataset, check to see if the two confidence intervals cover the true value of F (y) for y {.1,.2,...,.8,.9}. For example, for each simulation s = 1,..., S you might make a vector c N s, where c N s [k] indicates whether or not the normal interval covers the true value of F (y) at the value k/10. (c) For both types of intervals use the results of the simulation to evaluate the pointwise coverage rates at y {.1,.2,...,.8,.9}; the global coverage rate, i.e., for what fraction of datasets did the intervals cover the true values of F at all y {.1,.2,...,.8,.9}. (d) Comment on the relative widths of the intervals, and summarize your findings about coverage rates. 29. Let Y 1,..., Y n i.i.d. from a distribution P with CDF F. Let ˆF be the empirical CDF of Y 1,..., Y n. For two points x, y with x < y, calculate E[ ˆF (y)], E[ ˆF (x) ˆF (y)] and Cov[ ˆF (x), ˆF (y)]. 30. Suppose we wish to simulate a bootstrap dataset Y = (Y 1,..., Y n ) from the empirical distribution of the observed sample values y = (y 1,..., y n ). Explain mathematically why this can be done with the R command Ystar <- sample(y,replace=true). 31. Suppose you observe an outcome y and a predictor x for a random sample of n = 10 objects, with y-values (2.38,2.72,-0.13,2.66,3.72,0.48,2.86,4.27,3.86, 2.04) and x-values (-0.63,0.18,-0.84,1.60,0.33,-0.82,0.49,0.74,0.58,-0.31). In other words, you sample (X 1, Y 1 ),..., (X n, Y n ) i.i.d. from some bivariate distribution, and these are the numerical results you get. Con- 7

8 sider the normal linear regression model y i ɛ 1,... ɛ n i.i.d. N(0, σ 2 ). = β 0 + β 1 x i + ɛ i, with (a) Obtain the usual normal-theory standard error, 95% CI and p- value for the OLS estimate of β 1 (use the lm command in R). (b) Obtain a bootstrap estimate of the standard deviation of ˆβ 1 (this is the bootstrap standard error), obtain a normal-theory CI for β 1 using the bootstrap standard error, and compare to the results in (a). (c) Obtain the bootstrap distribution of the p-value and display it graphically, including the observed p-value as a reference. How might you describe the evidence that β 1 0? How stable is this evidence? 32. Let Y 1,..., Y n i.i.d. from a continuous distribution P. Given Y 1,..., Y n, let Y 1,..., Y n i.i.d. from ˆP, the empirical distribution of Y 1,..., Y n. Let Ȳ = Y i /n. (a) Compute E[Ȳ ˆP ] and Var[Ȳ ˆP ]. (b) Now compute E[Ȳ ] and Var[Ȳ ], the unconditional expectation and variance of Ȳ, marginal over i.i.d. samples Y 1,..., Y n from P. 33. Suppose Y 1,..., Y n i.i.d. N(µ, σ 2 ). (a) What is the distribution of n(ȳ µ)/σ, and why? (b) What is the distribution of (n 1)s 2 /σ 2, and why? (c) What is the distribution of n(ȳ µ)/s, and why? (d) For w [0, 1], find the coverage rate for the set C w (Y ) = (Ȳ + s n t α(1 w), Ȳ + s n t 1 αw ). 34. Let Y N(µ, 1) and consider testing the hypothesis H : µ = 0. Consider an acceptance region of the form A 0 = (Y : z α(1 w) < Y < z 1 αw ). 8

9 (a) Show that the type I error rate of such a test is α for w [0, 1]. (b) Obtain the power of the test, that is, Pr(Y A 0 µ) as a function of µ. Make a plot of this power function for w = 1/2 and w = 1/4. When would you use w = 1/2? When would you use w = 1/4? 35. Suppose treatment A is assigned to a random selection of 5 experimental units, and 5 remaining experimental units are assigned treatment B. The observed treatment assignments and measured responses are (X 1,..., X 10 ) = (B, A, A, B, A, B, A, B, B, A), and (Y 1,..., Y 10 ) = (7.5, 1.2, 5.5, 2.2, 9.1, 8.7, 3.2, 5.1, 6.2, 1.7). (a) Assuming the A and B outcomes are random samples from N(µ A, σ 2 ) and N(µ B, σ 2 ) populations, compute the appropriate t-statistic for testing H : µ A = µ B, state the distribution of the statistic under H, and compute the p-value, (b) Using the same test statistic, do a permutation test of H : no treatment effect. Specifically, obtain the permeation null distribution, and compute the corresponding p-value. (c) Graphically compare the two null distributions, and compare the p-values. Describe the differences in assumptions that the two testing procedures make. (d) Obtain the permutation null distributions and p-values for the test statistics ȲA ȲB and ȲA/ȲB. 36. Let Y be a random variable and t(y ) be a test statistic. The p-value is p(y ) = Pr(t(Y ) > t(y )), where the distribution of Y is the null distribution P 0, with CDF F 0. Show that the distribution of p(y ) under Y P 0 is uniform on [0, 1] (Hint: Find the CDF of the p-value in terms of F 0 ). 37. Let Y 1,..., Y n i.i.d. P θ0 P. Find the log likelihood function and the form of the MLE in the cases where P is the set of 9

10 (a) Poisson distributions with mean θ R; (b) the multinomial distributions with probabilities (θ 1,..., θ p ); (c) the uniform distribution on (θ 1 θ 2 /2, θ 1 + θ 2 /2), with θ 1 R and θ 2 R Let Y i = e X i where X 1,..., X n i.i.d. N(µ, σ 2 ). Let φ = E[Y i ]. (a) Find the expectation and variance of Y i, and the expectation and variance of Ȳ, in terms of µ and σ2. (b) Find the MLE ˆφ of φ based on Y 1,..., Y n, and find an approximation to the variance of ˆφ. Discuss the magnitude of Var[ ˆφ] relative to Var[Ȳ ]. (c) Perform a simulation study where you compare ˆφ and Ȳ in terms of bias, variance and MSE. 39. Let f and g be discrete pdfs on {0, 1, 2,...} and define D(f, g) = y log(f(y)/g(y)) f(y). (a) Show that D(f, g) > 0 if f g and D(f, f) = 0. (b) Let g θ be the Poisson pdf with mean θ. Find the value of θ that minimizes D(f, g θ ), in terms of moments of f. 40. Consider a one parameter exponential family model, P = {P θ : θ Θ}, with densities f(y θ) = c(y) exp(θt(y) A(θ)) for θ Θ R. Here, t(y) is a scalar-valued function of the data point y. (a) For a sample of size n, write out the log-likelihood function and simplify as much as possible. (b) Find the likelihood equation, with the data on one side of the equation and items involving the parameter on the other. (c) Now take the derivative of p(y θ) with respect to θ and integrate to obtain a formula for the expectation of t(y). equation to the one obtained on (b), and comment. Compare this 10

11 41. Let (X i, Y i ) i.i.d. with Y i X i binary(e β 0+β 1 X i /(1 + e β 0+β 1 X i )) and X i P X. Our goal is to infer θ = (β 0, β 1 ). (a) Find a formula for the log-likelihood and the score function, and obtain equations that determine the MLE (i.e., the likelihood equations ). (b) Write down the observed information for θ, and compute the Fisher information. (c) Find the asymptotic distribution of ˆθ MLE as a function of the Fisher information. Is this usable for inference if P X is unknown? (d) Find another asymptotic approximation to the distribution of ˆθ MLE that can be used if P X is unknown. Describe how this approximation can be used to provide a hypothesis test of H : β 1 = Let Y 1,..., Y n i.i.d. gamma(a, b), parameterized so that E[Y i ] = a/b. (a) Write down the log-likelihood and obtain the likelihood equations. (b) Compute the Fisher information and use this to obtain a joint asymptotic distribution for (â MLE, ˆb MLE ). (c) Let µ = a/b. Obtain the asymptotic distribution of ˆµ MLE = â MLE /ˆb MLE. 43. Suppose Y 1,..., Y n i.i.d. gamma(a, b) as in the previous problem, but the statistician thinks that Y 1,..., Y n i.i.d. N(µ, σ 2 ) for some unknown values of µ, σ 2. (a) What values of (µ, σ 2 ) will maximize the expected log likelihood, E[log p(y µ, σ 2 )]? Here, the expectation is with respect to the true gamma distribution for Y, and your answer should depend on (a, b). (b) Make an argument that (ˆµ MLE, ˆσ MLE 2 ) converges in probability to something, say what that something is and explain your reasoning. 11

12 (c) What is the standard error of ˆµ MLE for the statistician who assumes normality? How does this compare to the standard error of a statistician who correctly assumes the gamma model (as in the previous problem)? (d) Discuss the consequences of model misspecification in this case. 44. Information inequalities: (a) Adapt the derivation of the Cramer-Rao information inequality to obtain a lower bound on the variance of a biased estimator. (b) For the model Y 1,..., Y n i.i.d. N(µ, σ 2 ), the posterior mean estimator ˆµ of µ under the prior µ N(0, τ 2 ) is ȳ (n/σ 2 )/(n/σ 2 + 1/τ 2 ). Use (a) to obtain a lower bound on the variance of ˆµ and compare to the actual variance, Var[ˆµ µ]. 45. Let X 1,..., X n i.i.d. gamma(a x, b x ) and Y 1,..., Y n i.i.d. gamma(a y, b y ). (a) Compute the (-2 log) likelihood ratio statistic for testing H : a x = a y, b x = b y, and state the asymptotic null distribution. (b) Simulate the actual null distribution of the statistic in the case that a x = a y = b x = b y = 1 for the sample sizes n = 5, 10, 20, 40, and compare to the asymptotic null distribution. 46. Let X 1,..., X m i.i.d. N(µ x, σ 2 ) and Y 1,..., Y n i.i.d. N(µ y, σ 2 ). (a) For the case that σ 2 is known, compute and compare the AIC and BIC for the two models corresponding to µ x = µ y and µ x µ y. For each model selection criterion, give the decision rule for choosing µ x µ y over µ x = µ y. Also compare these decision rules to deciding based on a level-α z-test. (b) Repeat for the case that σ 2 is unknown, but now compare AIC and BIC to deciding based on a level-α t-test. 12

13 (c) Now compute and compare the AIC and BIC decision rules for the case that the variances of the two population are not necessarily equal. 47. Let Y 1,..., Y n i.i.d. N(θ, 1). Using level-α n z-tests where α n depends on n, develop a consistent model selection procedure for choosing between θ = 0 and θ Let p 1,..., p m i.i.d. (1 γ)p 0 + γp 1, where P 0 is the uniform distribution on [0, 1] and P 1 is some other distribution, with CDF F 1. (a) Write out the probability that p 1 < α/m in terms of α, m, F 1, γ. (b) Write out the probability that the Bonferroni procedure rejects the global null hypothesis H 0 : γ = 0 at level α, that is, the probability that the smallest p-value is less than α/m. (c) Approximate the above probability using the approximation that log(1 x) x for small x. (d) Based on this approximation, evaluate if the probability of rejection is increasing or decreasing in α and in γ. Explain why your answers make sense. (e) What are conditions on F 1 that suggest (based on the approximation) that the Bonferroni procedure will have good power as m? 49. Let Y i θ i N(θ i, 1) independently for i = 1,..., m with m = 100. Using a Monte Carlo approximation, compute the probability of rejecting the global null H 0 : θ 1 =... = θ m = 0 at level α =.05 using the Bonferroni procedure, Fisher s procedure, and a test based on the statistic Y 2 i, under the following scenarios: (a) θ 1,..., θ m i.i.d. N(0, K/100) for K {1, 2, 4, 8, 16, 32}. (b) θ 1 = K and θ 2 =... = θ m = 0, where K {1, 2, 3, 4, 5, 6}. 13

14 50. Consider a model for m p-values, p 1,..., p m i.i.d. from a mixture distribution P = (1 γ)p 0 + γp 1, where P 0 is uniform on [0, 1] and P 1 is a beta(1, b) distribution. (a) Propose a modified Benjamini-Hochberg procedure to control the FDR at level α, in the case that γ and b are known. (b) Compute the mean and variance of p 1 in terms of γ and b. Using these calculations, propose moment-based estimators of γ and b using the observed values of p 1,..., p m. Based on this, propose a modified BH procedure that can be used if γ and b are not known. (*) Compare the FDR and the number of discoveries made by the BH and modified BH procedure in a simulation study, for the case that b {1, 2, 4, 8} and some interesting values of α and γ. 14

Problem Selected Scores

Problem Selected Scores Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected

More information

Monte Carlo Studies. The response in a Monte Carlo study is a random variable.

Monte Carlo Studies. The response in a Monte Carlo study is a random variable. Monte Carlo Studies The response in a Monte Carlo study is a random variable. The response in a Monte Carlo study has a variance that comes from the variance of the stochastic elements in the data-generating

More information

STA 2201/442 Assignment 2

STA 2201/442 Assignment 2 STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution

More information

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis. 401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis

More information

STAT 461/561- Assignments, Year 2015

STAT 461/561- Assignments, Year 2015 STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and

More information

Part 4: Multi-parameter and normal models

Part 4: Multi-parameter and normal models Part 4: Multi-parameter and normal models 1 The normal model Perhaps the most useful (or utilized) probability model for data analysis is the normal distribution There are several reasons for this, e.g.,

More information

The Delta Method and Applications

The Delta Method and Applications Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order

More information

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:. MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida First Year Examination Department of Statistics, University of Florida August 19, 010, 8:00 am - 1:00 noon Instructions: 1. You have four hours to answer questions in this examination.. You must show your

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Chapter 2. Discrete Distributions

Chapter 2. Discrete Distributions Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Problem 1 (20) Log-normal. f(x) Cauchy

Problem 1 (20) Log-normal. f(x) Cauchy ORF 245. Rigollet Date: 11/21/2008 Problem 1 (20) f(x) f(x) 0.0 0.1 0.2 0.3 0.4 0.0 0.2 0.4 0.6 0.8 4 2 0 2 4 Normal (with mean -1) 4 2 0 2 4 Negative-exponential x x f(x) f(x) 0.0 0.1 0.2 0.3 0.4 0.5

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Statistics & Data Sciences: First Year Prelim Exam May 2018

Statistics & Data Sciences: First Year Prelim Exam May 2018 Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book

More information

Bias Variance Trade-off

Bias Variance Trade-off Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]

More information

Contents 1. Contents

Contents 1. Contents Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

Review. December 4 th, Review

Review. December 4 th, Review December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter

More information

BTRY 4090: Spring 2009 Theory of Statistics

BTRY 4090: Spring 2009 Theory of Statistics BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)

More information

HT Introduction. P(X i = x i ) = e λ λ x i

HT Introduction. P(X i = x i ) = e λ λ x i MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework

More information

Master s Written Examination - Solution

Master s Written Examination - Solution Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2

More information

Bayesian performance

Bayesian performance Bayesian performance Frequentist properties of estimators refer to the performance of an estimator (say the posterior mean) over repeated experiments under the same conditions. The posterior distribution

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

ECON 4160, Autumn term Lecture 1

ECON 4160, Autumn term Lecture 1 ECON 4160, Autumn term 2017. Lecture 1 a) Maximum Likelihood based inference. b) The bivariate normal model Ragnar Nymoen University of Oslo 24 August 2017 1 / 54 Principles of inference I Ordinary least

More information

Masters Comprehensive Examination Department of Statistics, University of Florida

Masters Comprehensive Examination Department of Statistics, University of Florida Masters Comprehensive Examination Department of Statistics, University of Florida May 6, 003, 8:00 am - :00 noon Instructions: You have four hours to answer questions in this examination You must show

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

STATISTICS SYLLABUS UNIT I

STATISTICS SYLLABUS UNIT I STATISTICS SYLLABUS UNIT I (Probability Theory) Definition Classical and axiomatic approaches.laws of total and compound probability, conditional probability, Bayes Theorem. Random variable and its distribution

More information

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30 MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Bivariate Paired Numerical Data

Bivariate Paired Numerical Data Bivariate Paired Numerical Data Pearson s correlation, Spearman s ρ and Kendall s τ, tests of independence University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.

More information

Economics 583: Econometric Theory I A Primer on Asymptotics

Economics 583: Econometric Theory I A Primer on Asymptotics Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:

More information

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables. Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just

More information

Association studies and regression

Association studies and regression Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation IEOR E4703: Monte-Carlo Simulation Output Analysis for Monte-Carlo Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Output Analysis

More information

Regression and Statistical Inference

Regression and Statistical Inference Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF

More information

Data Mining Stat 588

Data Mining Stat 588 Data Mining Stat 588 Lecture 02: Linear Methods for Regression Department of Statistics & Biostatistics Rutgers University September 13 2011 Regression Problem Quantitative generic output variable Y. Generic

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Model Checking and Improvement

Model Checking and Improvement Model Checking and Improvement Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Model Checking All models are wrong but some models are useful George E. P. Box So far we have looked at a number

More information

Lecture 21: Convergence of transformations and generating a random variable

Lecture 21: Convergence of transformations and generating a random variable Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous

More information

parameter space Θ, depending only on X, such that Note: it is not θ that is random, but the set C(X).

parameter space Θ, depending only on X, such that Note: it is not θ that is random, but the set C(X). 4. Interval estimation The goal for interval estimation is to specify the accurary of an estimate. A 1 α confidence set for a parameter θ is a set C(X) in the parameter space Θ, depending only on X, such

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood Regression Estimation - Least Squares and Maximum Likelihood Dr. Frank Wood Least Squares Max(min)imization Function to minimize w.r.t. β 0, β 1 Q = n (Y i (β 0 + β 1 X i )) 2 i=1 Minimize this by maximizing

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 1 Bootstrapped Bias and CIs Given a multiple regression model with mean and

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

Primer on statistics:

Primer on statistics: Primer on statistics: MLE, Confidence Intervals, and Hypothesis Testing ryan.reece@gmail.com http://rreece.github.io/ Insight Data Science - AI Fellows Workshop Feb 16, 018 Outline 1. Maximum likelihood

More information

Confidence Intervals, Testing and ANOVA Summary

Confidence Intervals, Testing and ANOVA Summary Confidence Intervals, Testing and ANOVA Summary 1 One Sample Tests 1.1 One Sample z test: Mean (σ known) Let X 1,, X n a r.s. from N(µ, σ) or n > 30. Let The test statistic is H 0 : µ = µ 0. z = x µ 0

More information

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1 Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Answer Key for STAT 200B HW No. 7

Answer Key for STAT 200B HW No. 7 Answer Key for STAT 200B HW No. 7 May 5, 2007 Problem 2.2 p. 649 Assuming binomial 2-sample model ˆπ =.75, ˆπ 2 =.6. a ˆτ = ˆπ 2 ˆπ =.5. From Ex. 2.5a on page 644: ˆπ ˆπ + ˆπ 2 ˆπ 2.75.25.6.4 = + =.087;

More information

Outline. Motivation Contest Sample. Estimator. Loss. Standard Error. Prior Pseudo-Data. Bayesian Estimator. Estimators. John Dodson.

Outline. Motivation Contest Sample. Estimator. Loss. Standard Error. Prior Pseudo-Data. Bayesian Estimator. Estimators. John Dodson. s s Practitioner Course: Portfolio Optimization September 24, 2008 s The Goal of s The goal of estimation is to assign numerical values to the parameters of a probability model. Considerations There are

More information

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Fall, 2013 Page 1 Random Variable and Probability Distribution Discrete random variable Y : Finite possible values {y

More information

Robustness and Distribution Assumptions

Robustness and Distribution Assumptions Chapter 1 Robustness and Distribution Assumptions 1.1 Introduction In statistics, one often works with model assumptions, i.e., one assumes that data follow a certain model. Then one makes use of methodology

More information

Estimation, Inference, and Hypothesis Testing

Estimation, Inference, and Hypothesis Testing Chapter 2 Estimation, Inference, and Hypothesis Testing Note: The primary reference for these notes is Ch. 7 and 8 of Casella & Berger 2. This text may be challenging if new to this topic and Ch. 7 of

More information

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That Statistics Lecture 2 August 7, 2000 Frank Porter Caltech The plan for these lectures: The Fundamentals; Point Estimation Maximum Likelihood, Least Squares and All That What is a Confidence Interval? Interval

More information

Practice Problems Section Problems

Practice Problems Section Problems Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1

TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1 TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1 1.1 The Probability Model...1 1.2 Finite Discrete Models with Equally Likely Outcomes...5 1.2.1 Tree Diagrams...6 1.2.2 The Multiplication Principle...8

More information

Mathematical statistics

Mathematical statistics October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu

More information

Maximum Likelihood, Logistic Regression, and Stochastic Gradient Training

Maximum Likelihood, Logistic Regression, and Stochastic Gradient Training Maximum Likelihood, Logistic Regression, and Stochastic Gradient Training Charles Elkan elkan@cs.ucsd.edu January 17, 2013 1 Principle of maximum likelihood Consider a family of probability distributions

More information

ST440/540: Applied Bayesian Statistics. (9) Model selection and goodness-of-fit checks

ST440/540: Applied Bayesian Statistics. (9) Model selection and goodness-of-fit checks (9) Model selection and goodness-of-fit checks Objectives In this module we will study methods for model comparisons and checking for model adequacy For model comparisons there are a finite number of candidate

More information

Estimators as Random Variables

Estimators as Random Variables Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until

More information

ECE531 Lecture 10b: Maximum Likelihood Estimation

ECE531 Lecture 10b: Maximum Likelihood Estimation ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So

More information

STA 732: Inference. Notes 10. Parameter Estimation from a Decision Theoretic Angle. Other resources

STA 732: Inference. Notes 10. Parameter Estimation from a Decision Theoretic Angle. Other resources STA 732: Inference Notes 10. Parameter Estimation from a Decision Theoretic Angle Other resources 1 Statistical rules, loss and risk We saw that a major focus of classical statistics is comparing various

More information

Statistics 135 Fall 2008 Final Exam

Statistics 135 Fall 2008 Final Exam Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations

More information

Linear Methods for Prediction

Linear Methods for Prediction This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

Linear models and their mathematical foundations: Simple linear regression

Linear models and their mathematical foundations: Simple linear regression Linear models and their mathematical foundations: Simple linear regression Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/21 Introduction

More information

Generalized Linear Models Introduction

Generalized Linear Models Introduction Generalized Linear Models Introduction Statistics 135 Autumn 2005 Copyright c 2005 by Mark E. Irwin Generalized Linear Models For many problems, standard linear regression approaches don t work. Sometimes,

More information

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Review Quiz 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Cramér Rao lower bound (CRLB). That is, if where { } and are scalars, then

More information

A General Overview of Parametric Estimation and Inference Techniques.

A General Overview of Parametric Estimation and Inference Techniques. A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

Lecture 6: Discrete Choice: Qualitative Response

Lecture 6: Discrete Choice: Qualitative Response Lecture 6: Instructor: Department of Economics Stanford University 2011 Types of Discrete Choice Models Univariate Models Binary: Linear; Probit; Logit; Arctan, etc. Multinomial: Logit; Nested Logit; GEV;

More information

Likelihood-based inference with missing data under missing-at-random

Likelihood-based inference with missing data under missing-at-random Likelihood-based inference with missing data under missing-at-random Jae-kwang Kim Joint work with Shu Yang Department of Statistics, Iowa State University May 4, 014 Outline 1. Introduction. Parametric

More information

P n. This is called the law of large numbers but it comes in two forms: Strong and Weak.

P n. This is called the law of large numbers but it comes in two forms: Strong and Weak. Large Sample Theory Large Sample Theory is a name given to the search for approximations to the behaviour of statistical procedures which are derived by computing limits as the sample size, n, tends to

More information

Probability and Estimation. Alan Moses

Probability and Estimation. Alan Moses Probability and Estimation Alan Moses Random variables and probability A random variable is like a variable in algebra (e.g., y=e x ), but where at least part of the variability is taken to be stochastic.

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

Stat 5102 Final Exam May 14, 2015

Stat 5102 Final Exam May 14, 2015 Stat 5102 Final Exam May 14, 2015 Name Student ID The exam is closed book and closed notes. You may use three 8 1 11 2 sheets of paper with formulas, etc. You may also use the handouts on brand name distributions

More information

COS513 LECTURE 8 STATISTICAL CONCEPTS

COS513 LECTURE 8 STATISTICAL CONCEPTS COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b) LECTURE 5 NOTES 1. Bayesian point estimators. In the conventional (frequentist) approach to statistical inference, the parameter θ Θ is considered a fixed quantity. In the Bayesian approach, it is considered

More information