SUFFICIENT STATISTICS
|
|
- Willa Bradford
- 5 years ago
- Views:
Transcription
1 SUFFICIENT STATISTICS. Introduction Let X (X,..., X n ) be a random sample from f θ, where θ Θ is unknown. We are interested using X to estimate θ. In the simple case where X i Bern(p), we found that the sample mean was an efficient estimator for p. Thus, if we observe a finite sequence of coin flips, in order to have an efficient estimate of the probability, p, that heads occurs in a single flip, we need only count the number of times we see heads (and divide by the total number of flips), and we need not worry about the order in which the heads and tails occurred. Note that the sequences 000 and 000 lead to the same estimate, when we use the sample mean. In what follows, we want to study the following question: do we get any additional information about p by making use of the order in which the heads and tails occurred? The sample mean does not make use of the order, and it does give us an efficient estimator, so, in short the answer in this case is no. Thus in this example, it appears that we can greatly simplify and reduce the amount of data, without affecting our ability to find good estimators. 2. Sufficient statistics Let X (X,..., X n ) be a random sample from f θ, where θ Θ is unknown. Recall that a T is a statistic if T T (X) u(x) for some deterministic function u. We will assume that u does not depend on θ. Some examples of that you are familiar with are when T is the sample mean, the sample variance, and the maximum. Let us remark that although in many important examples, T is one dimensional point estimator, it need not be, for example, T (X) X is a statistic. We say that T is sufficient for θ if the conditional distribution of X given T does not depend on θ. In the case were the random variables involved are not discrete, even this definition requires somewhat advanced mathematics, since we might have that P(T t) 0, in which case it is not immediate how one can make sense of P(X T t) We will first discuss the discrete case, and then we will extend our discussion to the continuous case.
2 2 SUFFICIENT STATISTICS 3. The discrete case Exercise. Let X (X,..., X n ) be a random sample, where X i Bern(p). Show that the sample sum given by T X + + X n is a sufficient statistic for p. Solution. Let x {0, } n and t {0,,... n}. We need to show that P(X x T t) does not depend on p. In fact, you already did this computation in the first homework! By definition, P(X x T t) P(X x, T t). P(T t) We may assume that t t(x) x + +x n, otherwise, P(X x, T t) 0. Thus, {X x} {X x} {T t}, and P(X x, T t) P(X x). We have that P(X x) L(x; p) p x i ( p) x i p t ( p) n t We also know that T Bin(n, p), so that ( ) n P(T t) p t ( p) n t. t Hence we obtain that which does not depend on p. P(X x T t) ( n t), Exercise 2. Discuss why you should expect that the final answer we obtained in Exercise is P(X x T t) ( n t). In the discrete setting, we have that T is sufficient for θ if and only if for all x and t t(x), we have P(X x T t) P(X x, T t) P(T t) for some function of H(x) which does not depend on θ. P(X x) P(T t) H(x), Exercise 3. Let X (X,..., X n ) be a random sample, where X i is a discrete random variable that is uniformly distributed in {, 2,..., θ}. Show that M max {X,..., X n } is a sufficient statistic for θ.
3 Solution. Let m {, 2,..., θ}. Note that and Hence SUFFICIENT STATISTICS 3 {M m} n {X i m} {M m} {M m} \ {M m }. P(M m) θ n (mn (m ) n ). Let x {, 2,..., θ} n and m max {x,..., x n }. so we are done. P(X x M m) P(X x, M m) P(M m) P(X x) P(M m) θ n (m θ n (m ) n ) n m n (m ) n, Exercise 4. Let X (X,..., X n ) be a random sample, where X i is a Poisson random variable with mean λ. Show that the sample sum given by T X + + X n is a sufficient statistic for λ. In order to have some more examples to discuss, recall that X is a geometric random variable with parameter p (0, ), if P(X k) p( p) k, for k, 2,.... Thus X is the number of Bernoulli p trials required to get a success. Here, EX /p. Let us remark that sometimes geometric random variables are defined so that P(X k) p( p) k, for k 0,, 2,... ; in this case X is the number of fails before a success, and EX p/( p). Before we find a sufficient statistic for p, we do a couple of preliminary exercises. Exercise 5. Let X (X,..., X n ) be a random sample, where X i is a geometric random variable with parameter p (0, ) and mean /p. Show that the mle for p is given by / X.
4 4 SUFFICIENT STATISTICS Exercise 6. Referring to Exercise 5, let T X + + X n. Show that for k n, n +,..., we have ( ) k P(T k) p n ( p) k n. n Solution. Note that T is the number for trials required to get n success. By counting we obtain the required formula: the last kth trial is a success, and you are left with k trials, of which n of them must be successes. Exercise 7. Referring to Exercise 5, show that the sample sum given by T X + + X n is a sufficient statistic for p. Solution. Let x {, 2, 3, 4,...} n and t x + + x n. We have that P(X x T t) which does not depend on p. P(X x) P(T t) n p( p)x i ) pn ( p) t n ( t n ( t n ), 4. The continuous case In the continuous case, as in the case of likelihoods we work with the density functions instead of the probabilities directly. Let X (X,..., X n ) be a random sample from f θ, where θ Θ is unknown. Let T u(x) be a statistic with density function q(t). Then T is a sufficient statistic for θ if for all x and t t(x), we have L(x; θ) n q(t(x)) f(x i; θ) H(x), q(t(x)) for some function H which does not depend on θ. Exercise 8. Let X (X,..., X n ) be a random sample, where X i Unif(0, θ), where θ is unknown. Show that M max {X,..., X n } is a sufficient statistic for θ. Exercise 9. Let X (X,..., X n ) be a random sample, where X i N(µ, ), where µ is unknown. Show that the sample mean is a sufficient statistic for µ.
5 SUFFICIENT STATISTICS 5 Solution. Luckily, we know that the distribution for X; we have that X N(µ, /n). However, even with this piece of knowledge, this is a tricky exercise. First, we need the following observation. Note that (x i x) 0. Thus (x i µ) 2 (x i x + x µ) 2 ) ((x i x) 2 + 2(x i x)( x µ) + ( x µ) 2 ) ((x i x) 2 + ( x µ) 2 (x i x) 2 + n( x µ) 2 With this algebra in hand, we have that n 2π e (x i µ)2 2 n 2π e n( x µ)2 2 n (x i x) 2 n(2π) (n )/2 e 2, which does not depend on µ. 5. Fisher-Neyman factorization We saw in the previous exercises that proving that a statistic is sufficient from the definition can be quite challenging. The following theorem factorization theorem makes life easier. Theorem 0. Let X (X,..., X n ) be a random sample from the pdf f θ, where θ Θ is unknown. A statistic T is sufficient for θ if and only if there exists nonnegative functions g(t; θ) and h(x) (which does not depend on θ) such that for all points x and all θ Θ, we have L(x; θ) f(x i ; θ) g(t(x); θ)h(x). Clearly, by definition, a factorization holds if T is sufficient, so one direction of the proof is trivial. It is also immediate for Theorem 0, that a - function of a sufficient statistic is again sufficient. Let us also remark in Theorem 0, g(t; θ) does not have to be the density
6 6 SUFFICIENT STATISTICS function for T (X), and in the discrete case, we do not require that g(t) P(T t). The factorization of Theorem 0 is not unique. The utility of Theorem 0 lies in the fact that we do not need to identify the distribution of T. Before we prove the non-trivial direction of Theorem 0, let us apply it Exercise 9. Exercise. Apply Theorem 0 to solve Exercise 9. Solution (Solution to Exercise 9). The difference here is we still need the somewhat tricky algebra, but we no longer need to know that sum of independent normals is again normal. L(x; µ) 2π e (x i µ)2 2 (2π) n/2 e 2 n (x i x) 2 e n 2 ( x µ)2. Thus we choose, g( x; µ) e n 2 ( x µ)2 and h(x) (2π) n/2 e 2 n (x i x) 2. Exercise 2. Let X (X,..., X n ) be a random sample, where X i N(0, θ), where the variance θ is unknown. Show that T n X2 i is a sufficient statistic for θ. Exercise 3. Let X (X,..., X n ) be a random sample, where X i N(µ, σ 2 ), where both µ and σ 2 are unknown. Set θ (µ, σ 2 ). Let T ( X, S 2 ), where X is the usual sample mean, and S 2 is the usual sample variance. Show that L(x; θ) g(t(x); θ)h(x), some functions g and h, so that T is a sufficient statistic for θ. Exercise 4. Apply Theorem 0 to solve Exercise 4. Solution. Let x {0,, 2...} n, and t t(x) x + + x n. We have that P(X x) e λ λx i x i! λt e nλ x i!. Thus we choose g(t; λ) λ t e nλ and h(x) n. x i! Exercise 5. Let X (X,..., X n ) be a random sample, where X is a real-valued continuous random variable with a pdf given by f(x ; θ) h(x )c(θ)e w(θ)u(x ) Show that T n u(x i) is a sufficient statistic for θ.
7 SUFFICIENT STATISTICS 7 Proof Theorem 0 (discrete case). Let t t(x). We have by assumption that P(X x) g(t; θ) h(x). P(T t) P(T t) Let us remark that we do not have that g(t; θ) P(T t). Of course, P(T t) P θ (T t) depends on θ, and the claim is that the θ s in g(t; θ) cancel out the θ s in P(T t). To see why, let A : {y : t(y) t(x)}. Of course, x A, but there could be other elements; think of t as the sample sum, then if t(x) t, any other permutation of y of x, we have t(y) t. Thus, Hence P(T t) P(A) y A which does not depend on θ. P(X x) P(T t) P(X y) g(t; θ) y A h(y). h(x) y A h(y), The proof in the continuous case is more technical; your text has a proof of a special case of the continuous case. The above proof is similar to the proof of the following elementary fact. Theorem 6. Let X be a discrete random variable with pdf f. g : R R, then Eg(X) g(x)f(x), x whenever the sum is absolutely convergent. Proof. We have that Eg(X) y yp(g(x) y). Suppose X takes values on the set A. Let A y : {x A : g(x) y}. Note that the sets A y partition the set A. Thus P(g(X) y) P(A y ) x A y f(x) If and Eg(X) yf(x) g(x)f(x) g(x)f(x). y x A y y x A y End of Midterm coverage x A
1 Probability Model. 1.1 Types of models to be discussed in the course
Sufficiency January 11, 2016 Debdeep Pati 1 Probability Model Model: A family of distributions {P θ : θ Θ}. P θ (B) is the probability of the event B when the parameter takes the value θ. P θ is described
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationBEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.
BEST TESTS Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. 1. Most powerful test Let {f θ } θ Θ be a family of pdfs. We will consider
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More informationMathematical statistics
October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:
More informationRandom variables. DS GA 1002 Probability and Statistics for Data Science.
Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities
More information1 Probability Model. 1.1 Types of models to be discussed in the course
Sufficiency January 18, 016 Debdeep Pati 1 Probability Model Model: A family of distributions P θ : θ Θ}. P θ (B) is the probability of the event B when the parameter takes the value θ. P θ is described
More information3 Multiple Discrete Random Variables
3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f
More informationFebruary 26, 2017 COMPLETENESS AND THE LEHMANN-SCHEFFE THEOREM
February 26, 2017 COMPLETENESS AND THE LEHMANN-SCHEFFE THEOREM Abstract. The Rao-Blacwell theorem told us how to improve an estimator. We will discuss conditions on when the Rao-Blacwellization of an estimator
More informationTopic 15: Simple Hypotheses
Topic 15: November 10, 2009 In the simplest set-up for a statistical hypothesis, we consider two values θ 0, θ 1 in the parameter space. We write the test as H 0 : θ = θ 0 versus H 1 : θ = θ 1. H 0 is
More informationLecture 3. Discrete Random Variables
Math 408 - Mathematical Statistics Lecture 3. Discrete Random Variables January 23, 2013 Konstantin Zuev (USC) Math 408, Lecture 3 January 23, 2013 1 / 14 Agenda Random Variable: Motivation and Definition
More informationMarch 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS
March 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS Abstract. We will introduce a class of distributions that will contain many of the discrete and continuous we are familiar with. This class will help
More informationChapter 3: Random Variables 1
Chapter 3: Random Variables 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More information1. (Regular) Exponential Family
1. (Regular) Exponential Family The density function of a regular exponential family is: [ ] Example. Poisson(θ) [ ] Example. Normal. (both unknown). ) [ ] [ ] [ ] [ ] 2. Theorem (Exponential family &
More informationLast Lecture - Key Questions. Biostatistics Statistical Inference Lecture 03. Minimal Sufficient Statistics
Last Lecture - Key Questions Biostatistics 602 - Statistical Inference Lecture 03 Hyun Min Kang January 17th, 2013 1 How do we show that a statistic is sufficient for θ? 2 What is a necessary and sufficient
More informationINTRODUCTION TO BAYESIAN METHODS II
INTRODUCTION TO BAYESIAN METHODS II Abstract. We will revisit point estimation and hypothesis testing from the Bayesian perspective.. Bayes estimators Let X = (X,..., X n ) be a random sample from the
More informationECE 275B Homework # 1 Solutions Version Winter 2015
ECE 275B Homework # 1 Solutions Version Winter 2015 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2
More informationChapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.
Chapter 2 Random Variable CLO2 Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance. 1 1. Introduction In Chapter 1, we introduced the concept
More informationFundamental Tools - Probability Theory II
Fundamental Tools - Probability Theory II MSc Financial Mathematics The University of Warwick September 29, 2015 MSc Financial Mathematics Fundamental Tools - Probability Theory II 1 / 22 Measurable random
More informationPage Max. Possible Points Total 100
Math 3215 Exam 2 Summer 2014 Instructor: Sal Barone Name: GT username: 1. No books or notes are allowed. 2. You may use ONLY NON-GRAPHING and NON-PROGRAMABLE scientific calculators. All other electronic
More informationECE 275B Homework # 1 Solutions Winter 2018
ECE 275B Homework # 1 Solutions Winter 2018 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2 < < x n Thus,
More informationChapters 9. Properties of Point Estimators
Chapters 9. Properties of Point Estimators Recap Target parameter, or population parameter θ. Population distribution f(x; θ). { probability function, discrete case f(x; θ) = density, continuous case The
More informationSTAT2201. Analysis of Engineering & Scientific Data. Unit 3
STAT2201 Analysis of Engineering & Scientific Data Unit 3 Slava Vaisman The University of Queensland School of Mathematics and Physics What we learned in Unit 2 (1) We defined a sample space of a random
More informationMAT 271E Probability and Statistics
MAT 71E Probability and Statistics Spring 013 Instructor : Class Meets : Office Hours : Textbook : Supp. Text : İlker Bayram EEB 1103 ibayram@itu.edu.tr 13.30 1.30, Wednesday EEB 5303 10.00 1.00, Wednesday
More informationTom Salisbury
MATH 2030 3.00MW Elementary Probability Course Notes Part V: Independence of Random Variables, Law of Large Numbers, Central Limit Theorem, Poisson distribution Geometric & Exponential distributions Tom
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 10: Expectation and Variance Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/ psarkar/teaching
More informationMath 564 Homework 1. Solutions.
Math 564 Homework 1. Solutions. Problem 1. Prove Proposition 0.2.2. A guide to this problem: start with the open set S = (a, b), for example. First assume that a >, and show that the number a has the properties
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationF79SM STATISTICAL METHODS
F79SM STATISTICAL METHODS SUMMARY NOTES 9 Hypothesis testing 9.1 Introduction As before we have a random sample x of size n of a population r.v. X with pdf/pf f(x;θ). The distribution we assign to X is
More informationSUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)
SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems
More informationPCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities
PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets
More informationChapter 3: Random Variables 1
Chapter 3: Random Variables 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More information3 Modeling Process Quality
3 Modeling Process Quality 3.1 Introduction Section 3.1 contains basic numerical and graphical methods. familiar with these methods. It is assumed the student is Goal: Review several discrete and continuous
More informationTHE QUEEN S UNIVERSITY OF BELFAST
THE QUEEN S UNIVERSITY OF BELFAST 0SOR20 Level 2 Examination Statistics and Operational Research 20 Probability and Distribution Theory Wednesday 4 August 2002 2.30 pm 5.30 pm Examiners { Professor R M
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationConditioning a random variable on an event
Conditioning a random variable on an event Let X be a continuous random variable and A be an event with P (A) > 0. Then the conditional pdf of X given A is defined as the nonnegative function f X A that
More information1 Review of Probability
1 Review of Probability Random variables are denoted by X, Y, Z, etc. The cumulative distribution function (c.d.f.) of a random variable X is denoted by F (x) = P (X x), < x
More informationChapter 2.5 Random Variables and Probability The Modern View (cont.)
Chapter 2.5 Random Variables and Probability The Modern View (cont.) I. Statistical Independence A crucially important idea in probability and statistics is the concept of statistical independence. Suppose
More informationStat410 Probability and Statistics II (F16)
Stat4 Probability and Statistics II (F6 Exponential, Poisson and Gamma Suppose on average every /λ hours, a Stochastic train arrives at the Random station. Further we assume the waiting time between two
More informationStat 134 Fall 2011: Notes on generating functions
Stat 3 Fall 0: Notes on generating functions Michael Lugo October, 0 Definitions Given a random variable X which always takes on a positive integer value, we define the probability generating function
More informationEXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS
EXAM Exam # Math 3342 Summer II, 2 July 2, 2 ANSWERS i pts. Problem. Consider the following data: 7, 8, 9, 2,, 7, 2, 3. Find the first quartile, the median, and the third quartile. Make a box and whisker
More informationNotes on Random Vectors and Multivariate Normal
MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution
More informationReview. December 4 th, Review
December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter
More information1 Complete Statistics
Complete Statistics February 4, 2016 Debdeep Pati 1 Complete Statistics Suppose X P θ, θ Θ. Let (X (1),..., X (n) ) denote the order statistics. Definition 1. A statistic T = T (X) is complete if E θ g(t
More informationChapter 8: Least squares (beginning of chapter)
Chapter 8: Least squares (beginning of chapter) Least Squares So far, we have been trying to determine an estimator which was unbiased and had minimum variance. Next we ll consider a class of estimators
More informationMath LM (27794) - Lectures 01
Math 37500 -LM (27794) - Lectures 01 Ethan Akin Office: NAC 6/287 Phone: 650-5136 Email: ethanakin@earthlink.net Fall, 2018 Contents Probability and Counting, Chapter 1 Counting, Sections 1.3, 1.4 Adjusting
More informationOutline. 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks
Outline 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks Likelihood A common and fruitful approach to statistics is to assume
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationMath Bootcamp 2012 Miscellaneous
Math Bootcamp 202 Miscellaneous Factorial, combination and permutation The factorial of a positive integer n denoted by n!, is the product of all positive integers less than or equal to n. Define 0! =.
More informationMathematical Statistics
Mathematical Statistics Chapter Three. Point Estimation 3.4 Uniformly Minimum Variance Unbiased Estimator(UMVUE) Criteria for Best Estimators MSE Criterion Let F = {p(x; θ) : θ Θ} be a parametric distribution
More informationDiscrete Distributions
Discrete Distributions STA 281 Fall 2011 1 Introduction Previously we defined a random variable to be an experiment with numerical outcomes. Often different random variables are related in that they have
More informationST5215: Advanced Statistical Theory
Department of Statistics & Applied Probability Wednesday, October 19, 2011 Lecture 17: UMVUE and the first method of derivation Estimable parameters Let ϑ be a parameter in the family P. If there exists
More informationMath 341: Probability Eighth Lecture (10/6/09)
Math 341: Probability Eighth Lecture (10/6/09) Steven J Miller Williams College Steven.J.Miller@williams.edu http://www.williams.edu/go/math/sjmiller/ public html/341/ Bronfman Science Center Williams
More informationExample: An experiment can either result in success or failure with probability θ and (1 θ) respectively. The experiment is performed independently
Chapter 3 Sufficient statistics and variance reduction Let X 1,X 2,...,X n be a random sample from a certain distribution with p.m/d.f fx θ. A function T X 1,X 2,...,X n = T X of these observations is
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationContinuous Distributions
Page 1 Chapter 6 Continuous Distributions In principle it is easy to calculate probabilities such as P{Bin(3, p) 17} for various values of p: one has only to sum the series ( ) ( ) 3 3 p 17 (1 p) 13 +
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationMA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems
MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions
More information1 Poisson processes, and Compound (batch) Poisson processes
Copyright c 2007 by Karl Sigman 1 Poisson processes, and Compound (batch) Poisson processes 1.1 Point Processes Definition 1.1 A simple point process ψ = {t n : n 1} is a sequence of strictly increasing
More informationMidterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm
Midterm Examination STA 215: Statistical Inference Due Wednesday, 2006 Mar 8, 1:15 pm This is an open-book take-home examination. You may work on it during any consecutive 24-hour period you like; please
More informationChapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic
Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic Unbiased estimation Unbiased or asymptotically unbiased estimation plays an important role in
More informationChapter 5. Random Variables (Continuous Case) 5.1 Basic definitions
Chapter 5 andom Variables (Continuous Case) So far, we have purposely limited our consideration to random variables whose ranges are countable, or discrete. The reason for that is that distributions on
More informationNorthwestern University Department of Electrical Engineering and Computer Science
Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability
More informationS n = x + X 1 + X X n.
0 Lecture 0 0. Gambler Ruin Problem Let X be a payoff if a coin toss game such that P(X = ) = P(X = ) = /2. Suppose you start with x dollars and play the game n times. Let X,X 2,...,X n be payoffs in each
More informationIEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008
IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008 Justify your answers; show your work. 1. A sequence of Events. (10 points) Let {B n : n 1} be a sequence of events in
More informationHypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes
Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis
More informationLecture 16. Lectures 1-15 Review
18.440: Lecture 16 Lectures 1-15 Review Scott Sheffield MIT 1 Outline Counting tricks and basic principles of probability Discrete random variables 2 Outline Counting tricks and basic principles of probability
More information1: PROBABILITY REVIEW
1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following
More informationStat Lecture 20. Last class we introduced the covariance and correlation between two jointly distributed random variables.
Stat 260 - Lecture 20 Recap of Last Class Last class we introduced the covariance and correlation between two jointly distributed random variables. Today: We will introduce the idea of a statistic and
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationTopic 3: The Expectation of a Random Variable
Topic 3: The Expectation of a Random Variable Course 003, 2017 Page 0 Expectation of a discrete random variable Definition (Expectation of a discrete r.v.): The expected value (also called the expectation
More informationParametric Models: from data to models
Parametric Models: from data to models Pradeep Ravikumar Co-instructor: Manuela Veloso Machine Learning 10-701 Jan 22, 2018 Recall: Model-based ML DATA MODEL LEARNING MODEL MODEL INFERENCE KNOWLEDGE Learning:
More informationChapter Generating Functions
Chapter 8.1.1-8.1.2. Generating Functions Prof. Tesler Math 184A Fall 2017 Prof. Tesler Ch. 8. Generating Functions Math 184A / Fall 2017 1 / 63 Ordinary Generating Functions (OGF) Let a n (n = 0, 1,...)
More informationReview of Discrete Probability (contd.)
Stat 504, Lecture 2 1 Review of Discrete Probability (contd.) Overview of probability and inference Probability Data generating process Observed data Inference The basic problem we study in probability:
More informationSTAT 135 Lab 3 Asymptotic MLE and the Method of Moments
STAT 135 Lab 3 Asymptotic MLE and the Method of Moments Rebecca Barter February 9, 2015 Maximum likelihood estimation (a reminder) Maximum likelihood estimation Suppose that we have a sample, X 1, X 2,...,
More informationGuidelines for Solving Probability Problems
Guidelines for Solving Probability Problems CS 1538: Introduction to Simulation 1 Steps for Problem Solving Suggested steps for approaching a problem: 1. Identify the distribution What distribution does
More informationChapter 8. Some Approximations to Probability Distributions: Limit Theorems
Chapter 8. Some Approximations to Probability Distributions: Limit Theorems Sections 8.2 -- 8.3: Convergence in Probability and in Distribution Jiaping Wang Department of Mathematical Science 04/22/2013,
More informationHomework for 1/13 Due 1/22
Name: ID: Homework for 1/13 Due 1/22 1. [ 5-23] An irregularly shaped object of unknown area A is located in the unit square 0 x 1, 0 y 1. Consider a random point distributed uniformly over the square;
More informationContinuous Distributions
Continuous Distributions 1.8-1.9: Continuous Random Variables 1.10.1: Uniform Distribution (Continuous) 1.10.4-5 Exponential and Gamma Distributions: Distance between crossovers Prof. Tesler Math 283 Fall
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationQuick Tour of Basic Probability Theory and Linear Algebra
Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationClassical Probability
Chapter 1 Classical Probability Probability is the very guide of life. Marcus Thullius Cicero The original development of probability theory took place during the seventeenth through nineteenth centuries.
More informationProbability inequalities 11
Paninski, Intro. Math. Stats., October 5, 2005 29 Probability inequalities 11 There is an adage in probability that says that behind every limit theorem lies a probability inequality (i.e., a bound on
More information2 Random Variable Generation
2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationMATH Solutions to Probability Exercises
MATH 5 9 MATH 5 9 Problem. Suppose we flip a fair coin once and observe either T for tails or H for heads. Let X denote the random variable that equals when we observe tails and equals when we observe
More informationBinomial random variable
Binomial random variable Toss a coin with prob p of Heads n times X: # Heads in n tosses X is a Binomial random variable with parameter n,p. X is Bin(n, p) An X that counts the number of successes in many
More informationThings to remember when learning probability distributions:
SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions
More informationIntroduction and Preliminaries
Chapter 1 Introduction and Preliminaries This chapter serves two purposes. The first purpose is to prepare the readers for the more systematic development in later chapters of methods of real analysis
More informationMultivariate Distributions (Hogg Chapter Two)
Multivariate Distributions (Hogg Chapter Two) STAT 45-1: Mathematical Statistics I Fall Semester 15 Contents 1 Multivariate Distributions 1 11 Random Vectors 111 Two Discrete Random Variables 11 Two Continuous
More informationCME 106: Review Probability theory
: Probability theory Sven Schmit April 3, 2015 1 Overview In the first half of the course, we covered topics from probability theory. The difference between statistics and probability theory is the following:
More informationProbability, Random Processes and Inference
INSTITUTO POLITÉCNICO NACIONAL CENTRO DE INVESTIGACION EN COMPUTACION Laboratorio de Ciberseguridad Probability, Random Processes and Inference Dr. Ponciano Jorge Escamilla Ambrosio pescamilla@cic.ipn.mx
More information1 Random Variable: Topics
Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?
More informationBrief Review of Probability
Maura Department of Economics and Finance Università Tor Vergata Outline 1 Distribution Functions Quantiles and Modes of a Distribution 2 Example 3 Example 4 Distributions Outline Distribution Functions
More informationTopic 9 Examples of Mass Functions and Densities
Topic 9 Examples of Mass Functions and Densities Discrete Random Variables 1 / 12 Outline Bernoulli Binomial Negative Binomial Poisson Hypergeometric 2 / 12 Introduction Write f X (x θ) = P θ {X = x} for
More informationProperties of Linear Transformations from R n to R m
Properties of Linear Transformations from R n to R m MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Topic Overview Relationship between the properties of a matrix transformation
More informationDiscrete Distributions
Chapter 2 Discrete Distributions 2.1 Random Variables of the Discrete Type An outcome space S is difficult to study if the elements of S are not numbers. However, we can associate each element/outcome
More informationMATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation)
MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation) Last modified: March 7, 2009 Reference: PRP, Sections 3.6 and 3.7. 1. Tail-Sum Theorem
More informationExercises with solutions (Set D)
Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where
More information