STAT 6385 Survey of Nonparametric Statistics. Order Statistics, EDF and Censoring
|
|
- Austin Sullivan
- 6 years ago
- Views:
Transcription
1 STAT 6385 Survey of Nonparametric Statistics Order Statistics, EDF and Censoring
2 Quantile Function A quantile (or a percentile) of a distribution is that value of X such that a specific percentage of the probability is at or below it. A quantile divides the area under the pdf into two parts of specific amounts. The p-th quantile (or the 100p-th percentile) is that value of the random variable X, say X p, such that 100p% if the values of X in the population are less than or equal to X p, for any positive fraction p (0 < p < 1). If X is continuous, X p is a parameter that satisfies Pr(X X p ) = p, or, in terms of cdf, F X (X p ) = p. 1
3 Quantile Function Moreover, if F X is strictly increasing, the p-th quantile is the unique solution to the equation X p = F 1 X (p) = Q X(p) say, for a given p. The inverse of the cdf Q X (p), 0 < p < 1, is called the quantile function of the random variable X. Since the cdf may not be increasing for all values, we define the p-th quantile Q X (p) as the smallest X value at which the cdf is at least equal to p, or Q X (p) = F 1 X (p) = inf{x : F X(x) p},0 < p < 1. This definition gives a unique value for the quantile Q X (p) even when F X is flat or is a step function (X is discrete). 2
4 Quantile Function The cdf and the quantile function provide similar information regarding the distribution, however, there are situations where one is more natural than the other. Formulas for the moments of X can be expressed in terms of the quantile function: E(X) = 1 0 Q X (p)dp and E(X 2 ) = 1 0 Q X (p) 2 dp Let f X (x) = d dx F X(x) be the pdf of X, the first and the second derivatives of the quantile function Q X (p) are d dp Q X(p) = 1 f X [Q X (p)] d2 and dp 2Q X(p) = f X [Q X(p)] {f X [Q X (p)]} 3 3
5 Empirical Distribution Function For a random sample (X 1, X 2,..., X n ) with the common cdf F(x). the empirical distribution function (EDF) is defined as ˆF n (x) = the proportion of sample values less than or equal to the specified value x = 1 n 1 {Xi x}, n i=1 where 1 {A} is an indicator function 1 {A} = { 1 if A is true, 0 otherwise. The EDF can be defined in terms of the order statistic of a sample as ˆF n (x) = { 0 if x < X1:n, i/n if X i:n < x < X i+1:n, 1 if x X n:n. Example: n = 5, (x 1, x 2,..., x 5 ) = (7,4,8,12,10) 4
6 Empirical Distribution Function ˆF n (x) is a jump (or a step) function, with jumps occurring at the (distinct) ordered sample values, where the height of each jump is equal to the reciprocal of the sample size. When there are ties, the EDF is still a step function but it jumps only at the distinct ordered sample vaules X j:n and the height of the jump is equal to k/n, where k is the number of values tied at X j:n. Theorem 1.1. For any fixed real constant x, the random variable nˆf n (x) has a binomial distribution with parameter n and F(x). 5
7 Empirical Distribution Function Corollary 1.2. The mean and variance of the EDF ˆF n (x) are E[ˆF n (x)] = var[ˆf n (x)] = ˆF n (x) is an unbiased estimator of F(x) and the variance of ˆF n (x) 0 as n. By Chebyshev s inequality, we can show that ˆF n (x) is a consistent estimator of F(x). In other words, ˆF n (x) converges to F(x) in probability. Corollary 1.3. E[(nˆF n (x))(nˆf n (y))] = nf(x)+ n(n 1)F(x)F(y), for x < y. 6
8 Empirical Distribution Function Theorem 1.4. (Glivenko-Cantelli Theorem) ˆF n (x) converges uniformly to F(x) with probability 1, i.e., Pr [ lim n sup ˆF n (x) F(x) = 0 <x< ] = 1. Theorem 1.5. As n, the limiting probability distribution of the standardized ˆF n (x) is standard normal, i.e., lim n Pr n[ˆf n (x) F(x)] F(x)[1 F(x)] t = Φ(t), where Φ( ) is the cdf of standard normal distribution. 7
9 Order Statistics Suppose that X 1, X 2,..., X n are n independent variates from a population with continuous cumulative distribution function (cdf) F(x) and probability density function (pdf) f(x). The probability that any two or more of these random variables have equal magnitudes is zero. There exists an unique ordered arrangement within the sample. X 1:n < X 2:n < < X n:n denote the order statistics obtained by arranging the n random variates in an increasing order of magnitude. X r:n, r = 1,2,...,n, is called the r-th order statistic. 8
10 Probability-Integral Transformation One key reason why the order statistic are so important in nonparametric statistics is that for any order statistics X r:n from a continuous cdf F, the transformed random variable U r:n = F(X r:n ) has the same distribution as that of the r-th order statistic from the continuous uniform distribution on (0,1), as long as F is continuous. In this sense, F(X r:n ) may be viewed as distribution-free. This important property of continuous order statistics is called the probability-integral transformation. Theorem. Let X be a random variable with cdf F X. If F X is continuous, the random variable Y produced by the transformation Y = F X (X) has the continuous uniform probability distribution over the interval (0, 1). 9
11 Distribution of Single Order Statistic The cdf of the largest (n-th) order statistic X n:n is given by F n:n (x) = Likewise, we have the cdf of the smallest (first) order statistic X 1:n F 1:n (x) = These are special cases of the general results for the cdf of the r-th order statistic F r:n (x) = 10
12 Distribution of Single Order Statistic The joint density function of all n order statistics is f X1:n,,X n:n (x 1:n,, x n:n ) = Upon integrating out (n 1) variables in the joint density function, we obtain the marginal density function of the r-th order statistic. For instance, for the largest order statistic f n:n (x) = The smallest order statistic f 1:n (x) = The r-th order statistic f r:n (x) = 11
13 Joint Densities and Markovian Property Use the above technique, upon integrating out appropriate variables, we obtain the joint density functions of (X 1:n, X 2:n,, X i:n ) and (X 1:n,, X i:n, X j:n ) for 1 i < j n, as and f X1:n,,X i:n (x 1:n,, x i:n ) { i } = n! f(x l:n ) {1 F(x i:n )} n i, (n i)! l=1 x 1:n < < x i:n, f X1:n,,X i:n,x j:n (x 1:n,, x i:n, x j:n ) { i } n! = f(x l:n ) f(x j:n ) (j i 1)!(n j)! respectively. l=1 {F(x j:n ) F(x i:n )} j i 1 {1 F(x j:n )} n j, x 1:n < < x i:n < x j:n, 12
14 Joint Densities and Markovian Property The conditional density function of X j:n, given (X 1:n = x 1:n,, X i:n = x i:n ) for 1 i < j n, is = f Xj:n X 1:n =x 1:n,,X i:n =x i:n (x j:n x 1:n,, x i:n ) Theorem 1.5. The conditional density function of X i:n, given X j:n = x j:n, for 1 i < j n, is exactly the same as the density function of the i-th order statistic in a sample of size j 1 from the distribution F( ) truncated on the right at x j:n, i.e., from the distribution with density function f(x)/{f(x j:n )}, x < x j:n. 13
15 Joint Densities and Markovian Property Theorem 1.6. Order statistics {X i:n,1 i n} from a continuous distribution form a Markov chain. Proof. We see that the conditional density function of X j:n, given (X 1:n = x 1:n,..., X i:n = x i:n ) for 1 i < j n depends only on x i:n which implies that the order statistics form a Markov chain. Theorem 1.7. The conditional density function of X j:n, given X i:n = x i:n, for 1 i < j n, is exactly the same as the density function of the (j i)-th order statistic in a sample of size n i from the distribution F( ) truncated on the left at x i:n, i.e., from the distribution with density function f(x)/[1 F(x i:n )], x > x i:n. 14
16 Moments We have the single and product moments of order statistics as E(Xi:n l ) = n! (i 1)!(n i)! and E(X l 1 i:n X l 2 j:n ) = x l i:n {F(x i:n)} i 1 {1 F(x i:n )} n i f(x i:n )dx i:n, n! (i 1)!(j i 1)!(n j)! x i:n x l1 i:n x l 2 j:n {F(x i:n )} i 1 {F(x j:n ) F(x i:n )} j i 1 {1 F(x j:n )} n j f(x i:n )f(x j:n )dx j:n dx i:n, respectively. 15
17 Moments Upon making the probability-integral transformation u i = F(x i:n ), for 1 i n, and noting that it is order-preserving, we can rewrite the single and product moments as and E(X l i:n) = n! (i 1)!(n i)! 1 0 {F 1 (u i )} l u i 1 i (1 u i ) n i du i, E(X l 1 i:n Xl 2 j:n ) = n! (i 1)!(j i 1)!(n j)! 1 1 { F 1 (u i ) } { l1 F 1 (u j ) } l2 0 u i (u j u i ) j i 1 (1 u j ) n j du j du i. u i 1 i 16
18 Results for the Uniform Distribution For the Uniform(0,1) distribution with density function f(x) = 1 for 0 < x < 1, expressions for distributions and moments of order statistics simplify considerably. The joint density of all n order statistics simply becomes f X1:n,,X n:n (x 1:n,, x n:n ) = n!, 0 < x 1:n < < x n:n < 1. The joint density of (X 1:n,, X i:n ) becomes f X1:n,,X i:n (x 1:n,, x i:n ) = n! (n i)! (1 x i:n) n i, 0 < x 1:n < < x i:n < 1. The marginal density function of X i:n (for 1 i n) becomes f Xi:n (x i:n ) = n! (i 1)!(n i)! xi 1 i:n (1 x i:n) n i, 0 < x i:n < 1, 17
19 Results for the Uniform Distribution This simply reveals that X i:n in this case is distributed as Beta(i,n-i+1) We immediately obtain E(Xi:n) k = from which we get E(X i:n ) = and V ar(x i:n ) =. The joint density function of X i:n and X j:n (for 1 i < j n) becomes = f Xi:n,X j:n (x i:n, x j:n ) n! (i 1)!(j i 1)!(n j)! xi 1 i:n (x j:n x i:n ) j i 1 (1 x j:n ) n j, 0 < x i:n < x j:n < 1. We can show that i(j + 1) E(X i:n X j:n ) = (n + 1)(n + 2) i(n j + 1) and Cov(X i:n, X j:n ) = (n + 1) 2 (n + 2). 18
20 Censoring Censoring restricts our ability to observe the random variable. Special source of difficulty in statistical analysis. Arise when subjects are not observed for full duration of time to failure incomplete information May be due to loss, early termination, death due to other causes, etc. 19
21 Left, Right, Doubly Censored Data Right-censoring: A survival time is said to be right-censored if the event of interest occurs after the subject is observed in the study. Left-censoring: A survival time is said to be left-censored if the event of interest occurs before the subject is observed in the study. 20
22 Left, Right, Doubly Censored Data Example for Left-censoring: In early childhood learning centers, interest often focus upon testing children to determine when a child learns to accomplish certain specified tasks. The age at which a child learns the task would be considered the time-to-event. Often, some children can already perform the task when they start in the study. Such event times are considered left-censored. Doubly censoring: In a study, it may happen that some survival times are left-censored and some are right-censored. In this case we say that the sampling scheme of the study is doubly-censored. 21
23 Type-I Censoring Type-I censoring Consider an experiment which will be stopped at a pre-specified time c. All observations survive after c cannot be observed. 6 O BS 5 : censored obs. 4 : observed failure N O Time to failure c 22
24 Type-I Censoring Only observations smaller than c can be observed. It is also called time-censoring. The observations greater than c are called Type-I right censored. The number of observations being observed is a random variable. The experimental time is fixed. 23
25 Type-I Censoring Generalized Type-I censoring Sometimes the subjects entered into the study at different time point and the study will be terminated at a pre-specified time c. 6 O BS 5 : censored obs. N O : observed failure : entry time 1 Calendar Time c 24
26 Type-I Censoring Of course we can rescale all the starting time to 0 and we have the following representation: 6 O BS 5 : censored obs. N O : observed failure 1 Time to failure 25
27 Type-II Censoring Type-II Censoring One of the question in Type-I censoring is the determination of the censoring time c. If c is large, the expense of the experiment is large. If c is small, it might turn out only small portion of sample can be observed. To avoid these situations, one might terminate the experiment after the first r failures have been observed. This is also called item censoring. 26
28 Type-II Censoring Type-II Censoring Start T (1) T (2) T (r 1) censored T (r) n r End The observations are T (1) T (2) T (r). The number of observations being observed is fixed. The experimental time is a random variable. 27
29 Interval Censoring Interval Censoring In some situations the lifetime is known to occur only within an interval. A survival time is said to be interval-censored if the event of interest is only known to have occurred in a given time interval. Example: The patients in a clinical trial have periodic follow-up b 1 b 2 b k b k+1 event occurs Under this situation we only know that t i [b k, b k+1 ) 28
30 Censoring Time as random variable Censoring time as a random variable: Example: In a cancer research study, some patients might drop out from the study for some reasons (pass away for reason(s) other than cancer). Therefore, the censoring time can be a random variable. Let C i denote the censoring time for the i-th subject. Observed data = min(t i, C i ) 29
31 Censoring Time as random variable Censoring time as a random variable: Important assumption: C i and T i are independent random variables, i.e., the reason for observing a censored observation is completely unrelated to the disease process of interest. For example, the subject that moved to another city, didn t move in search for better patient care. The survival data are usually coded as (X i, δ i ), i = 1,2,..., n; where δ i = 1 if the i-th observation is uncensored, and 0 otherwise. 30
Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More informationProbability and Probability Distributions. Dr. Mohammed Alahmed
Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about
More informationMath 151. Rumbos Spring Solutions to Review Problems for Exam 3
Math 151. Rumbos Spring 2014 1 Solutions to Review Problems for Exam 3 1. Suppose that a book with n pages contains on average λ misprints per page. What is the probability that there will be at least
More informationQuick Tour of Basic Probability Theory and Linear Algebra
Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions
More informationHomework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples
Homework #3 36-754, Spring 27 Due 14 May 27 1 Convergence of the empirical CDF, uniform samples In this problem and the next, X i are IID samples on the real line, with cumulative distribution function
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationChapter 3. Chapter 3 sections
sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional Distributions 3.7 Multivariate Distributions
More informationSummary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016
8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying
More informationSection 27. The Central Limit Theorem. Po-Ning Chen, Professor. Institute of Communications Engineering. National Chiao Tung University
Section 27 The Central Limit Theorem Po-Ning Chen, Professor Institute of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 3000, R.O.C. Identically distributed summands 27- Central
More informationNotes largely based on Statistical Methods for Reliability Data by W.Q. Meeker and L. A. Escobar, Wiley, 1998 and on their class notes.
Unit 2: Models, Censoring, and Likelihood for Failure-Time Data Notes largely based on Statistical Methods for Reliability Data by W.Q. Meeker and L. A. Escobar, Wiley, 1998 and on their class notes. Ramón
More informationConditional independence of blocked ordered data
Conditional independence of blocked ordered data G. Iliopoulos 1 and N. Balakrishnan 2 Abstract In this paper, we prove that blocks of ordered data formed by some conditioning events are mutually independent.
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More informationOrder Statistics and Distributions
Order Statistics and Distributions 1 Some Preliminary Comments and Ideas In this section we consider a random sample X 1, X 2,..., X n common continuous distribution function F and probability density
More informationThis exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.
Probability and Statistics FS 2017 Session Exam 22.08.2017 Time Limit: 180 Minutes Name: Student ID: This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided
More information5.6 The Normal Distributions
STAT 41 Lecture Notes 13 5.6 The Normal Distributions Definition 5.6.1. A (continuous) random variable X has a normal distribution with mean µ R and variance < R if the p.d.f. of X is f(x µ, ) ( π ) 1/
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationLecture 4a: Continuous-Time Markov Chain Models
Lecture 4a: Continuous-Time Markov Chain Models Continuous-time Markov chains are stochastic processes whose time is continuous, t [0, ), but the random variables are discrete. Prominent examples of continuous-time
More informationNonparametric Model Construction
Nonparametric Model Construction Chapters 4 and 12 Stat 477 - Loss Models Chapters 4 and 12 (Stat 477) Nonparametric Model Construction Brian Hartman - BYU 1 / 28 Types of data Types of data For non-life
More informationMATH4427 Notebook 4 Fall Semester 2017/2018
MATH4427 Notebook 4 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 4 MATH4427 Notebook 4 3 4.1 K th Order Statistics and Their
More informationExact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring
Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring A. Ganguly, S. Mitra, D. Samanta, D. Kundu,2 Abstract Epstein [9] introduced the Type-I hybrid censoring scheme
More informationProblem 1. Problem 2. Problem 3. Problem 4
Problem Let A be the event that the fungus is present, and B the event that the staph-bacteria is present. We have P A = 4, P B = 9, P B A =. We wish to find P AB, to do this we use the multiplication
More informationLecture 21: Convergence of transformations and generating a random variable
Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous
More informationChapter 2. Discrete Distributions
Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation
More informationWhat to do today (Nov 22, 2018)?
What to do today (Nov 22, 2018)? Part 1. Introduction and Review (Chp 1-5) Part 2. Basic Statistical Inference (Chp 6-9) Part 3. Important Topics in Statistics (Chp 10-13) Part 4. Further Topics (Selected
More informationLecture 3. Truncation, length-bias and prevalence sampling
Lecture 3. Truncation, length-bias and prevalence sampling 3.1 Prevalent sampling Statistical techniques for truncated data have been integrated into survival analysis in last two decades. Truncation in
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More informationSTAT Section 2.1: Basic Inference. Basic Definitions
STAT 518 --- Section 2.1: Basic Inference Basic Definitions Population: The collection of all the individuals of interest. This collection may be or even. Sample: A collection of elements of the population.
More informationLecture 2: CDF and EDF
STAT 425: Introduction to Nonparametric Statistics Winter 2018 Instructor: Yen-Chi Chen Lecture 2: CDF and EDF 2.1 CDF: Cumulative Distribution Function For a random variable X, its CDF F () contains all
More informationSampling Distributions
In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of random sample. For example,
More informationORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES
ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES 1. Order statistics Let X 1,...,X n be n real-valued observations. One can always arrangetheminordertogettheorder statisticsx (1) X (2) X (n). SinceX (k)
More informationHomework 10 (due December 2, 2009)
Homework (due December, 9) Problem. Let X and Y be independent binomial random variables with parameters (n, p) and (n, p) respectively. Prove that X + Y is a binomial random variable with parameters (n
More informationST745: Survival Analysis: Nonparametric methods
ST745: Survival Analysis: Nonparametric methods Eric B. Laber Department of Statistics, North Carolina State University February 5, 2015 The KM estimator is used ubiquitously in medical studies to estimate
More informationExercises. (a) Prove that m(t) =
Exercises 1. Lack of memory. Verify that the exponential distribution has the lack of memory property, that is, if T is exponentially distributed with parameter λ > then so is T t given that T > t for
More informationA Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.
A Probability Primer A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes. Are you holding all the cards?? Random Events A random event, E,
More informationNormal Random Variables and Probability
Normal Random Variables and Probability An Undergraduate Introduction to Financial Mathematics J. Robert Buchanan 2015 Discrete vs. Continuous Random Variables Think about the probability of selecting
More informationLecture 16: Sample quantiles and their asymptotic properties
Lecture 16: Sample quantiles and their asymptotic properties Estimation of quantiles (percentiles Suppose that X 1,...,X n are i.i.d. random variables from an unknown nonparametric F For p (0,1, G 1 (p
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationContinuous r.v. s: cdf s, Expected Values
Continuous r.v. s: cdf s, Expected Values Engineering Statistics Section 4.2 Josh Engwer TTU 29 February 2016 Josh Engwer (TTU) Continuous r.v. s: cdf s, Expected Values 29 February 2016 1 / 17 PART I
More information1.1 Review of Probability Theory
1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,
More informationLecture 11. Probability Theory: an Overveiw
Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the
More informationKernel Density Estimation
Kernel Density Estimation Univariate Density Estimation Suppose tat we ave a random sample of data X 1,..., X n from an unknown continuous distribution wit probability density function (pdf) f(x) and cumulative
More informationST 371 (IX): Theories of Sampling Distributions
ST 371 (IX): Theories of Sampling Distributions 1 Sample, Population, Parameter and Statistic The major use of inferential statistics is to use information from a sample to infer characteristics about
More informationMath 151. Rumbos Fall Solutions to Review Problems for Final Exam
Math 5. Rumbos Fall 23 Solutions to Review Problems for Final Exam. Three cards are in a bag. One card is red on both sides. Another card is white on both sides. The third card in red on one side and white
More informationSurvival Distributions, Hazard Functions, Cumulative Hazards
BIO 244: Unit 1 Survival Distributions, Hazard Functions, Cumulative Hazards 1.1 Definitions: The goals of this unit are to introduce notation, discuss ways of probabilistically describing the distribution
More information4.1 The Expectation of a Random Variable
STAT 42 Lecture Notes 93 4. The Expectation of a Random Variable This chapter begins the discussion of properties of random variables. The focus of this chapter is on expectations of random variables.
More informationAnalysis of Time-to-Event Data: Chapter 2 - Nonparametric estimation of functions of survival time
Analysis of Time-to-Event Data: Chapter 2 - Nonparametric estimation of functions of survival time Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term
More informationECON 5350 Class Notes Review of Probability and Distribution Theory
ECON 535 Class Notes Review of Probability and Distribution Theory 1 Random Variables Definition. Let c represent an element of the sample space C of a random eperiment, c C. A random variable is a one-to-one
More informationIt can be shown that if X 1 ;X 2 ;:::;X n are independent r.v. s with
Example: Alternative calculation of mean and variance of binomial distribution A r.v. X has the Bernoulli distribution if it takes the values 1 ( success ) or 0 ( failure ) with probabilities p and (1
More informationExpectation, inequalities and laws of large numbers
Chapter 3 Expectation, inequalities and laws of large numbers 3. Expectation and Variance Indicator random variable Let us suppose that the event A partitions the sample space S, i.e. A A S. The indicator
More informationCensoring mechanisms
Censoring mechanisms Patrick Breheny September 3 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/23 Fixed vs. random censoring In the previous lecture, we derived the contribution to the likelihood
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More informationA Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints
Noname manuscript No. (will be inserted by the editor) A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Mai Zhou Yifan Yang Received: date / Accepted: date Abstract In this note
More informationStatistics for scientists and engineers
Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3
More information3 Continuous Random Variables
Jinguo Lian Math437 Notes January 15, 016 3 Continuous Random Variables Remember that discrete random variables can take only a countable number of possible values. On the other hand, a continuous random
More information1. Point Estimators, Review
AMS571 Prof. Wei Zhu 1. Point Estimators, Review Example 1. Let be a random sample from. Please find a good point estimator for Solutions. There are the typical estimators for and. Both are unbiased estimators.
More informationSTA 584 Supplementary Examples (not to be graded) Fall, 2003
Page 1 of 8 Central Michigan University Department of Mathematics STA 584 Supplementary Examples (not to be graded) Fall, 003 1. (a) If A and B are independent events, P(A) =.40 and P(B) =.70, find (i)
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional
More informationSurvival Analysis. Lu Tian and Richard Olshen Stanford University
1 Survival Analysis Lu Tian and Richard Olshen Stanford University 2 Survival Time/ Failure Time/Event Time We will introduce various statistical methods for analyzing survival outcomes What is the survival
More informationDepartment of Statistical Science FIRST YEAR EXAM - SPRING 2017
Department of Statistical Science Duke University FIRST YEAR EXAM - SPRING 017 Monday May 8th 017, 9:00 AM 1:00 PM NOTES: PLEASE READ CAREFULLY BEFORE BEGINNING EXAM! 1. Do not write solutions on the exam;
More informationFinal Exam # 3. Sta 230: Probability. December 16, 2012
Final Exam # 3 Sta 230: Probability December 16, 2012 This is a closed-book exam so do not refer to your notes, the text, or any other books (please put them on the floor). You may use the extra sheets
More informationParameters Estimation for a Linear Exponential Distribution Based on Grouped Data
International Mathematical Forum, 3, 2008, no. 33, 1643-1654 Parameters Estimation for a Linear Exponential Distribution Based on Grouped Data A. Al-khedhairi Department of Statistics and O.R. Faculty
More informationMultiple Random Variables
Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An
More informationTransformations and Expectations
Transformations and Expectations 1 Distributions of Functions of a Random Variable If is a random variable with cdf F (x), then any function of, say g(), is also a random variable. Sine Y = g() is a function
More informationLecture 11: Probability, Order Statistics and Sampling
5-75: Graduate Algorithms February, 7 Lecture : Probability, Order tatistics and ampling Lecturer: David Whitmer cribes: Ilai Deutel, C.J. Argue Exponential Distributions Definition.. Given sample space
More informationSUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)
SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems
More informationRandom Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.
Random Variables Amappingthattransformstheeventstotherealline. Example 1. Toss a fair coin. Define a random variable X where X is 1 if head appears and X is if tail appears. P (X =)=1/2 P (X =1)=1/2 Example
More information2 (Statistics) Random variables
2 (Statistics) Random variables References: DeGroot and Schervish, chapters 3, 4 and 5; Stirzaker, chapters 4, 5 and 6 We will now study the main tools use for modeling experiments with unknown outcomes
More informationThis exam contains 6 questions. The questions are of equal weight. Print your name at the top of this page in the upper right hand corner.
GROUND RULES: This exam contains 6 questions. The questions are of equal weight. Print your name at the top of this page in the upper right hand corner. This exam is closed book and closed notes. Show
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More information16.4. Power Series. Introduction. Prerequisites. Learning Outcomes
Power Series 6.4 Introduction In this Section we consider power series. These are examples of infinite series where each term contains a variable, x, raised to a positive integer power. We use the ratio
More informationStatistical Methods in Particle Physics
Statistical Methods in Particle Physics. Probability Distributions Prof. Dr. Klaus Reygers (lectures) Dr. Sebastian Neubert (tutorials) Heidelberg University WS 07/8 Gaussian g(x; µ, )= p exp (x µ) https://en.wikipedia.org/wiki/normal_distribution
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections Chapter 3 - continued 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions
More information1 Glivenko-Cantelli type theorems
STA79 Lecture Spring Semester Glivenko-Cantelli type theorems Given i.i.d. observations X,..., X n with unknown distribution function F (t, consider the empirical (sample CDF ˆF n (t = I [Xi t]. n Then
More informationStatistics 100A Homework 5 Solutions
Chapter 5 Statistics 1A Homework 5 Solutions Ryan Rosario 1. Let X be a random variable with probability density function a What is the value of c? fx { c1 x 1 < x < 1 otherwise We know that for fx to
More informationNonparametric Function Estimation with Infinite-Order Kernels
Nonparametric Function Estimation with Infinite-Order Kernels Arthur Berg Department of Statistics, University of Florida March 15, 2008 Kernel Density Estimation (IID Case) Let X 1,..., X n iid density
More informationReview. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda
Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with
More informationHybrid Censoring; An Introduction 2
Hybrid Censoring; An Introduction 2 Debasis Kundu Department of Mathematics & Statistics Indian Institute of Technology Kanpur 23-rd November, 2010 2 This is a joint work with N. Balakrishnan Debasis Kundu
More informationEXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS
EXAM Exam # Math 3342 Summer II, 2 July 2, 2 ANSWERS i pts. Problem. Consider the following data: 7, 8, 9, 2,, 7, 2, 3. Find the first quartile, the median, and the third quartile. Make a box and whisker
More informationThe Type I Generalized Half Logistic Distribution Downloaded from jirss.irstat.ir at 11: on Monday August 20th 2018
JIRSS (2014) Vol. 13, No. 1, pp 69-82 The Type I Generalized Half Logistic Distribution A. K. Olapade Department of Mathematics, Obafemi Awolowo University, Ile-Ife, Nigeria. Abstract. In this paper, we
More informationStatistics 135 Fall 2008 Final Exam
Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing
More informationSTAT 830 Non-parametric Inference Basics
STAT 830 Non-parametric Inference Basics Richard Lockhart Simon Fraser University STAT 801=830 Fall 2012 Richard Lockhart (Simon Fraser University)STAT 830 Non-parametric Inference Basics STAT 801=830
More informationStatistics Ph.D. Qualifying Exam: Part I October 18, 2003
Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer
More informationDistribution Fitting (Censored Data)
Distribution Fitting (Censored Data) Summary... 1 Data Input... 2 Analysis Summary... 3 Analysis Options... 4 Goodness-of-Fit Tests... 6 Frequency Histogram... 8 Comparison of Alternative Distributions...
More informationMath Spring Practice for the Second Exam.
Math 4 - Spring 27 - Practice for the Second Exam.. Let X be a random variable and let F X be the distribution function of X: t < t 2 t < 4 F X (t) : + t t < 2 2 2 2 t < 4 t. Find P(X ), P(X ), P(X 2),
More informationExam P Review Sheet. for a > 0. ln(a) i=0 ari = a. (1 r) 2. (Note that the A i s form a partition)
Exam P Review Sheet log b (b x ) = x log b (y k ) = k log b (y) log b (y) = ln(y) ln(b) log b (yz) = log b (y) + log b (z) log b (y/z) = log b (y) log b (z) ln(e x ) = x e ln(y) = y for y > 0. d dx ax
More informationCS145: Probability & Computing
CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,
More informationIAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES
IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES VARIABLE Studying the behavior of random variables, and more importantly functions of random variables is essential for both the
More informationReview of Probability Theory
Review of Probability Theory Arian Maleki and Tom Do Stanford University Probability theory is the study of uncertainty Through this class, we will be relying on concepts from probability theory for deriving
More informationHW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100]
HW7 Solutions. 5 pts.) James Bond James Bond, my favorite hero, has again jumped off a plane. The plane is traveling from from base A to base B, distance km apart. Now suppose the plane takes off from
More informationProbability and Measure
Chapter 4 Probability and Measure 4.1 Introduction In this chapter we will examine probability theory from the measure theoretic perspective. The realisation that measure theory is the foundation of probability
More informationStatistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University
Statistics for Economists Lectures 6 & 7 Asrat Temesgen Stockholm University 1 Chapter 4- Bivariate Distributions 41 Distributions of two random variables Definition 41-1: Let X and Y be two random variables
More informationSTAT 302 Introduction to Probability Learning Outcomes. Textbook: A First Course in Probability by Sheldon Ross, 8 th ed.
STAT 302 Introduction to Probability Learning Outcomes Textbook: A First Course in Probability by Sheldon Ross, 8 th ed. Chapter 1: Combinatorial Analysis Demonstrate the ability to solve combinatorial
More informationUNIVERSITY OF CALIFORNIA, SAN DIEGO
UNIVERSITY OF CALIFORNIA, SAN DIEGO Estimation of the primary hazard ratio in the presence of a secondary covariate with non-proportional hazards An undergraduate honors thesis submitted to the Department
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More information5 Introduction to the Theory of Order Statistics and Rank Statistics
5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order
More informationLecture 17: The Exponential and Some Related Distributions
Lecture 7: The Exponential and Some Related Distributions. Definition Definition: A continuous random variable X is said to have the exponential distribution with parameter if the density of X is e x if
More informationStatistics 427: Sample Final Exam
Statistics 427: Sample Final Exam Instructions: The following sample exam was given several quarters ago in Stat 427. The same topics were covered in the class that year. This sample exam is meant to be
More information