STAT 6385 Survey of Nonparametric Statistics. Order Statistics, EDF and Censoring

Size: px
Start display at page:

Download "STAT 6385 Survey of Nonparametric Statistics. Order Statistics, EDF and Censoring"

Transcription

1 STAT 6385 Survey of Nonparametric Statistics Order Statistics, EDF and Censoring

2 Quantile Function A quantile (or a percentile) of a distribution is that value of X such that a specific percentage of the probability is at or below it. A quantile divides the area under the pdf into two parts of specific amounts. The p-th quantile (or the 100p-th percentile) is that value of the random variable X, say X p, such that 100p% if the values of X in the population are less than or equal to X p, for any positive fraction p (0 < p < 1). If X is continuous, X p is a parameter that satisfies Pr(X X p ) = p, or, in terms of cdf, F X (X p ) = p. 1

3 Quantile Function Moreover, if F X is strictly increasing, the p-th quantile is the unique solution to the equation X p = F 1 X (p) = Q X(p) say, for a given p. The inverse of the cdf Q X (p), 0 < p < 1, is called the quantile function of the random variable X. Since the cdf may not be increasing for all values, we define the p-th quantile Q X (p) as the smallest X value at which the cdf is at least equal to p, or Q X (p) = F 1 X (p) = inf{x : F X(x) p},0 < p < 1. This definition gives a unique value for the quantile Q X (p) even when F X is flat or is a step function (X is discrete). 2

4 Quantile Function The cdf and the quantile function provide similar information regarding the distribution, however, there are situations where one is more natural than the other. Formulas for the moments of X can be expressed in terms of the quantile function: E(X) = 1 0 Q X (p)dp and E(X 2 ) = 1 0 Q X (p) 2 dp Let f X (x) = d dx F X(x) be the pdf of X, the first and the second derivatives of the quantile function Q X (p) are d dp Q X(p) = 1 f X [Q X (p)] d2 and dp 2Q X(p) = f X [Q X(p)] {f X [Q X (p)]} 3 3

5 Empirical Distribution Function For a random sample (X 1, X 2,..., X n ) with the common cdf F(x). the empirical distribution function (EDF) is defined as ˆF n (x) = the proportion of sample values less than or equal to the specified value x = 1 n 1 {Xi x}, n i=1 where 1 {A} is an indicator function 1 {A} = { 1 if A is true, 0 otherwise. The EDF can be defined in terms of the order statistic of a sample as ˆF n (x) = { 0 if x < X1:n, i/n if X i:n < x < X i+1:n, 1 if x X n:n. Example: n = 5, (x 1, x 2,..., x 5 ) = (7,4,8,12,10) 4

6 Empirical Distribution Function ˆF n (x) is a jump (or a step) function, with jumps occurring at the (distinct) ordered sample values, where the height of each jump is equal to the reciprocal of the sample size. When there are ties, the EDF is still a step function but it jumps only at the distinct ordered sample vaules X j:n and the height of the jump is equal to k/n, where k is the number of values tied at X j:n. Theorem 1.1. For any fixed real constant x, the random variable nˆf n (x) has a binomial distribution with parameter n and F(x). 5

7 Empirical Distribution Function Corollary 1.2. The mean and variance of the EDF ˆF n (x) are E[ˆF n (x)] = var[ˆf n (x)] = ˆF n (x) is an unbiased estimator of F(x) and the variance of ˆF n (x) 0 as n. By Chebyshev s inequality, we can show that ˆF n (x) is a consistent estimator of F(x). In other words, ˆF n (x) converges to F(x) in probability. Corollary 1.3. E[(nˆF n (x))(nˆf n (y))] = nf(x)+ n(n 1)F(x)F(y), for x < y. 6

8 Empirical Distribution Function Theorem 1.4. (Glivenko-Cantelli Theorem) ˆF n (x) converges uniformly to F(x) with probability 1, i.e., Pr [ lim n sup ˆF n (x) F(x) = 0 <x< ] = 1. Theorem 1.5. As n, the limiting probability distribution of the standardized ˆF n (x) is standard normal, i.e., lim n Pr n[ˆf n (x) F(x)] F(x)[1 F(x)] t = Φ(t), where Φ( ) is the cdf of standard normal distribution. 7

9 Order Statistics Suppose that X 1, X 2,..., X n are n independent variates from a population with continuous cumulative distribution function (cdf) F(x) and probability density function (pdf) f(x). The probability that any two or more of these random variables have equal magnitudes is zero. There exists an unique ordered arrangement within the sample. X 1:n < X 2:n < < X n:n denote the order statistics obtained by arranging the n random variates in an increasing order of magnitude. X r:n, r = 1,2,...,n, is called the r-th order statistic. 8

10 Probability-Integral Transformation One key reason why the order statistic are so important in nonparametric statistics is that for any order statistics X r:n from a continuous cdf F, the transformed random variable U r:n = F(X r:n ) has the same distribution as that of the r-th order statistic from the continuous uniform distribution on (0,1), as long as F is continuous. In this sense, F(X r:n ) may be viewed as distribution-free. This important property of continuous order statistics is called the probability-integral transformation. Theorem. Let X be a random variable with cdf F X. If F X is continuous, the random variable Y produced by the transformation Y = F X (X) has the continuous uniform probability distribution over the interval (0, 1). 9

11 Distribution of Single Order Statistic The cdf of the largest (n-th) order statistic X n:n is given by F n:n (x) = Likewise, we have the cdf of the smallest (first) order statistic X 1:n F 1:n (x) = These are special cases of the general results for the cdf of the r-th order statistic F r:n (x) = 10

12 Distribution of Single Order Statistic The joint density function of all n order statistics is f X1:n,,X n:n (x 1:n,, x n:n ) = Upon integrating out (n 1) variables in the joint density function, we obtain the marginal density function of the r-th order statistic. For instance, for the largest order statistic f n:n (x) = The smallest order statistic f 1:n (x) = The r-th order statistic f r:n (x) = 11

13 Joint Densities and Markovian Property Use the above technique, upon integrating out appropriate variables, we obtain the joint density functions of (X 1:n, X 2:n,, X i:n ) and (X 1:n,, X i:n, X j:n ) for 1 i < j n, as and f X1:n,,X i:n (x 1:n,, x i:n ) { i } = n! f(x l:n ) {1 F(x i:n )} n i, (n i)! l=1 x 1:n < < x i:n, f X1:n,,X i:n,x j:n (x 1:n,, x i:n, x j:n ) { i } n! = f(x l:n ) f(x j:n ) (j i 1)!(n j)! respectively. l=1 {F(x j:n ) F(x i:n )} j i 1 {1 F(x j:n )} n j, x 1:n < < x i:n < x j:n, 12

14 Joint Densities and Markovian Property The conditional density function of X j:n, given (X 1:n = x 1:n,, X i:n = x i:n ) for 1 i < j n, is = f Xj:n X 1:n =x 1:n,,X i:n =x i:n (x j:n x 1:n,, x i:n ) Theorem 1.5. The conditional density function of X i:n, given X j:n = x j:n, for 1 i < j n, is exactly the same as the density function of the i-th order statistic in a sample of size j 1 from the distribution F( ) truncated on the right at x j:n, i.e., from the distribution with density function f(x)/{f(x j:n )}, x < x j:n. 13

15 Joint Densities and Markovian Property Theorem 1.6. Order statistics {X i:n,1 i n} from a continuous distribution form a Markov chain. Proof. We see that the conditional density function of X j:n, given (X 1:n = x 1:n,..., X i:n = x i:n ) for 1 i < j n depends only on x i:n which implies that the order statistics form a Markov chain. Theorem 1.7. The conditional density function of X j:n, given X i:n = x i:n, for 1 i < j n, is exactly the same as the density function of the (j i)-th order statistic in a sample of size n i from the distribution F( ) truncated on the left at x i:n, i.e., from the distribution with density function f(x)/[1 F(x i:n )], x > x i:n. 14

16 Moments We have the single and product moments of order statistics as E(Xi:n l ) = n! (i 1)!(n i)! and E(X l 1 i:n X l 2 j:n ) = x l i:n {F(x i:n)} i 1 {1 F(x i:n )} n i f(x i:n )dx i:n, n! (i 1)!(j i 1)!(n j)! x i:n x l1 i:n x l 2 j:n {F(x i:n )} i 1 {F(x j:n ) F(x i:n )} j i 1 {1 F(x j:n )} n j f(x i:n )f(x j:n )dx j:n dx i:n, respectively. 15

17 Moments Upon making the probability-integral transformation u i = F(x i:n ), for 1 i n, and noting that it is order-preserving, we can rewrite the single and product moments as and E(X l i:n) = n! (i 1)!(n i)! 1 0 {F 1 (u i )} l u i 1 i (1 u i ) n i du i, E(X l 1 i:n Xl 2 j:n ) = n! (i 1)!(j i 1)!(n j)! 1 1 { F 1 (u i ) } { l1 F 1 (u j ) } l2 0 u i (u j u i ) j i 1 (1 u j ) n j du j du i. u i 1 i 16

18 Results for the Uniform Distribution For the Uniform(0,1) distribution with density function f(x) = 1 for 0 < x < 1, expressions for distributions and moments of order statistics simplify considerably. The joint density of all n order statistics simply becomes f X1:n,,X n:n (x 1:n,, x n:n ) = n!, 0 < x 1:n < < x n:n < 1. The joint density of (X 1:n,, X i:n ) becomes f X1:n,,X i:n (x 1:n,, x i:n ) = n! (n i)! (1 x i:n) n i, 0 < x 1:n < < x i:n < 1. The marginal density function of X i:n (for 1 i n) becomes f Xi:n (x i:n ) = n! (i 1)!(n i)! xi 1 i:n (1 x i:n) n i, 0 < x i:n < 1, 17

19 Results for the Uniform Distribution This simply reveals that X i:n in this case is distributed as Beta(i,n-i+1) We immediately obtain E(Xi:n) k = from which we get E(X i:n ) = and V ar(x i:n ) =. The joint density function of X i:n and X j:n (for 1 i < j n) becomes = f Xi:n,X j:n (x i:n, x j:n ) n! (i 1)!(j i 1)!(n j)! xi 1 i:n (x j:n x i:n ) j i 1 (1 x j:n ) n j, 0 < x i:n < x j:n < 1. We can show that i(j + 1) E(X i:n X j:n ) = (n + 1)(n + 2) i(n j + 1) and Cov(X i:n, X j:n ) = (n + 1) 2 (n + 2). 18

20 Censoring Censoring restricts our ability to observe the random variable. Special source of difficulty in statistical analysis. Arise when subjects are not observed for full duration of time to failure incomplete information May be due to loss, early termination, death due to other causes, etc. 19

21 Left, Right, Doubly Censored Data Right-censoring: A survival time is said to be right-censored if the event of interest occurs after the subject is observed in the study. Left-censoring: A survival time is said to be left-censored if the event of interest occurs before the subject is observed in the study. 20

22 Left, Right, Doubly Censored Data Example for Left-censoring: In early childhood learning centers, interest often focus upon testing children to determine when a child learns to accomplish certain specified tasks. The age at which a child learns the task would be considered the time-to-event. Often, some children can already perform the task when they start in the study. Such event times are considered left-censored. Doubly censoring: In a study, it may happen that some survival times are left-censored and some are right-censored. In this case we say that the sampling scheme of the study is doubly-censored. 21

23 Type-I Censoring Type-I censoring Consider an experiment which will be stopped at a pre-specified time c. All observations survive after c cannot be observed. 6 O BS 5 : censored obs. 4 : observed failure N O Time to failure c 22

24 Type-I Censoring Only observations smaller than c can be observed. It is also called time-censoring. The observations greater than c are called Type-I right censored. The number of observations being observed is a random variable. The experimental time is fixed. 23

25 Type-I Censoring Generalized Type-I censoring Sometimes the subjects entered into the study at different time point and the study will be terminated at a pre-specified time c. 6 O BS 5 : censored obs. N O : observed failure : entry time 1 Calendar Time c 24

26 Type-I Censoring Of course we can rescale all the starting time to 0 and we have the following representation: 6 O BS 5 : censored obs. N O : observed failure 1 Time to failure 25

27 Type-II Censoring Type-II Censoring One of the question in Type-I censoring is the determination of the censoring time c. If c is large, the expense of the experiment is large. If c is small, it might turn out only small portion of sample can be observed. To avoid these situations, one might terminate the experiment after the first r failures have been observed. This is also called item censoring. 26

28 Type-II Censoring Type-II Censoring Start T (1) T (2) T (r 1) censored T (r) n r End The observations are T (1) T (2) T (r). The number of observations being observed is fixed. The experimental time is a random variable. 27

29 Interval Censoring Interval Censoring In some situations the lifetime is known to occur only within an interval. A survival time is said to be interval-censored if the event of interest is only known to have occurred in a given time interval. Example: The patients in a clinical trial have periodic follow-up b 1 b 2 b k b k+1 event occurs Under this situation we only know that t i [b k, b k+1 ) 28

30 Censoring Time as random variable Censoring time as a random variable: Example: In a cancer research study, some patients might drop out from the study for some reasons (pass away for reason(s) other than cancer). Therefore, the censoring time can be a random variable. Let C i denote the censoring time for the i-th subject. Observed data = min(t i, C i ) 29

31 Censoring Time as random variable Censoring time as a random variable: Important assumption: C i and T i are independent random variables, i.e., the reason for observing a censored observation is completely unrelated to the disease process of interest. For example, the subject that moved to another city, didn t move in search for better patient care. The survival data are usually coded as (X i, δ i ), i = 1,2,..., n; where δ i = 1 if the i-th observation is uncensored, and 0 otherwise. 30

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique

More information

Probability and Probability Distributions. Dr. Mohammed Alahmed

Probability and Probability Distributions. Dr. Mohammed Alahmed Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about

More information

Math 151. Rumbos Spring Solutions to Review Problems for Exam 3

Math 151. Rumbos Spring Solutions to Review Problems for Exam 3 Math 151. Rumbos Spring 2014 1 Solutions to Review Problems for Exam 3 1. Suppose that a book with n pages contains on average λ misprints per page. What is the probability that there will be at least

More information

Quick Tour of Basic Probability Theory and Linear Algebra

Quick Tour of Basic Probability Theory and Linear Algebra Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions

More information

Homework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples

Homework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples Homework #3 36-754, Spring 27 Due 14 May 27 1 Convergence of the empirical CDF, uniform samples In this problem and the next, X i are IID samples on the real line, with cumulative distribution function

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Chapter 3. Chapter 3 sections

Chapter 3. Chapter 3 sections sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional Distributions 3.7 Multivariate Distributions

More information

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying

More information

Section 27. The Central Limit Theorem. Po-Ning Chen, Professor. Institute of Communications Engineering. National Chiao Tung University

Section 27. The Central Limit Theorem. Po-Ning Chen, Professor. Institute of Communications Engineering. National Chiao Tung University Section 27 The Central Limit Theorem Po-Ning Chen, Professor Institute of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 3000, R.O.C. Identically distributed summands 27- Central

More information

Notes largely based on Statistical Methods for Reliability Data by W.Q. Meeker and L. A. Escobar, Wiley, 1998 and on their class notes.

Notes largely based on Statistical Methods for Reliability Data by W.Q. Meeker and L. A. Escobar, Wiley, 1998 and on their class notes. Unit 2: Models, Censoring, and Likelihood for Failure-Time Data Notes largely based on Statistical Methods for Reliability Data by W.Q. Meeker and L. A. Escobar, Wiley, 1998 and on their class notes. Ramón

More information

Conditional independence of blocked ordered data

Conditional independence of blocked ordered data Conditional independence of blocked ordered data G. Iliopoulos 1 and N. Balakrishnan 2 Abstract In this paper, we prove that blocks of ordered data formed by some conditioning events are mutually independent.

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

Order Statistics and Distributions

Order Statistics and Distributions Order Statistics and Distributions 1 Some Preliminary Comments and Ideas In this section we consider a random sample X 1, X 2,..., X n common continuous distribution function F and probability density

More information

This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam. Probability and Statistics FS 2017 Session Exam 22.08.2017 Time Limit: 180 Minutes Name: Student ID: This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided

More information

5.6 The Normal Distributions

5.6 The Normal Distributions STAT 41 Lecture Notes 13 5.6 The Normal Distributions Definition 5.6.1. A (continuous) random variable X has a normal distribution with mean µ R and variance < R if the p.d.f. of X is f(x µ, ) ( π ) 1/

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Probability Models. 4. What is the definition of the expectation of a discrete random variable? 1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions

More information

Lecture 4a: Continuous-Time Markov Chain Models

Lecture 4a: Continuous-Time Markov Chain Models Lecture 4a: Continuous-Time Markov Chain Models Continuous-time Markov chains are stochastic processes whose time is continuous, t [0, ), but the random variables are discrete. Prominent examples of continuous-time

More information

Nonparametric Model Construction

Nonparametric Model Construction Nonparametric Model Construction Chapters 4 and 12 Stat 477 - Loss Models Chapters 4 and 12 (Stat 477) Nonparametric Model Construction Brian Hartman - BYU 1 / 28 Types of data Types of data For non-life

More information

MATH4427 Notebook 4 Fall Semester 2017/2018

MATH4427 Notebook 4 Fall Semester 2017/2018 MATH4427 Notebook 4 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 4 MATH4427 Notebook 4 3 4.1 K th Order Statistics and Their

More information

Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring

Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring A. Ganguly, S. Mitra, D. Samanta, D. Kundu,2 Abstract Epstein [9] introduced the Type-I hybrid censoring scheme

More information

Problem 1. Problem 2. Problem 3. Problem 4

Problem 1. Problem 2. Problem 3. Problem 4 Problem Let A be the event that the fungus is present, and B the event that the staph-bacteria is present. We have P A = 4, P B = 9, P B A =. We wish to find P AB, to do this we use the multiplication

More information

Lecture 21: Convergence of transformations and generating a random variable

Lecture 21: Convergence of transformations and generating a random variable Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous

More information

Chapter 2. Discrete Distributions

Chapter 2. Discrete Distributions Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation

More information

What to do today (Nov 22, 2018)?

What to do today (Nov 22, 2018)? What to do today (Nov 22, 2018)? Part 1. Introduction and Review (Chp 1-5) Part 2. Basic Statistical Inference (Chp 6-9) Part 3. Important Topics in Statistics (Chp 10-13) Part 4. Further Topics (Selected

More information

Lecture 3. Truncation, length-bias and prevalence sampling

Lecture 3. Truncation, length-bias and prevalence sampling Lecture 3. Truncation, length-bias and prevalence sampling 3.1 Prevalent sampling Statistical techniques for truncated data have been integrated into survival analysis in last two decades. Truncation in

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are

More information

STAT Section 2.1: Basic Inference. Basic Definitions

STAT Section 2.1: Basic Inference. Basic Definitions STAT 518 --- Section 2.1: Basic Inference Basic Definitions Population: The collection of all the individuals of interest. This collection may be or even. Sample: A collection of elements of the population.

More information

Lecture 2: CDF and EDF

Lecture 2: CDF and EDF STAT 425: Introduction to Nonparametric Statistics Winter 2018 Instructor: Yen-Chi Chen Lecture 2: CDF and EDF 2.1 CDF: Cumulative Distribution Function For a random variable X, its CDF F () contains all

More information

Sampling Distributions

Sampling Distributions In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of random sample. For example,

More information

ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES

ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES 1. Order statistics Let X 1,...,X n be n real-valued observations. One can always arrangetheminordertogettheorder statisticsx (1) X (2) X (n). SinceX (k)

More information

Homework 10 (due December 2, 2009)

Homework 10 (due December 2, 2009) Homework (due December, 9) Problem. Let X and Y be independent binomial random variables with parameters (n, p) and (n, p) respectively. Prove that X + Y is a binomial random variable with parameters (n

More information

ST745: Survival Analysis: Nonparametric methods

ST745: Survival Analysis: Nonparametric methods ST745: Survival Analysis: Nonparametric methods Eric B. Laber Department of Statistics, North Carolina State University February 5, 2015 The KM estimator is used ubiquitously in medical studies to estimate

More information

Exercises. (a) Prove that m(t) =

Exercises. (a) Prove that m(t) = Exercises 1. Lack of memory. Verify that the exponential distribution has the lack of memory property, that is, if T is exponentially distributed with parameter λ > then so is T t given that T > t for

More information

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes. A Probability Primer A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes. Are you holding all the cards?? Random Events A random event, E,

More information

Normal Random Variables and Probability

Normal Random Variables and Probability Normal Random Variables and Probability An Undergraduate Introduction to Financial Mathematics J. Robert Buchanan 2015 Discrete vs. Continuous Random Variables Think about the probability of selecting

More information

Lecture 16: Sample quantiles and their asymptotic properties

Lecture 16: Sample quantiles and their asymptotic properties Lecture 16: Sample quantiles and their asymptotic properties Estimation of quantiles (percentiles Suppose that X 1,...,X n are i.i.d. random variables from an unknown nonparametric F For p (0,1, G 1 (p

More information

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample

More information

Continuous r.v. s: cdf s, Expected Values

Continuous r.v. s: cdf s, Expected Values Continuous r.v. s: cdf s, Expected Values Engineering Statistics Section 4.2 Josh Engwer TTU 29 February 2016 Josh Engwer (TTU) Continuous r.v. s: cdf s, Expected Values 29 February 2016 1 / 17 PART I

More information

1.1 Review of Probability Theory

1.1 Review of Probability Theory 1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,

More information

Lecture 11. Probability Theory: an Overveiw

Lecture 11. Probability Theory: an Overveiw Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the

More information

Kernel Density Estimation

Kernel Density Estimation Kernel Density Estimation Univariate Density Estimation Suppose tat we ave a random sample of data X 1,..., X n from an unknown continuous distribution wit probability density function (pdf) f(x) and cumulative

More information

ST 371 (IX): Theories of Sampling Distributions

ST 371 (IX): Theories of Sampling Distributions ST 371 (IX): Theories of Sampling Distributions 1 Sample, Population, Parameter and Statistic The major use of inferential statistics is to use information from a sample to infer characteristics about

More information

Math 151. Rumbos Fall Solutions to Review Problems for Final Exam

Math 151. Rumbos Fall Solutions to Review Problems for Final Exam Math 5. Rumbos Fall 23 Solutions to Review Problems for Final Exam. Three cards are in a bag. One card is red on both sides. Another card is white on both sides. The third card in red on one side and white

More information

Survival Distributions, Hazard Functions, Cumulative Hazards

Survival Distributions, Hazard Functions, Cumulative Hazards BIO 244: Unit 1 Survival Distributions, Hazard Functions, Cumulative Hazards 1.1 Definitions: The goals of this unit are to introduce notation, discuss ways of probabilistically describing the distribution

More information

4.1 The Expectation of a Random Variable

4.1 The Expectation of a Random Variable STAT 42 Lecture Notes 93 4. The Expectation of a Random Variable This chapter begins the discussion of properties of random variables. The focus of this chapter is on expectations of random variables.

More information

Analysis of Time-to-Event Data: Chapter 2 - Nonparametric estimation of functions of survival time

Analysis of Time-to-Event Data: Chapter 2 - Nonparametric estimation of functions of survival time Analysis of Time-to-Event Data: Chapter 2 - Nonparametric estimation of functions of survival time Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term

More information

ECON 5350 Class Notes Review of Probability and Distribution Theory

ECON 5350 Class Notes Review of Probability and Distribution Theory ECON 535 Class Notes Review of Probability and Distribution Theory 1 Random Variables Definition. Let c represent an element of the sample space C of a random eperiment, c C. A random variable is a one-to-one

More information

It can be shown that if X 1 ;X 2 ;:::;X n are independent r.v. s with

It can be shown that if X 1 ;X 2 ;:::;X n are independent r.v. s with Example: Alternative calculation of mean and variance of binomial distribution A r.v. X has the Bernoulli distribution if it takes the values 1 ( success ) or 0 ( failure ) with probabilities p and (1

More information

Expectation, inequalities and laws of large numbers

Expectation, inequalities and laws of large numbers Chapter 3 Expectation, inequalities and laws of large numbers 3. Expectation and Variance Indicator random variable Let us suppose that the event A partitions the sample space S, i.e. A A S. The indicator

More information

Censoring mechanisms

Censoring mechanisms Censoring mechanisms Patrick Breheny September 3 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/23 Fixed vs. random censoring In the previous lecture, we derived the contribution to the likelihood

More information

Lecture 2: Repetition of probability theory and statistics

Lecture 2: Repetition of probability theory and statistics Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:

More information

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Noname manuscript No. (will be inserted by the editor) A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Mai Zhou Yifan Yang Received: date / Accepted: date Abstract In this note

More information

Statistics for scientists and engineers

Statistics for scientists and engineers Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3

More information

3 Continuous Random Variables

3 Continuous Random Variables Jinguo Lian Math437 Notes January 15, 016 3 Continuous Random Variables Remember that discrete random variables can take only a countable number of possible values. On the other hand, a continuous random

More information

1. Point Estimators, Review

1. Point Estimators, Review AMS571 Prof. Wei Zhu 1. Point Estimators, Review Example 1. Let be a random sample from. Please find a good point estimator for Solutions. There are the typical estimators for and. Both are unbiased estimators.

More information

STA 584 Supplementary Examples (not to be graded) Fall, 2003

STA 584 Supplementary Examples (not to be graded) Fall, 2003 Page 1 of 8 Central Michigan University Department of Mathematics STA 584 Supplementary Examples (not to be graded) Fall, 003 1. (a) If A and B are independent events, P(A) =.40 and P(B) =.70, find (i)

More information

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued Chapter 3 sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional

More information

Survival Analysis. Lu Tian and Richard Olshen Stanford University

Survival Analysis. Lu Tian and Richard Olshen Stanford University 1 Survival Analysis Lu Tian and Richard Olshen Stanford University 2 Survival Time/ Failure Time/Event Time We will introduce various statistical methods for analyzing survival outcomes What is the survival

More information

Department of Statistical Science FIRST YEAR EXAM - SPRING 2017

Department of Statistical Science FIRST YEAR EXAM - SPRING 2017 Department of Statistical Science Duke University FIRST YEAR EXAM - SPRING 017 Monday May 8th 017, 9:00 AM 1:00 PM NOTES: PLEASE READ CAREFULLY BEFORE BEGINNING EXAM! 1. Do not write solutions on the exam;

More information

Final Exam # 3. Sta 230: Probability. December 16, 2012

Final Exam # 3. Sta 230: Probability. December 16, 2012 Final Exam # 3 Sta 230: Probability December 16, 2012 This is a closed-book exam so do not refer to your notes, the text, or any other books (please put them on the floor). You may use the extra sheets

More information

Parameters Estimation for a Linear Exponential Distribution Based on Grouped Data

Parameters Estimation for a Linear Exponential Distribution Based on Grouped Data International Mathematical Forum, 3, 2008, no. 33, 1643-1654 Parameters Estimation for a Linear Exponential Distribution Based on Grouped Data A. Al-khedhairi Department of Statistics and O.R. Faculty

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An

More information

Transformations and Expectations

Transformations and Expectations Transformations and Expectations 1 Distributions of Functions of a Random Variable If is a random variable with cdf F (x), then any function of, say g(), is also a random variable. Sine Y = g() is a function

More information

Lecture 11: Probability, Order Statistics and Sampling

Lecture 11: Probability, Order Statistics and Sampling 5-75: Graduate Algorithms February, 7 Lecture : Probability, Order tatistics and ampling Lecturer: David Whitmer cribes: Ilai Deutel, C.J. Argue Exponential Distributions Definition.. Given sample space

More information

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems

More information

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline. Random Variables Amappingthattransformstheeventstotherealline. Example 1. Toss a fair coin. Define a random variable X where X is 1 if head appears and X is if tail appears. P (X =)=1/2 P (X =1)=1/2 Example

More information

2 (Statistics) Random variables

2 (Statistics) Random variables 2 (Statistics) Random variables References: DeGroot and Schervish, chapters 3, 4 and 5; Stirzaker, chapters 4, 5 and 6 We will now study the main tools use for modeling experiments with unknown outcomes

More information

This exam contains 6 questions. The questions are of equal weight. Print your name at the top of this page in the upper right hand corner.

This exam contains 6 questions. The questions are of equal weight. Print your name at the top of this page in the upper right hand corner. GROUND RULES: This exam contains 6 questions. The questions are of equal weight. Print your name at the top of this page in the upper right hand corner. This exam is closed book and closed notes. Show

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

16.4. Power Series. Introduction. Prerequisites. Learning Outcomes

16.4. Power Series. Introduction. Prerequisites. Learning Outcomes Power Series 6.4 Introduction In this Section we consider power series. These are examples of infinite series where each term contains a variable, x, raised to a positive integer power. We use the ratio

More information

Statistical Methods in Particle Physics

Statistical Methods in Particle Physics Statistical Methods in Particle Physics. Probability Distributions Prof. Dr. Klaus Reygers (lectures) Dr. Sebastian Neubert (tutorials) Heidelberg University WS 07/8 Gaussian g(x; µ, )= p exp (x µ) https://en.wikipedia.org/wiki/normal_distribution

More information

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued Chapter 3 sections Chapter 3 - continued 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions

More information

1 Glivenko-Cantelli type theorems

1 Glivenko-Cantelli type theorems STA79 Lecture Spring Semester Glivenko-Cantelli type theorems Given i.i.d. observations X,..., X n with unknown distribution function F (t, consider the empirical (sample CDF ˆF n (t = I [Xi t]. n Then

More information

Statistics 100A Homework 5 Solutions

Statistics 100A Homework 5 Solutions Chapter 5 Statistics 1A Homework 5 Solutions Ryan Rosario 1. Let X be a random variable with probability density function a What is the value of c? fx { c1 x 1 < x < 1 otherwise We know that for fx to

More information

Nonparametric Function Estimation with Infinite-Order Kernels

Nonparametric Function Estimation with Infinite-Order Kernels Nonparametric Function Estimation with Infinite-Order Kernels Arthur Berg Department of Statistics, University of Florida March 15, 2008 Kernel Density Estimation (IID Case) Let X 1,..., X n iid density

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

Hybrid Censoring; An Introduction 2

Hybrid Censoring; An Introduction 2 Hybrid Censoring; An Introduction 2 Debasis Kundu Department of Mathematics & Statistics Indian Institute of Technology Kanpur 23-rd November, 2010 2 This is a joint work with N. Balakrishnan Debasis Kundu

More information

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS EXAM Exam # Math 3342 Summer II, 2 July 2, 2 ANSWERS i pts. Problem. Consider the following data: 7, 8, 9, 2,, 7, 2, 3. Find the first quartile, the median, and the third quartile. Make a box and whisker

More information

The Type I Generalized Half Logistic Distribution Downloaded from jirss.irstat.ir at 11: on Monday August 20th 2018

The Type I Generalized Half Logistic Distribution Downloaded from jirss.irstat.ir at 11: on Monday August 20th 2018 JIRSS (2014) Vol. 13, No. 1, pp 69-82 The Type I Generalized Half Logistic Distribution A. K. Olapade Department of Mathematics, Obafemi Awolowo University, Ile-Ife, Nigeria. Abstract. In this paper, we

More information

Statistics 135 Fall 2008 Final Exam

Statistics 135 Fall 2008 Final Exam Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

STAT 830 Non-parametric Inference Basics

STAT 830 Non-parametric Inference Basics STAT 830 Non-parametric Inference Basics Richard Lockhart Simon Fraser University STAT 801=830 Fall 2012 Richard Lockhart (Simon Fraser University)STAT 830 Non-parametric Inference Basics STAT 801=830

More information

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer

More information

Distribution Fitting (Censored Data)

Distribution Fitting (Censored Data) Distribution Fitting (Censored Data) Summary... 1 Data Input... 2 Analysis Summary... 3 Analysis Options... 4 Goodness-of-Fit Tests... 6 Frequency Histogram... 8 Comparison of Alternative Distributions...

More information

Math Spring Practice for the Second Exam.

Math Spring Practice for the Second Exam. Math 4 - Spring 27 - Practice for the Second Exam.. Let X be a random variable and let F X be the distribution function of X: t < t 2 t < 4 F X (t) : + t t < 2 2 2 2 t < 4 t. Find P(X ), P(X ), P(X 2),

More information

Exam P Review Sheet. for a > 0. ln(a) i=0 ari = a. (1 r) 2. (Note that the A i s form a partition)

Exam P Review Sheet. for a > 0. ln(a) i=0 ari = a. (1 r) 2. (Note that the A i s form a partition) Exam P Review Sheet log b (b x ) = x log b (y k ) = k log b (y) log b (y) = ln(y) ln(b) log b (yz) = log b (y) + log b (z) log b (y/z) = log b (y) log b (z) ln(e x ) = x e ln(y) = y for y > 0. d dx ax

More information

CS145: Probability & Computing

CS145: Probability & Computing CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,

More information

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES VARIABLE Studying the behavior of random variables, and more importantly functions of random variables is essential for both the

More information

Review of Probability Theory

Review of Probability Theory Review of Probability Theory Arian Maleki and Tom Do Stanford University Probability theory is the study of uncertainty Through this class, we will be relying on concepts from probability theory for deriving

More information

HW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100]

HW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100] HW7 Solutions. 5 pts.) James Bond James Bond, my favorite hero, has again jumped off a plane. The plane is traveling from from base A to base B, distance km apart. Now suppose the plane takes off from

More information

Probability and Measure

Probability and Measure Chapter 4 Probability and Measure 4.1 Introduction In this chapter we will examine probability theory from the measure theoretic perspective. The realisation that measure theory is the foundation of probability

More information

Statistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University

Statistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University Statistics for Economists Lectures 6 & 7 Asrat Temesgen Stockholm University 1 Chapter 4- Bivariate Distributions 41 Distributions of two random variables Definition 41-1: Let X and Y be two random variables

More information

STAT 302 Introduction to Probability Learning Outcomes. Textbook: A First Course in Probability by Sheldon Ross, 8 th ed.

STAT 302 Introduction to Probability Learning Outcomes. Textbook: A First Course in Probability by Sheldon Ross, 8 th ed. STAT 302 Introduction to Probability Learning Outcomes Textbook: A First Course in Probability by Sheldon Ross, 8 th ed. Chapter 1: Combinatorial Analysis Demonstrate the ability to solve combinatorial

More information

UNIVERSITY OF CALIFORNIA, SAN DIEGO

UNIVERSITY OF CALIFORNIA, SAN DIEGO UNIVERSITY OF CALIFORNIA, SAN DIEGO Estimation of the primary hazard ratio in the presence of a secondary covariate with non-proportional hazards An undergraduate honors thesis submitted to the Department

More information

Master s Written Examination - Solution

Master s Written Examination - Solution Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2

More information

5 Introduction to the Theory of Order Statistics and Rank Statistics

5 Introduction to the Theory of Order Statistics and Rank Statistics 5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order

More information

Lecture 17: The Exponential and Some Related Distributions

Lecture 17: The Exponential and Some Related Distributions Lecture 7: The Exponential and Some Related Distributions. Definition Definition: A continuous random variable X is said to have the exponential distribution with parameter if the density of X is e x if

More information

Statistics 427: Sample Final Exam

Statistics 427: Sample Final Exam Statistics 427: Sample Final Exam Instructions: The following sample exam was given several quarters ago in Stat 427. The same topics were covered in the class that year. This sample exam is meant to be

More information