On Statistical Methods for Zero-Inflated Models

Size: px
Start display at page:

Download "On Statistical Methods for Zero-Inflated Models"

Transcription

1 U.U.D.M. Project Report 2015:9 On Statistical Methods for Zero-Inflated Models Julia Eggers Examensarbete i matematik, 15 hp Handledare och examinator: Silvelyn Zwanzig Juni 2015 Department of Mathematics Uppsala University

2

3 Abstract Data with excess zeros arise in many contexts. Conventional probability distributions often cannot explain large proportions of zero observations. In this paper we shall study statistical models which take large proportions of zero observations into account. We will consider both discrete and continuous distributions.

4 Contents 1 Introduction 2 2 Models for Zero-Inflated Data 7 3 Models for Semicontinuous Data with Excess Zeros Tobit Models Sample Selection Models Double Hurdle Models Two-Part Models Inference in Models for Zero-Inflated Data The Likelihood Function Maximum Likelihood Estimators Moment Estimators Cold Spells in Uppsala Exponential Family Inference in Two-Part Models The Likelihood Function Maximum Likelihood Estimators Moment Estimators Exponential Family Hypothesis Testing References 26 1

5 1. Introduction In this paper we will study models for data with a large proportion of zeros. For this we will introduce a few terms. Definition Discrete probability distributions with a large probability mass at zero are said to be zero-inflated. Conventional distributions usually cannot explain the large proportion of zeros in zero-inflated data. For this reason different models which can account for a large proportion of zero observations must be applied instead. Definition Probability distributions which are continuous on the entire sample space with the exception of one value at which there is a positive point mass are said to be semicontinuous. In this paper we will study models for zero-inflated distributions as well as for semicontinuous distributions with a positive probability mass at zero. We will only consider distributions with non-negative support. Remark Unlike in the case of left-censored data, zeros in semicontinuous data correspond to actual observations and do not represent negative or missing values which have been coded as zero. Data with excess zeros may arise in many different contexts. We will start by giving a few examples. Examples of zero-inflated data Cold spells In his paper on trends for warm and cold spells in Uppsala, Jesper Rydén studied the yearly number of cold spells in Uppsala, Sweden for the period from 1840 to 2012 [07]. He defined a cold spell as a period of at least six consecutive days during which the daily minimum temperature was less 2

6 than 13.4 C. The threshold of 13.4 C was chosen as it corresponds to the 5%-quantile of daily minimum temperatures for the reference period The yearly number of cold spells in Uppsala appears to be zero-inflated as can be seen from the data below. There is a large proportion of zero observations, i.e. years during which there were no cold spells. Figure 1.1: Yearly number of cold spells in Uppsala, Defects in manufacturing In manufacturing processes defects usually only occur when manufacturing equipment is not properly aligned. If the equipment is misaligned, defects can be found to occur according to a Poisson distribution [08]. This implies that defects in manufacturing occur according to a Poisson distribution with inflation at zero. Examples of semicontinuous data with excess zeros Household expenditures on durable goods The amount of money a household spends monthly on certain durable goods such as cars or appliances like washing machines or refrigerators is distributed according to a semicontinuous distribution. During most months no such goods are purchased and the expense is zero. If durable goods are purchased, the household expenditure on durable goods for that 3

7 month amounts to some positive value, namely the price of the purchased items. Alcohol consumption Consider the alcohol consumption of a population during a certain period of study. Some people belonging to the population may not drink any alcohol at all, thus consuming zero liters of alcohol. These people account for a point mass at zero. People who do consume alcohol may consume arbitrarily large, but positive, amounts. Thus we have a continuous distribution for positive values. Similarly, the tobacco consumption or consumption of drugs in general is semicontinuously distributed. Insurance benefits The Swedish Social Insurance Agency Forsäkringskassan publishes annual reports on its expenditures. The publication Social Insurance in Figures 2014 states that in 2013 a total amount of approximately 24.1 million SEK were paid out as sickness benefits. These sickness benefits are meant to compensate insured for the inability to work due to illness. In 2013, people in Sweden received sickness benefits. This corresponds to around 9% of all insured between the ages of 16 and 64. The amounts of sickness benefits paid out to insured during the year 2013 are semicontinuously distributed. 91% of all insured received no such benefits. We thus have a positive probability mass at zero. Those people who did receive sickness benefits got positive amounts which varied according to factors like income and time spent on sick leave. Therefore, we have a continuous distribution for positive values. Table 1.1 below gives an account of the average amounts of sickness benefits paid out to insured depending on gender and age group. 4

8 Table 1.1: Sickness benefits, 2013 We know from the data that insured received sickness benefits in 2013 while around insured received no such benefits. Since 24.1 million SEK were paid out in total, the average positive amount that was paid out per person that year amounted to approximately SEK. Assuming that the paid out benefits are exponentially distributed with parameter λ given that they are positive, we may estimate ˆλ = 1/ Generating a sample from this distribution in R, we may illustrate how the sickness benefits may have been distributed. Figure 1.2: A possible distribution of the amount of sickness benefits paid out to insured during the year

9 Healthcare expenditures Healthcare expenditures in general can be found to be semicontinuously distributed. Individuals may or may not choose to seek medical treatment during a certain period of study. There are no costs arising for people who do not seek medical treatment. If medical treatment is sought, however, the cost for the treatment amounts to some positive value. Thus healthcare expenditures are continuously distributed for positive values and have a positive probability mass at zero. 6

10 2. Models for Zero-Inflated Data The models for zero-inflated data which we will present here are variations of the following mixture model Y = Z 1 + (1 )Z 2 with Ber(p), Z 1 P Z1 and Z 2 P Z2. If we let P Z2 = δ {0} and assume Z 1 to be discrete, we obtain a model for zeroinflated data. When modeling count data we have the additional assumption that P (Z 1 0) = 1. For the above model we have that Y δ {0} with probability 1 p and Y P Z1 with probability p. Letting p Z1 denote the probability mass function of the random variable { Z 1 we obtain 1 p + ppz1 (0), y = 0 P (Y = y) = pp Z1 (y), y > 0 When modeling count data, the negative binomial and the Poisson distribution are common distributions for Z 1. If Z 1 P o(λ) the above model is referred to as the zero-inflated Poisson model, abbreviated ZIP. Zero-Inflated Poisson Regression Zero-inflated Poisson regression is an extension of the zero-inflated Poisson model which was proposed by Diane Lambert in 1992 [08]. The model assumes Y = (Y 1,..., Y n ) to be a sample of independent, but not necessarily identically distributed random variables Y i. In this model we assume Y i P o(λ i ) with probability p i. { 1 pi + p i exp( λ i ), y i = 0 Thus P (Y i = y i ) = p i λ y i i exp( λ i) y i!, y i > 0 The parameters p = (p 1,..., p n ) and λ = (λ 1,..., λ n ) are assumed to satisfy 7

11 log(λ) = Bβ and logit(p) = log( p 1 p ) = Gγ with B and G denoting matrices with explanatory variables. β and γ are matrices with coefficients to adequately describe the linear dependency of log(λ) and logit(p) on B and G respectively. If p and λ depend on the same explanatory variables, the number of model parameters may be reduced by expressing p as a function of λ. Lambert proposes the relation logit(p) = τ log(λ) for some τ R. This implies that p i = 1 1+λ i. The resulting model is denoted by ZIP (τ). 8

12 3. Models for Semicontinuous Data with Excess Zeros There are a number of different models which can be applied to semicontinuous data with excess zeros. We will present a few of the most common ones. In all of these models we will let Y denote the observed variable. The models we will present are all special kinds of two-component mixture models. A mixture model with two components has the form Y = Z 1 + (1 )Z 2 with Ber(p), Z 1 P Z1 and Z 2 P Z2. In the models we will present, we have that P Z2 = δ {0}. 3.1 Tobit Models The Tobit model which was proposed by James Tobin in 1958 [03] assumes that Y can be expressed in terms of a latent variable Y which can only be observed for values greater than zero. The random variable Y is defined as follows. { Y, Y Y = > 0 0, Y 0 The latent variable Y is assumed to be linearly dependent on a number of explanatory (and observable) variables and can be expressed as a linear combination of these, i.e. Y = Xβ + ɛ where X is a row vector containing the explanatory variables and β is a column vector with the corresponding coefficients describing the linear dependency of Y on X. The error terms ɛ are assumed to be independently and identically distributed 9

13 according to N(0, σ 2 ). Thus the Tobit model assumes an underlying normal distribution. The probability that Y takes the value zero is given by P (Y = 0) = P (Y 0) = P (Xβ + ɛ 0) = P (ɛ Xβ) = = P ( ɛ σ Xβ ) ( Xβ ) (Xβ ) = Φ = 1 Φ. σ σ σ This part of the Tobit model corresponds to the so-called Probit model. The name Tobit alludes to the Tobit model having been proposed by Tobin and being based on the Probit model. The likelihood function L of the uncensored positive values of Y is given by the probability density function of the latent variable Y given that it is positive, i.e. L(y y > 0) = L(y y > 0) = 1 σ φ( y Xβ ). σ In Tobit models the probability of a zero observation depends on the same random variable that determines the magnitude of the observation given that it is positive. Note that in the Tobit model zeros do not represent actual responses. The Tobit model is therefore not appropriate for semicontinuous data. It is, however, often applied to such data in spite of this. Remark There are many variations of the Tobit model. Censoring can for instance be performed at values other than zero. There are also models where censoring is done from above instead of below or from both above and below. Remark The mixture model above corresponds to the Tobit model if Z 1 = Y and p = P (Y > 0) = Φ ( ) Xβ σ. 3.2 Sample Selection Models The sample selection model was first proposed by J. Heckman in 1979 as an extension of the Tobit model. Sample selection models are based on two latent variables Y 1 and Y 2. The first latent variable Y1 is assumed to be of the form Y1 = X 1 β 1 + ɛ 1 with X 1 being a row vector of explanatory variables and β 1 being the corresponding vector of coefficients describing the linear dependency of Y1 on X 1. Similarly, the second latent variable Y 2 is assumed be of the form Y 2 = X 2 β 2 + ɛ 2, again with X 2 being a row vector of explanatory variables and β 2 being the corresponding vector of coefficients. 10

14 Sample selection models thus allow for the latent variables to depend on different covariates. The error terms (ɛ 1, ɛ 2 ) are assumed to be independently and identically distributed according to a bivariate normal distribution. They may thus be correlated. { Y The observed variable is defined as Y = 2, Y1 > 0 0, Y1 0. The sample selection model coincides with the Tobit model if X 1 = X 2 and β 1 = β 2 (i.e. Y1 = Y2 ). Remark The mixture model above corresponds to the sample selection model if Z 1 = Y2 and p = P (Y1 > 0). 3.3 Double Hurdle Models Similarly to sample selection models, double hurdle models are based on two latent variables Y1 and Y2. These latent variables are again assumed to be of the form Y1 = X 1 β 1 + ɛ 1 and Y2 = X 2 β 2 + ɛ 2 with X 1 and X 2 denoting row vectors with observed values of explanatory variables and β 1 and β 2 denoting column vectors that contain the corresponding coefficients describing the linear dependency of Y1 and Y2 on X 1 and X 2 respectively. (ɛ 1, ɛ 2 ) are again assumed to be independently and identically distributed according to a bivariate normal distribution. In double { hurdle models, the observed variable is defined as Y Y = 2, Y1 > 0 and Y2 > 0. 0, otherwise To illustrate the idea behind double hurdle models, we will apply it to the example of tobacco consumption. We thus let Y denote the amount of tobacco consumed by an individual during a certain period of time. The first latent variable may Y1 determine whether an individual is a smoker or non-smoker. This may depend on certain socioeconomic factors which can be accounted for by the dependency of Y1 on X 1. The second latent variable Y 2 may thereafter be used to determine how much tobacco is consumed by an individual given that the individual is a smoker. This quantity may depend on other covariates than the ones that affected the probability of the individual being a smoker in the first place. Note that it is possible for a smoker not to consume any tobacco during the period of the study, in other words we may have Y2 0 Y1 > 0. 11

15 We see that in order to observe positive values of Y two hurdles need to be overcome. The individual must be a smoker and smoke during the period of the study. Hence the name double hurdle model. Remark The mixture model above corresponds to the double hurdle model if Z 1 = Y2 and p = P (Y1 > 0, Y2 > 0). 3.4 Two-Part Models As the name suggests, two-part models consist of two parts. In the first part of the model a random variable determines whether the observation is zero or positive. In the second part another random variable Z determines the magnitude of the observation given that it is positive. The value of the random variable Z is not observed if has taken the value zero. The random variables and Z are assumed to be independent. Moreover, we assume that P (Z > 0) = 1. In other words we have the following model for the random variable Y Y = 1 {1} ( )Z = Z, (3.1) with Ber(θ 1 ) and Z P Z {P θ2 } being independent and P (Z > 0) = 1. Thus Y P Y {P θ, θ = (θ 1, θ 2 )}. Two-part models do not assume an underlying normal distribution and can therefore be applied to a wider range of data than for instance Tobit models. Note that in two-part models we do not have a latent variable. Zeros correspond to actual observations, and are not the result of censoring as in the previously presented models. Consequently, two-part models are more appropriate for modeling semicontinuous data than the other models we have presented. In the following, we will therefore restrict ourselves to the study of two-part models. Remark The mixture model above corresponds to the two-part model if Z 1 = Z. Remark Note that in all the models for semicontinuous data with excess zeros we have that P (Z 1 = 0) = 0 and P (Z 2 = 0) = 1. We can therefore distinguish between observations from Z 1 and Z 2. For zero-inflated count data, however, we have that P (Z 2 = 0) = 1 and P (Z 1 = 0) > 0. Here we are unable to distinguish between zero observations from Z 1 and zero observations from Z 2. 12

16 4. Inference in Models for Zero- Inflated Data Let Y denote the observed variable. We will assume the following model P Y {P θ, θ = (p, λ)} for Y. Y = Z 1 + (1 )Z 2 with Ber(p), Z 1 P Z1 {P λ } and Z 2 δ {0} being independent. Moreover, we assume that Z 1 is discrete and that P (Z 1 0) = 1. Note that the observed variable Y has non-negative support. Thus the above model can be applied to, for instance, count data. { 1 p + ppz1 (0), y = 0 For this model we have P (Y = y) = pp Z1 (y), y > 0. with p Z1 denoting the probability mass function of Z 1. In the case that { Z 1 P o(λ) we obtain a zero-inflated Poisson model with 1 p + p exp( λ), y = 0 P (Y = y) = p λy exp( λ) y!, y > 0. Theorem The expected value E[Y ] and variance V ar[y ] of Y are given by E[Y ] = pe[z 1 ] and V ar[y ] = pv ar[z 1 ] + (1 p)pe[z 1 ] 2. Proof. The expected value of Y is given by E[Y ] = E[ Z 1 + (1 )Z 2 ] = E[ ]E[Z 1 ] + E[Z 2 ] E[ ]E[Z 2 ] = pe[z 1 ]. The variance of Y is given by V ar[y ] = V ar[ Z 1 ] + V ar[z 2 ] + V ar[ Z 2 ] = V ar[ Z 1 ] + V ar[ Z 2 ] = = E[ 2 ]E[Z 2 1] E[ ] 2 E[Z 1 ] 2 + E[ 2 ]E[Z 2 2] E[ ] 2 E[Z 2 ] 2 = = E[ 2 ]E[Z 2 1] E[ ] 2 E[Z 1 ] 2 = = (V ar[ ] + E[ ] 2 )(V ar[z 1 ] + E[Z 1 ] 2 ) E[ ] 2 E[Z 1 ] 2 = = (p(1 p) + p 2 )(V ar[z 1 ] + E[Z 1 ] 2 ) p 2 E[Z 1 ] 2 = = pv ar[z 1 ] + p(1 p)e[z 1 ] 2 Corollary In the zero-inflated Poisson model the expected value E[Y ] and variance V ar[y ] of Y are given by E[Y ] = pλ and V ar[y ] = pλ(1+λ pλ). 13

17 Proof. In the zero-inflated Poisson model Z 1 P o(λ) and E[Z 1 ] = V ar[z 1 ] = λ. Plugging these values in to the expressions for E[Y ] and V ar[y ] yields the above result. 4.1 The Likelihood Function Definition The likelihood function L(θ, y) : Θ R + for an observation y of a random variable Y with probability function p(θ, y) is given by L(θ, y) := p(θ, y). (4.1) For a sample Y = (Y 1,..., Y n ) of independent and identically distributed random variables the likelihood function is given by n L(θ, y) := p(θ, y i ). (4.2) We will now consider a sample y = (y 1,..., y n ) of independent and identically distributed random variables Y i P Y. Let r denote the number of zero observations in the sample y. The likelihood function L(p, λ, y) of the sample y is given by L(p, λ, y) = n P (Y i = y i ) = (1 p + pp Z1 (0)) pp Z1 (y i ) = If Z 1 P o(λ) we have y i=0 = (1 p + pp Z1 (0)) r p n r p Z1 (y i ). y i>0 y i>0 L(p, λ, y) = (1 p + p exp( λ)) r p n r λ y i exp( λ(n r)) n. y i! 4.2 Maximum Likelihood Estimators Definition The maximum likelihood estimator ˆθ MLE of a variable θ is a value of θ which maximizes the likelihood function, i.e. with χ denoting the sample space of Y. ˆθ MLE max L(θ, y) y χ (4.3) θ Θ Theorem Let p (0, 1) and Z 1 P o(λ), i.e. assume a zero-inflated Poisson model for the sample y. The maximum likelihood estimators ˆp MLE (Y ) and ˆλ MLE (Y ) are given by ˆp MLE (Y ) = n r n(1 e ˆλ MLE ). 14

18 and (1 e ˆλ MLE ) y i = ˆλ MLE (n r). Proof. The likelihood function L(p, λ, y) of the sample y is given by L(p, λ, y) = (1 p + p exp( λ)) r p n r λ y i exp( λ(n r)) n y i! The values of p for which L(p, λ, y) is maximized satisfy pl(p, λ, y) = 0 p p ln(l(p, λ, y)) = 0 ( r ln(1 p+p exp( λ))+(n r) ln(p)+ y i ln(λ) λ(n r) ln( n y i!) ) = 0 r( 1+e λ ) 1 p+pe λ p = n r n(1 e λ ) + n r p = 0 The values of λ which maximize L(p, λ, y) satisfy λl(p, λ, y) = 0 ( λ r ln(1 p+p exp( λ))+(n r) ln(p)+ y i ln(λ) λ(n r) ln( n y i!) ) = 0 Inserting p = n r n(1 e λ ) gives (1 e λ ) n y i = λ(n r) Remark Numerical methods must be applied to solve the equation above. (1 e λ ) y i = λ(n r) 4.3 Moment Estimators Definition Let Y = (Y 1, Y 2,..., Y n ) be a sample from independent and identically distributed random variables with distributions depending on a parameter θ. The moment estimator of order k for θ is given by the value of θ for which E[Y k ] = g(θ) = 1 n Y k i 15

19 where g is some function specifying the expected value. Theorem Let Z 1 P o(λ), i.e. assume a zero-inflated Poisson model for the sample y. The moment estimators ˆp MME (Y ) and ˆλ MME (Y ) are given by ) 2 Y i and ˆp MME (Y ) = 1 n ˆλ MME (Y ) = ( 1 n Y 2 i 1 n Y i Y 2 i 1. Y i Proof. The moment estimators ˆp MME (Y ) and ˆλ MME (Y ) are given by values of p and λ which satisfy E[Y ] = 1 n Y i = pλ E[Y 2 ] = 1 n Yi 2 = V ar[y ] + E[Y ] 2 = pλ(1 + λ pλ) + p 2 λ 2 = pλ(1 + λ) (1 + λ) = λ = p = Y 2 i 1 n Y 2 i = Y i Y i 1 n Y i 1 1 n Y i λ = 1 n Y i Y i 2 Y i Y 2 i Y i = 1 n ( ) 2 1 n Y i Yi 2 1 n Y i 4.4 Cold Spells in Uppsala We will now assume a zero-inflated Poisson model for the data regarding cold spells in Uppsala (see chapter 1, figure 1.1) [07]. We let y = (y 1,..., y 169 ) denote the corresponding sample and assume that the observations are independent and identically distributed. Note that, in reality, the number of cold spells that occur during two consecutive years may not actually be independent so this assumption may not hold. For this sample we have 169 y i = 142 and 169 y 2 i = 340. We thus obtain the following moment estimates for the model parameters p and λ. 16

20 ˆp MME (y) = ( ˆλ MME (y) = y i ) 2 y 2 i yi = y i y i = ( ) = Note that ˆp MME (y) 0.6 < 1 so the yearly number of cold spells in Uppsala does indeed appear to be zero-inflated. Theorem An approximate level α test for the testing problem is given by where φ(y) = H 0 : p = 1, λ = 1.39 H 1 : 0 < p < 1, λ = 1.39 { 1, 2 ln(λ(y)) χ 2 α (1) 0, 2 ln(λ(y)) < χ 2 α(1) n r Λ(y) = (1 n(1 e 1.39 ) + n r n(1 e 1.39 ) e 1.39 ) r n r ( n(1 e 1.39 ) ) (n r) e 1.39r. Proof. The likelihood ratio Λ(Y ) is given by Λ(Y ) = max{p 0(y) : p = 1, λ = 1.39} max{p(y) : 0 < p 1, λ = 1.39} = 1 max (1 p + 0<p 1 pe 1.39 ) r p n r e 1.39r = n r = (1 n(1 e 1.39 ) + n r n(1 e 1.39 ) e 1.39 ) r n r ( n(1 e 1.39 ) ) (n r) e 1.39r According to Wilk s Theorem 2 ln(λ(y )) χ 2 (1) approximately as n. We can thus reject H 0 at significance level α if 2 ln(λ(y )) χ 2 α(1). This yields the above test. For the sample y we obtain 2 ln(λ(y)) = χ (1) = 3.84 It therefore follows that H 0 can be rejected at significance level Moreover, the p-value for the test is given by p = P 0 ( 2 ln(λ(y )) 66.92) = = 1 P 0 ( 2 ln(λ(y )) < 66.92) = , so H 0 can be rejected at any significance level α > We conclude that the yearly number of cold spells in Uppsala are zero-inflated. 17

21 4.5 Exponential Family Definition A class of probability measures P = {P θ : θ Θ} is called an exponential family if ( k ) L(y; θ) = A(θ) exp ζ j (θ)t j (y) h(y) (4.4) for some k N, real-valued functions ζ 1,..., ζ k on Θ, real-valued statistics T 1,..., T k and a function h on the sample space χ. j=1 Theorem If the class of probability measures P = {P θ : θ Θ} forms an exponential family, then all P θ are pairwise equivalent, i.e. for any P,Q P we have P(N)=0 iff Q(N)=0 [06]. Theorem If p [0, 1], then P Y does not form an exponential family. Proof. Consider the probability measures P (0,λ1) and P (p2,λ 2) with p 2 (0, 1]. We have that P (0,λ1)(R\{0}) = 0 but P (p2,λ 2)(R\{0}) > 0. Therefore it follows that P Y does not form an exponential family. Theorem If P Z1 does not form an exponential family, then P Y does not form an exponential family. Proof. The likelihood L(p, λ, y) of y is given by L(p, λ, y) = ((1 p) + pp Z1 (0)) r p n r p Z1 (y i ). y i>0 Since P Z1 does not form an exponential family, p Z1 (y i ) is not of the form (4.4). Thus L(p, λ, y) is not of the form (4.4). Consequently, P Y is not an exponential family. Theorem If p (0, 1] and P Z1 is a k-parameter exponential family with natural parameters ζ j (λ) and sufficient statistics T j (z), j = 1,..., k, then P Y is a k+1 parameter exponential family with natural parameters ζ j (λ), j = 1,..., k and ζ k+1 (p) = ln( 1 p+pp Z 1 (0) p ) and sufficient statistics T j (y), j = 1,..., k and T k+1 (y) = r. Proof. We have that L(p, λ, y) = ((1 p) + pp Z1 (0)) r p n r p Z1 (y i ) = y i>0 = exp ( r ln(1 p + pp Z1 (0)) + (n r) ln(p) ) p Z1 (y i ) = y i>0 = p n exp ( ) r ln( 1 p+pp Z 1 (0) p ) p Z1 (y i ) = y i>0 = p n A(λ) exp ( r ln( 1 p+pp Z 1 (0) p ) + k ) ζ j (λ)t j (y) h(y) j=1 Thus P Y {P (p,λ) } forms an exponential family with natural parameters ζ j (λ), j = 1,..., k and ζ k+1 (p) = ln( 1 p+pp Z 1 (0) p ) and sufficient statistics T j (y), j = 1,..., k and T k+1 (y) = r. 18

22 5. Inference in Two-Part Models Consider a random variable Y P Y {P θ, θ = (θ 1, θ 2 )} distributed according to the two-part model, i.e. let Y = 1 {1} ( )Z = Z, with and Z being independent, Ber(θ 1 ), Z P Z {P θ2 } and P (Z > 0) = 1. Theorem The expected value E[Y ] and variance V ar[y ] of Y are given by E[Y ] = θ 1 E[Z] and V ar[y ] = θ 1 V ar[z] + (1 θ 1 )θ 1 E[Z] 2. Proof. The expected value of Y is given by E[Y ] = E[ Z] = E[ ]E[Z] = θ 1 E[Z]. The variance of Y is given by V ar[y ] = V ar[ Z] = E[ 2 ]E[Z 2 ] E[ ] 2 E[Z] 2 since and Z are independent.thus V ar[y ] = E[ 2 ]E[Z 2 ] E[ ] 2 E[Z] 2 = = (V ar[ ] + E[ ] 2 )(V ar[z] + E[Z] 2 ) E[ ] 2 E[Z] 2 = = V ar[ ]V ar[z] + V ar[ ]E[Z] 2 + E[ ] 2 V ar[z] + E[ ] 2 E[Z] 2 E[ ] 2 E[Z] 2 = = V ar[ ]V ar[z] + V ar[ ]E[Z] 2 + E[ ] 2 V ar[z] = = (V ar[ ] + E[ ] 2 )V ar[z] + V ar[ ]E[Z] 2 = = ((1 θ 1 )θ 1 + θ 2 1)V ar[z] + (1 θ 1 )θ 1 E[Z] 2 = = θ 1 V ar[z] + (1 θ 1 )θ 1 E[Z] 2 It can be shown that if E[Z], ˆ E[Z] 2 and V ar[z], then and ˆ V ar[y ] = ˆ E[Y ] = { n r n ˆ E[Z] 2 and { n r n ˆ V ar[z] are unbiased estimators of E[Z], E[Z] ˆ, n r > 0 0, n r = 0 V ar[z] ˆ + n r r ˆ n n 1E[Z] 2, n r > 0 0, n r = 0 are unbiased estimators of E[Y ] and V ar[y ] respectively, see [10]. 19

23 5.1 The Likelihood Function Theorem Let y = (y 1,..., y n ) be a sample from independent and identically distributed random variables Y i distributed according to the two-part model. Moreover, let r denote the number of zero observations in the sample y and let z = (z 1,..., z n r ) be the subsample of positive observations of y. The likelihood function L(θ 1, θ 2, y) of the sample y is given by L(θ 1, θ 2, y) = L 1 (θ 1, y)l 2 (θ 2, y) where L 1 (θ 1, y) = (1 θ 1 ) r θ n r 1 and L 2 (θ 2, y) = L(θ 2, z). Proof. The likelihood function of the sample y is given by L(θ 1, θ 2, y) = P (Y i = 0) P (Y i > 0)f(y i y i > 0) = = (1 θ 1 ) y i=0 y i>0 y i=0 y i>0 θ 1 f(y i y i > 0) = (1 θ 1 ) r θ n r 1 f(y i y i > 0) y i>0 with f(y i y i > 0) denoting the probability density function of the observations given that they are positive, i.e. the probability density function of Z. Thus f(y i y i > 0) = f(z i ). It follows that L(θ 1, θ 2, y) = L 1 (θ 1, y)l 2 (θ 2, y) with L 1 (θ 1, y) = (1 θ 1 ) r θ1 n r and L 2 (θ 2, y) = f(y i y i > 0) = n r f(z i ) = L(θ 2, z). y i>0 Remark Here L(θ 2, z) denotes the likelihood of the subsample z. 5.2 Maximum Likelihood Estimators Theorem Let θ 1 (0, 1). The likelihood estimates of θ 1 and θ 2 are given by ˆθ 1,MLE = n r n and ˆθ 2,MLE max θ 2 Θ 2 L(θ 2, z). Proof. Since the likelihood function L of the sample y can be expressed as a product of two functions L 1 and L 2 that each depend on only θ 1 or θ 2, L may be maximized by maximizing L 1 and L 2 respectively. In other words, we have max L(θ 1, θ 2, y) = max L 1 (θ 1, y) max L 2 (θ 2, y). θ Θ θ 1 Θ 1 θ 2 Θ 2 20

24 Maximum likelihood estimation of θ 1 We have that ˆθ 1,MLE max L 1 (θ 1, y). In other words the maximum likelihood θ 1 Θ 1 estimate ˆθ 1MLE of θ 1 is a value of θ 1 which maximizes L 1 (θ 1, y) = (1 θ 1 ) r θ1 n r. The values of θ 1 for which L 1 (θ 1, y) is maximized satisfy the following equation. θ 1 L(θ 1, y) = 0 θ 1 ln(l(θ 1, y)) = 0 θ 1 r ln(1 θ 1 ) + (n r) ln(θ 1 ) = 0 θ 1 = n r n Maximum likelihood estimation of θ 2 The maximum likelihood estimate ˆθ 2MLE of θ 2 is a value of θ 2 which maximizes L 2 (θ 2, y) = L(θ 2, z), i.e. ˆθ2,MLE max L(θ 2, z). To obtain a maximum θ 2 Θ 2 likelihood estimate of θ 2 we therefore only need to consider the subsample z of positive observations of y. 5.3 Moment Estimators Theorem The moment estimators for θ 1 and θ 2 satisfy θ 1 E[Z] = 1 n Y i θ 1 V ar[z] + θ 1 E[Z] 2 = 1 n Yi 2 Proof. The first moment of Y is given by E[Y ] = 1 n Y i = θ 1 E[Z]. The second moment is given by E[Y 2 ] = 1 n Yi 2 = V ar[y ]+E[Y ] 2 = θ 1 V ar[z]+(1 θ 1 )θ 1 E[Z] 2 +θ1e[z] 2 2 = = θ 1 V ar[z] + θ 1 E[Z] 2. This yields the above result. Corollary If Z Exp(θ 2 ), the moment estimators ˆθ 1,MME (Y ) and ˆθ 2,MME (Y ) are given by 2 ( ) 2 Y i ˆθ 1,MME (Y ) = n n Y 2 i 21

25 and 2 n ˆθ 2,MME (Y ) = Proof. If Z Exp(θ 2 ), E[Z] = 1 θ 2 and V ar[z] = 1. Plugging these values into θ2 2 the equations for the first and second moment, we obtain θ 1 θ 2 = 1 n 2θ 1 θ θ 2 1 n θ 2 = Y i = 1 n Y 2 i Y i = 1 n θ 1 = θ2 n 2 n Y i Yi 2 Y i = Y 2 i 2( n Y i) 2 n n Yi 2 Y i Y 2 i. We will now consider the two parts of the two-part model separately. First we consider the part of the two-part model which determines whether the observation is zero or positive, i.e. the part corresponding to { the random variable. Consider the sample δ = (δ 1,..., δ n ) defined by δ i := 0,if yi = 0 1,if y i > 0. We have that δ is a sample from i.i.d. random variables i Ber(θ 1 ). The moment estimator of order 1 for θ 1 can be determined as follows. E[ ] = θ 1 = 1 n i The moment estimator ˆθ 1,MME for θ 1 is given by ˆθ 1,MME ( ) = 1 i. n Now consider the second part of the two-part model. Again, letting z = (z 1,..., z n r ) denote subsample of positive observations of y, we get that Z Exp(θ 2 ). The first moment estimator for θ 2 is given by E[Z] = 1/θ 2 = 1 n r θ 2 = n r/( n r n r Z i Z i ) The moment estimator of order 1 for θ 2 is given by ˆθ 2,MME (Z) = n r. n r Z i 22

26 Remark We see that the moment estimators for the separate parts of the models are not the same as the moment estimators for the joint model. When deriving moment estimates for the joint model, the two parts of the two-part model may therefore not be considered separately. 5.4 Exponential Family Let Y P Y {P θ, θ = (θ 1, θ 2 )} be distributed according to the two-part model, i.e. let Y = Z with and Z being independent, Ber(θ 1 ), Z P Z {P θ2 } and P (Z > 0) = 1. Let θ 1 Θ 1 and θ 2 Θ 2. Moreover, let y be a random sample of size n from independent and identically distributed random variables Y i. Let r denote the number of zero observations in the sample. Theorem If Θ 1 = [0, 1], then P Y {P θ } does not form an exponential family. Proof. Consider the probability measures P (1,α1), P (0,α2) and P (β,α3) with α 1, α 2, α 3 Θ 2 and β (0; 1). We have that P (1,α1)(R {0}) = 0 but P (β,α3)(r {0}) > 0 since P (β,α3)(0) = 1 β > 0. Moreover, P (0,α2)(R\{0}) = 0 but P (β,α3)(r\{0}) > 0. Therefore it follows that P Y does not form an exponential family. Theorem If P Z {P θ2 } is not an exponential family, then P Y {P θ } does not form an exponential family. Proof. We have that the likelihood L(θ 1, θ 2, y) of the sample y is of the form L(θ 1, θ 2, y) = L 1 (θ 1, y)l 2 (θ 2, y) with L 1 (θ 1, y) = (1 θ 1 ) r θ n r 1 and L 2 (θ 2, y) = y i>0 f(y i y i > 0) = n r f(z i ) = L(θ 2, z). Since P Z {P θ2 } is not an exponential family, L(θ 2, z) is not of the form (4.4). Thus L(θ 1, θ 2, y) is not of the form (4.4) either. This implies that P Y {P θ } does not form an exponential family. Theorem Let P Z {P θ2 } form a k-parameter exponential family with natural parameters ζ j (θ 2 ) and sufficient statistics T j (z), j = 1,..., k. Moreover, let Θ 1 = (0, 1). Then P Y {P θ } with θ = (θ 1, θ 2 ) forms a k+1 parameter exponential family with natural parameters ζ j (θ 2 ), j = 1,..., k and ζ k+1 (θ 1 ) = ln( 1 θ1 θ 1 ) and sufficient statistics T j (y), j = 1,..., k and T k+1 (y) = r. 23

27 Proof. We have that L(θ 1, θ 2, y) = L 1 (θ 1, y)l 2 (θ 2, y) with L 1 (θ 1, y) = (1 θ 1 ) r θ1 n r and L 2 (θ 2, y) = f(y i y i > 0) = n r f(z i ) = L(θ 2, z). y i>0 Since P Z {P θ2 } is a k-parameter exponential family, ( k ) L(θ 2, z) = L 2 (θ 2, y) = A(θ 2 ) ζ j (θ 2 )T j (z) h(z). j=1 Moreover, we have that L 1 (θ 1, y) = (1 θ 1 ) r θ n r 1 = θ n 1 exp(r ln( 1 θ1 θ 1 )). It follows that ( L(θ 1, θ 2, y) = θ1 n A(θ 2 ) exp r ln( 1 θ1 θ 1 ) + k which yields the above result. 5.5 Hypothesis Testing j=1 ) ζ j (θ 2 )T j (y)) h(y) We assume the following two-part model for the sample y = (y 1,..., y n ) of observations from independent and identically distributed random variables Y i Y i = 1 {1} ( i )Z i, where i Ber(θ 1 ), θ 1 (0, 1) and Z i Exp(θ 2 ), θ 2 R + \{0} are independent random variables. Thus Y i P θ where θ = (θ 1, θ 2 ). Note that P θ belongs to an exponential family. Let r denote the number of zero observations in the sample y. is an observation from the random variable R Bin(n, 1 θ 1 ) Note that r Neyman-Pearson tests for simple hypotheses and known θ 2 Theorem The Neyman-Pearson test of size α for the testing problem is given by H 0 : Y i P 0 i.e. θ 1 = α 1, θ 2 = β H 1 : Y i P 1 i.e. θ 1 = α 2, θ 2 = β 1, R < c α P φ(y) = 0(R<c) P 0(R c), R = c 0, R > c The value of c is given by the solution to P 0 (R < c) = α or, if such c doesn t exist, P 0 (R < c) < α < P 0 (R c). Note that R Bin(n, 1 α 1 ) under H 0. Proof. We want to test H 0 : Y i P 0 i.e. θ 1 = α 1, θ 2 = β H 1 : Y i P 1 i.e. θ 1 = α 2, θ 2 = β. The Neyman-Pearson tests are of the form 24

28 1, p 0 (y) < kp 1 (y) φ(y) = γ, p 0 (y) = kp 1 (y) 0, p 0 (y) > kp 1 (y) We have that p 0 (y) = (1 α 1 ) r α1 n r β n r e β p 1 (y) = (1 α 2 ) r α2 n r β n r e β y i y i To obtain a test of size α k is chosen so that P 0 (p 0 (Y ) < kp 1 (Y )) = α or, if such k doesn t exist, so that P 0 (p 0 (Y ) < kp 1 (Y )) < α < P 0 (p 0 (Y ) kp 1 (Y )). P 0 (p 0 (Y ) < kp 1 (Y )) = P 0 ( p0 (Y ) p 1 (Y ) < k ) ( (1 α1 ) r α1 n r = P 0 (1 α 2 ) r α2 n r ) < k = = P 0 ( ( (1 α 1 )α 2 (1 α 2 )α 1 ) r < k (α 2 α 1 ) n ) = P 0 ( r ln( (1 α 1)α 2 (1 α 2 )α 1 ) < ln(k ( α 2 α 1 ) n) ) = ( = P 0 r < ln(k ( α 2 ) n)/ (1 α 1 )α ) 2 ln( ) α 1 (1 α 2 )α 1 Let c = ln(k ( α 2 α 1 ) n)/ ln( (1 α 1)α 2 (1 α 2)α 1 ). Under H 0, r is an observation of the random variable R Bin(n, 1 α 1 ). We have that P 0 (p 0 (Y ) < kp 1 (Y )) = P 0 (R < c) Similarly, P 0 (p 0 (Y ) kp 1 (Y )) = P 0 (R c). The value for c such that either P 0 (p 0 (Y ) < kp 1 (Y )) = P 0 (R < c) = α or P 0 (p 0 (Y ) < kp 1 (Y )) = P 0 (R < c) < α < P 0 (p 0 (Y ) kp 1 (Y )) = P 0 (R c) can easily be computed. From this we can derive γ = α P 0(R < c). P 0 (R c) Finally, k can be obtained through k = ( (1 α1)α2 (1 α 2)α 1 ) c n ( α1 α 2 ) n. Remark The Neyman-Pearson test is the most powerful test of size α for the above testing problem [06]. 25

29 References [01] Social Insurance in Figures 2014 (2014). Försäkringskassan. [02] Duan N., Manning W. G., Morris C. N. and J. P. Newhouse (1983). A Comparison of Alternative Models of the Demand for Medical Care. Journal of Business Economics and Statistics 1, pp [03] Tobin J. (1958). Estimation of Relationships for Limited Dependent Variables. Econometrica, Vol. 26, pp [04] Maren K. Olsen, Joseph L. Schafer. A Class of Models for Semicontinuous Longitudinal Data. [05] Yongyi Min, Agresti, A. (2002). Modeling Nonnegative Data with Clumping at Zero: A Survey. JIRSS, Vol. 1, Nos. 1-2, pp [06] Liero H., Zwanzig S. (2012). Introduction to the Theory of Statistical Inference. [07] Rydén J. (2014). A Statistical Analysis of Trends for Warm and Cold Spells in Uppsala by Means of Counts. Geografiska Annaler: Series A, Physical Geography [08] Lambert D. (1992). Zero-Inflated Poisson Regression, With an Application to Defects in Manufacturing. Technometrics, Vol. 34, No. 1, pp [09] Jie Gao (2007). Modeling Individual Healthcare Expenditures by Extending the Two-part Model, pp [10] Aitchison J. (1955). On the Distribution of a Positive Random Variable Having a Discrete Probability Mass at the Origin. Journal of the American Statistical Association, Vol.50, No. 271, pp [11] Heckman, J. (1979). Sample Selection Bias as a Specification Error. Econometrica, Vol. 47, pp

Truncation and Censoring

Truncation and Censoring Truncation and Censoring Laura Magazzini laura.magazzini@univr.it Laura Magazzini (@univr.it) Truncation and Censoring 1 / 35 Truncation and censoring Truncation: sample data are drawn from a subset of

More information

NELS 88. Latent Response Variable Formulation Versus Probability Curve Formulation

NELS 88. Latent Response Variable Formulation Versus Probability Curve Formulation NELS 88 Table 2.3 Adjusted odds ratios of eighth-grade students in 988 performing below basic levels of reading and mathematics in 988 and dropping out of school, 988 to 990, by basic demographics Variable

More information

Maximum Likelihood Methods

Maximum Likelihood Methods Maximum Likelihood Methods Some of the models used in econometrics specify the complete probability distribution of the outcomes of interest rather than just a regression function. Sometimes this is because

More information

Lecture 8. Poisson models for counts

Lecture 8. Poisson models for counts Lecture 8. Poisson models for counts Jesper Rydén Department of Mathematics, Uppsala University jesper.ryden@math.uu.se Statistical Risk Analysis Spring 2014 Absolute risks The failure intensity λ(t) describes

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

SCHOOL OF MATHEMATICS AND STATISTICS. Linear and Generalised Linear Models

SCHOOL OF MATHEMATICS AND STATISTICS. Linear and Generalised Linear Models SCHOOL OF MATHEMATICS AND STATISTICS Linear and Generalised Linear Models Autumn Semester 2017 18 2 hours Attempt all the questions. The allocation of marks is shown in brackets. RESTRICTED OPEN BOOK EXAMINATION

More information

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or

More information

Tobit and Selection Models

Tobit and Selection Models Tobit and Selection Models Class Notes Manuel Arellano November 24, 2008 Censored Regression Illustration : Top-coding in wages Suppose Y log wages) are subject to top coding as is often the case with

More information

Lecture 01: Introduction

Lecture 01: Introduction Lecture 01: Introduction Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology Medical University of South Carolina Lecture 01: Introduction

More information

Linear Models and Estimation by Least Squares

Linear Models and Estimation by Least Squares Linear Models and Estimation by Least Squares Jin-Lung Lin 1 Introduction Causal relation investigation lies in the heart of economics. Effect (Dependent variable) cause (Independent variable) Example:

More information

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain 0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher

More information

Generalized Linear Models Introduction

Generalized Linear Models Introduction Generalized Linear Models Introduction Statistics 135 Autumn 2005 Copyright c 2005 by Mark E. Irwin Generalized Linear Models For many problems, standard linear regression approaches don t work. Sometimes,

More information

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

Econometrics Lecture 5: Limited Dependent Variable Models: Logit and Probit

Econometrics Lecture 5: Limited Dependent Variable Models: Logit and Probit Econometrics Lecture 5: Limited Dependent Variable Models: Logit and Probit R. G. Pierse 1 Introduction In lecture 5 of last semester s course, we looked at the reasons for including dichotomous variables

More information

Mathematical statistics

Mathematical statistics October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation

More information

Hypothesis testing: theory and methods

Hypothesis testing: theory and methods Statistical Methods Warsaw School of Economics November 3, 2017 Statistical hypothesis is the name of any conjecture about unknown parameters of a population distribution. The hypothesis should be verifiable

More information

Some General Types of Tests

Some General Types of Tests Some General Types of Tests We may not be able to find a UMP or UMPU test in a given situation. In that case, we may use test of some general class of tests that often have good asymptotic properties.

More information

SOLUTION FOR HOMEWORK 8, STAT 4372

SOLUTION FOR HOMEWORK 8, STAT 4372 SOLUTION FOR HOMEWORK 8, STAT 4372 Welcome to your 8th homework. Here you have an opportunity to solve classical estimation problems which are the must to solve on the exam due to their simplicity. 1.

More information

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Instructions This exam has 7 pages in total, numbered 1 to 7. Make sure your exam has all the pages. This exam will be 2 hours

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

Statistics 135 Fall 2008 Final Exam

Statistics 135 Fall 2008 Final Exam Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations

More information

Mohammed. Research in Pharmacoepidemiology National School of Pharmacy, University of Otago

Mohammed. Research in Pharmacoepidemiology National School of Pharmacy, University of Otago Mohammed Research in Pharmacoepidemiology (RIPE) @ National School of Pharmacy, University of Otago What is zero inflation? Suppose you want to study hippos and the effect of habitat variables on their

More information

Lecture 12: Application of Maximum Likelihood Estimation:Truncation, Censoring, and Corner Solutions

Lecture 12: Application of Maximum Likelihood Estimation:Truncation, Censoring, and Corner Solutions Econ 513, USC, Department of Economics Lecture 12: Application of Maximum Likelihood Estimation:Truncation, Censoring, and Corner Solutions I Introduction Here we look at a set of complications with the

More information

A Guide to Modern Econometric:

A Guide to Modern Econometric: A Guide to Modern Econometric: 4th edition Marno Verbeek Rotterdam School of Management, Erasmus University, Rotterdam B 379887 )WILEY A John Wiley & Sons, Ltd., Publication Contents Preface xiii 1 Introduction

More information

Solutions to the Spring 2015 CAS Exam ST

Solutions to the Spring 2015 CAS Exam ST Solutions to the Spring 2015 CAS Exam ST (updated to include the CAS Final Answer Key of July 15) There were 25 questions in total, of equal value, on this 2.5 hour exam. There was a 10 minute reading

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process

Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Applied Mathematical Sciences, Vol. 4, 2010, no. 62, 3083-3093 Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Julia Bondarenko Helmut-Schmidt University Hamburg University

More information

Parameter Estimation

Parameter Estimation Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y

More information

Deal with Excess Zeros in the Discrete Dependent Variable, the Number of Homicide in Chicago Census Tract

Deal with Excess Zeros in the Discrete Dependent Variable, the Number of Homicide in Chicago Census Tract Deal with Excess Zeros in the Discrete Dependent Variable, the Number of Homicide in Chicago Census Tract Gideon D. Bahn 1, Raymond Massenburg 2 1 Loyola University Chicago, 1 E. Pearson Rm 460 Chicago,

More information

Selection endogenous dummy ordered probit, and selection endogenous dummy dynamic ordered probit models

Selection endogenous dummy ordered probit, and selection endogenous dummy dynamic ordered probit models Selection endogenous dummy ordered probit, and selection endogenous dummy dynamic ordered probit models Massimiliano Bratti & Alfonso Miranda In many fields of applied work researchers need to model an

More information

Introduction to Regression Analysis. Dr. Devlina Chatterjee 11 th August, 2017

Introduction to Regression Analysis. Dr. Devlina Chatterjee 11 th August, 2017 Introduction to Regression Analysis Dr. Devlina Chatterjee 11 th August, 2017 What is regression analysis? Regression analysis is a statistical technique for studying linear relationships. One dependent

More information

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer

More information

Peaks-Over-Threshold Modelling of Environmental Data

Peaks-Over-Threshold Modelling of Environmental Data U.U.D.M. Project Report 2014:33 Peaks-Over-Threshold Modelling of Environmental Data Esther Bommier Examensarbete i matematik, 30 hp Handledare och examinator: Jesper Rydén September 2014 Department of

More information

Investigation of an Automated Approach to Threshold Selection for Generalized Pareto

Investigation of an Automated Approach to Threshold Selection for Generalized Pareto Investigation of an Automated Approach to Threshold Selection for Generalized Pareto Kate R. Saunders Supervisors: Peter Taylor & David Karoly University of Melbourne April 8, 2015 Outline 1 Extreme Value

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Generalized Linear Models (1/29/13)

Generalized Linear Models (1/29/13) STA613/CBB540: Statistical methods in computational biology Generalized Linear Models (1/29/13) Lecturer: Barbara Engelhardt Scribe: Yangxiaolu Cao When processing discrete data, two commonly used probability

More information

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45 Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS 21 June 2010 9:45 11:45 Answer any FOUR of the questions. University-approved

More information

STA 260: Statistics and Probability II

STA 260: Statistics and Probability II Al Nosedal. University of Toronto. Winter 2017 1 Properties of Point Estimators and Methods of Estimation 2 3 If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 71. Decide in each case whether the hypothesis is simple

More information

A nonparametric test for path dependence in discrete panel data

A nonparametric test for path dependence in discrete panel data A nonparametric test for path dependence in discrete panel data Maximilian Kasy Department of Economics, University of California - Los Angeles, 8283 Bunche Hall, Mail Stop: 147703, Los Angeles, CA 90095,

More information

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015 STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis

More information

Parametric Modelling of Over-dispersed Count Data. Part III / MMath (Applied Statistics) 1

Parametric Modelling of Over-dispersed Count Data. Part III / MMath (Applied Statistics) 1 Parametric Modelling of Over-dispersed Count Data Part III / MMath (Applied Statistics) 1 Introduction Poisson regression is the de facto approach for handling count data What happens then when Poisson

More information

Chapter 2. Discrete Distributions

Chapter 2. Discrete Distributions Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science

UNIVERSITY OF TORONTO Faculty of Arts and Science UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator

More information

Economics 536 Lecture 21 Counts, Tobit, Sample Selection, and Truncation

Economics 536 Lecture 21 Counts, Tobit, Sample Selection, and Truncation University of Illinois Fall 2016 Department of Economics Roger Koenker Economics 536 Lecture 21 Counts, Tobit, Sample Selection, and Truncation The simplest of this general class of models is Tobin s (1958)

More information

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1, Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Econometrics Master in Business and Quantitative Methods

Econometrics Master in Business and Quantitative Methods Econometrics Master in Business and Quantitative Methods Helena Veiga Universidad Carlos III de Madrid This chapter deals with truncation and censoring. Truncation occurs when the sample data are drawn

More information

Theory of Statistical Tests

Theory of Statistical Tests Ch 9. Theory of Statistical Tests 9.1 Certain Best Tests How to construct good testing. For simple hypothesis H 0 : θ = θ, H 1 : θ = θ, Page 1 of 100 where Θ = {θ, θ } 1. Define the best test for H 0 H

More information

2.6.3 Generalized likelihood ratio tests

2.6.3 Generalized likelihood ratio tests 26 HYPOTHESIS TESTING 113 263 Generalized likelihood ratio tests When a UMP test does not exist, we usually use a generalized likelihood ratio test to verify H 0 : θ Θ against H 1 : θ Θ\Θ It can be used

More information

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX Term Test 3 December 5, 2003 Name Math 52 Student Number Direction: This test is worth 250 points and each problem worth 4 points DO ANY SIX PROBLEMS You are required to complete this test within 50 minutes

More information

Endogenous Treatment Effects for Count Data Models with Endogenous Participation or Sample Selection

Endogenous Treatment Effects for Count Data Models with Endogenous Participation or Sample Selection Endogenous Treatment Effects for Count Data Models with Endogenous Participation or Sample Selection Massimilano Bratti & Alfonso Miranda Institute of Education University of London c Bratti&Miranda (p.

More information

Statistics II Lesson 1. Inference on one population. Year 2009/10

Statistics II Lesson 1. Inference on one population. Year 2009/10 Statistics II Lesson 1. Inference on one population Year 2009/10 Lesson 1. Inference on one population Contents Introduction to inference Point estimators The estimation of the mean and variance Estimating

More information

Mathematics Qualifying Examination January 2015 STAT Mathematical Statistics

Mathematics Qualifying Examination January 2015 STAT Mathematical Statistics Mathematics Qualifying Examination January 2015 STAT 52800 - Mathematical Statistics NOTE: Answer all questions completely and justify your derivations and steps. A calculator and statistical tables (normal,

More information

UNIVERSITÄT POTSDAM Institut für Mathematik

UNIVERSITÄT POTSDAM Institut für Mathematik UNIVERSITÄT POTSDAM Institut für Mathematik Testing the Acceleration Function in Life Time Models Hannelore Liero Matthias Liero Mathematische Statistik und Wahrscheinlichkeitstheorie Universität Potsdam

More information

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently

More information

Regression - Modeling a response

Regression - Modeling a response Regression - Modeling a response We often wish to construct a model to Explain the association between two or more variables Predict the outcome of a variable given values of other variables. Regression

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

MATH20802: STATISTICAL METHODS EXAMPLES

MATH20802: STATISTICAL METHODS EXAMPLES MATH20802: STATISTICAL METHODS EXAMPLES 1 1. If X N(µ, σ 2 ) show that its mgf is M X (t) = exp ( µt + σ2 t 2 2 2. If X 1 N(µ 1, σ 2 1 ) and X 2 N(µ 2, σ 2 2 ) are independent then show that ax 1 + bx

More information

Lecture 10: Alternatives to OLS with limited dependent variables. PEA vs APE Logit/Probit Poisson

Lecture 10: Alternatives to OLS with limited dependent variables. PEA vs APE Logit/Probit Poisson Lecture 10: Alternatives to OLS with limited dependent variables PEA vs APE Logit/Probit Poisson PEA vs APE PEA: partial effect at the average The effect of some x on y for a hypothetical case with sample

More information

Basic Concepts of Inference

Basic Concepts of Inference Basic Concepts of Inference Corresponds to Chapter 6 of Tamhane and Dunlop Slides prepared by Elizabeth Newton (MIT) with some slides by Jacqueline Telford (Johns Hopkins University) and Roy Welsch (MIT).

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

Agricultural and Applied Economics 637 Applied Econometrics II. Assignment III Maximum Likelihood Estimation (Due: March 25, 2014)

Agricultural and Applied Economics 637 Applied Econometrics II. Assignment III Maximum Likelihood Estimation (Due: March 25, 2014) Agricultural and Applied Economics 637 Applied Econometrics II Assignment III Maximum Likelihood Estimation (Due: March 5, 014) In this assignment I would like you to extend some of the theoretical Maximum

More information

Hidden Markov models for time series of counts with excess zeros

Hidden Markov models for time series of counts with excess zeros Hidden Markov models for time series of counts with excess zeros Madalina Olteanu and James Ridgway University Paris 1 Pantheon-Sorbonne - SAMM, EA4543 90 Rue de Tolbiac, 75013 Paris - France Abstract.

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

A simple bivariate count data regression model. Abstract

A simple bivariate count data regression model. Abstract A simple bivariate count data regression model Shiferaw Gurmu Georgia State University John Elder North Dakota State University Abstract This paper develops a simple bivariate count data regression model

More information

Graduate Econometrics I: Maximum Likelihood I

Graduate Econometrics I: Maximum Likelihood I Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood

More information

Lecture 14: Introduction to Poisson Regression

Lecture 14: Introduction to Poisson Regression Lecture 14: Introduction to Poisson Regression Ani Manichaikul amanicha@jhsph.edu 8 May 2007 1 / 52 Overview Modelling counts Contingency tables Poisson regression models 2 / 52 Modelling counts I Why

More information

Modelling counts. Lecture 14: Introduction to Poisson Regression. Overview

Modelling counts. Lecture 14: Introduction to Poisson Regression. Overview Modelling counts I Lecture 14: Introduction to Poisson Regression Ani Manichaikul amanicha@jhsph.edu Why count data? Number of traffic accidents per day Mortality counts in a given neighborhood, per week

More information

Inference and Regression

Inference and Regression Name Inference and Regression Final Examination, 2015 Department of IOMS This course and this examination are governed by the Stern Honor Code. Instructions Please write your name at the top of this page.

More information

Chapter 7. Hypothesis Testing

Chapter 7. Hypothesis Testing Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function

More information

Greene, Econometric Analysis (6th ed, 2008)

Greene, Econometric Analysis (6th ed, 2008) EC771: Econometrics, Spring 2010 Greene, Econometric Analysis (6th ed, 2008) Chapter 17: Maximum Likelihood Estimation The preferred estimator in a wide variety of econometric settings is that derived

More information

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn!

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Questions?! C. Porciani! Estimation & forecasting! 2! Cosmological parameters! A branch of modern cosmological research focuses

More information

Fractional Imputation in Survey Sampling: A Comparative Review

Fractional Imputation in Survey Sampling: A Comparative Review Fractional Imputation in Survey Sampling: A Comparative Review Shu Yang Jae-Kwang Kim Iowa State University Joint Statistical Meetings, August 2015 Outline Introduction Fractional imputation Features Numerical

More information

TMA 4275 Lifetime Analysis June 2004 Solution

TMA 4275 Lifetime Analysis June 2004 Solution TMA 4275 Lifetime Analysis June 2004 Solution Problem 1 a) Observation of the outcome is censored, if the time of the outcome is not known exactly and only the last time when it was observed being intact,

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

STA216: Generalized Linear Models. Lecture 1. Review and Introduction

STA216: Generalized Linear Models. Lecture 1. Review and Introduction STA216: Generalized Linear Models Lecture 1. Review and Introduction Let y 1,..., y n denote n independent observations on a response Treat y i as a realization of a random variable Y i In the general

More information

Ch. 5 Hypothesis Testing

Ch. 5 Hypothesis Testing Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,

More information

Nuisance parameters and their treatment

Nuisance parameters and their treatment BS2 Statistical Inference, Lecture 2, Hilary Term 2008 April 2, 2008 Ancillarity Inference principles Completeness A statistic A = a(x ) is said to be ancillary if (i) The distribution of A does not depend

More information

Stat 579: Generalized Linear Models and Extensions

Stat 579: Generalized Linear Models and Extensions Stat 579: Generalized Linear Models and Extensions Mixed models Yan Lu March, 2018, week 8 1 / 32 Restricted Maximum Likelihood (REML) REML: uses a likelihood function calculated from the transformed set

More information

BTRY 4090: Spring 2009 Theory of Statistics

BTRY 4090: Spring 2009 Theory of Statistics BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)

More information

Estimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1

Estimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1 Estimation September 24, 2018 STAT 151 Class 6 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0 1 1 1

More information

New Bayesian methods for model comparison

New Bayesian methods for model comparison Back to the future New Bayesian methods for model comparison Murray Aitkin murray.aitkin@unimelb.edu.au Department of Mathematics and Statistics The University of Melbourne Australia Bayesian Model Comparison

More information

Diploma Part 2. Quantitative Methods. Examiners Suggested Answers

Diploma Part 2. Quantitative Methods. Examiners Suggested Answers Diploma Part 2 Quantitative Methods Examiners Suggested Answers Q1 (a) A frequency distribution is a table or graph (i.e. a histogram) that shows the total number of measurements that fall in each of a

More information

SPRING 2007 EXAM C SOLUTIONS

SPRING 2007 EXAM C SOLUTIONS SPRING 007 EXAM C SOLUTIONS Question #1 The data are already shifted (have had the policy limit and the deductible of 50 applied). The two 350 payments are censored. Thus the likelihood function is L =

More information

Review of Discrete Probability (contd.)

Review of Discrete Probability (contd.) Stat 504, Lecture 2 1 Review of Discrete Probability (contd.) Overview of probability and inference Probability Data generating process Observed data Inference The basic problem we study in probability:

More information

MC3: Econometric Theory and Methods. Course Notes 4

MC3: Econometric Theory and Methods. Course Notes 4 University College London Department of Economics M.Sc. in Economics MC3: Econometric Theory and Methods Course Notes 4 Notes on maximum likelihood methods Andrew Chesher 25/0/2005 Course Notes 4, Andrew

More information

ST495: Survival Analysis: Hypothesis testing and confidence intervals

ST495: Survival Analysis: Hypothesis testing and confidence intervals ST495: Survival Analysis: Hypothesis testing and confidence intervals Eric B. Laber Department of Statistics, North Carolina State University April 3, 2014 I remember that one fateful day when Coach took

More information

Linear Regression With Special Variables

Linear Regression With Special Variables Linear Regression With Special Variables Junhui Qian December 21, 2014 Outline Standardized Scores Quadratic Terms Interaction Terms Binary Explanatory Variables Binary Choice Models Standardized Scores:

More information

Two-stage Adaptive Randomization for Delayed Response in Clinical Trials

Two-stage Adaptive Randomization for Delayed Response in Clinical Trials Two-stage Adaptive Randomization for Delayed Response in Clinical Trials Guosheng Yin Department of Statistics and Actuarial Science The University of Hong Kong Joint work with J. Xu PSI and RSS Journal

More information

Bivariate Relationships Between Variables

Bivariate Relationships Between Variables Bivariate Relationships Between Variables BUS 735: Business Decision Making and Research 1 Goals Specific goals: Detect relationships between variables. Be able to prescribe appropriate statistical methods

More information

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics A short review of the principles of mathematical statistics (or, what you should have learned in EC 151).

More information

Qualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf

Qualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 1: Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section

More information