Bayesian Estimation of Small Proportions Using Binomial Group Test

Size: px
Start display at page:

Download "Bayesian Estimation of Small Proportions Using Binomial Group Test"

Transcription

1 Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School Bayesian Estimation of Small Proportions Using Binomial Group Test Shihua Luo Florida International University, DOI: /etd.FI Follow this and additional works at: Recommended Citation Luo, Shihua, "Bayesian Estimation of Small Proportions Using Binomial Group Test" (212). FIU Electronic Theses and Dissertations This work is brought to you for free and open access by the University Graduate School at FIU Digital Commons. It has been accepted for inclusion in FIU Electronic Theses and Dissertations by an authorized administrator of FIU Digital Commons. For more information, please contact

2 FLORIDA INTERNATIONAL UNIVERSITY Miami, Florida BAYESIAN ESTIMATION OF SMALL PROPORTIONS USING BINOMIAL GROUP TEST A thesis submitted in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE in STATISTICS by Shihua Luo 212

3 To: Dean Kenneth Furton College of Arts and Sciences This thesis, written by Shihua Luo, and entitled Bayesian Estimation of Small Proportions Using Binomial Group Test, having been approved in respect to style and intellectual content, is referred to you for judgement. We have read this thesis and recommend that it be approved. Florence George Kai Huang, Co-Major Professor Jie Mi, Co-Major Professor Date of Defense: November 9, 212 The thesis of Shihua Luo is approved. Dean Kenneth Furton College of Arts and Sciences Dean Lakshmi N. Reddi University Graduate School Florida International University, 212 ii

4 ACKNOWLEDGMENTS First of all, I would like to express my sincere thanks to my major professor, Dr. Jie Mi for his patient guidance, enthusiasm, encouragement and friendship throughout this whole study. I couldn t finish my research without his great support. I would like to express deepest appreciation to my co-major professor, Dr. Kai Huang, for his technical support and patient guidance in Latex and Matlab. I would also like to thank the member of my committee, Dr. Florence George for her time, valuable advice and great encouragement. In addition, I would like to thank the Department of Mathematics and Statistics, all the professors who supported and encouraged me throughout my life in FIU. I wish to thank aunt Miao, Dr. Jie Mi s wife. She gave me substantial big help in my life. I also thank my classmates, friends. I am so happy I have them around me. Finally, I wish to thank my family for their strong support and their absolute confidence in me. iii

5 ABSTRACT OF THE THESIS BAYSIAN ESTIMATION OF SMALL PROPORTIONS USING BINOMIAL GROUP TEST by Shihua Luo Florida International University, 212 Miami, Florida Professor Jie Mi, Co-Major Professor Professor Kai Huang, Co-Major Professor Group testing has long been considered as a safe and sensible relative to oneat-a-time testing in applications where the prevalence rate p is small. In this thesis, we applied Bayes approach to estimate p using Beta-type prior distribution. First, we showed two Bayes estimators of p from prior on p derived from two different loss functions. Second, we presented two more Bayes estimators of p from prior on π according to two loss functions. We also displayed credible and HPD interval for p. In addition, we did intensive numerical studies. All results showed that the Bayes estimator was preferred over the usual maximum likelihood estimator (MLE) for small p. We also presented the optimal β for different p,m, and k. Keywords: Group Test, Bayes Estimator, Beta Distribution, MLE, MSE. iv

6 TABLE OF CONTENTS CHAPTER PAGE 1 Introduction Bayes Inferences from Prior on p Bayes Point Estimators of p Approximation to Bayes Estimators of p Comparison in the Special Case of Y= Credible and HPD Interval for p Bayes Estimators of p from Prior on π Numerical Studies Conclusion REFERENCES APPENDIX v

7 LIST OF TABLES TABLE PAGE 1 Optimal β For k = 5, α = 1, N = Optimal β For k = 1, α = 1, N = Optimal β For k = 15, α = 1, N = MSE For p =.5, k = 5, m = 5, α = 1, N = MSE For p =.1, k = 5, m = 5, α = 1, N = MSE For p =.5, k = 5, m = 15, α = 1, N = MSE For p =.1, k = 5, m = 15, α = 1, N = MSE For p =.5,.6,.7,.8, k = 5, α = 1, N = MSE For p =.9,.1,.2,.3, k = 5, α = 1, N = MSE For p =.4,.5,.6,.7, k = 5, α = 1, N = MSE For p =.8,.9,.1, k = 5, α = 1, N = MSE For p =.5,.6,.7,.8, k = 1, α = 1, N = MSE For p =.9,.1,.2,.3, k = 1, α = 1, N = MSE For p =.4,.5,.6,.7, k = 1, α = 1, N = MSE For p =.8,.9,.1, k = 1, α = 1, N = MSE For p =.5,.6,.7,.8, k = 15, α = 1, N = MSE For p =.9,.1,.2,.3, k = 15, α = 1, N = MSE For p =.4,.5,.6,.7, k = 15, α = 1, N = MSE For p =.8,.9,.1, k = 15, α = 1, N = vi

8 LIST OF FIGURES FIGURE PAGE 1 Optimal β for k = 5, m = Optimal β for k = 5, m = Optimal β for k = 5, m = 1 or Optimal β for k = 1, m = Optimal β for k = 5 or 1, m = vii

9 1 Introduction Almost daily, the news media report the results of some poll. Pollsters often give some data, for example, the percentage of people in favor of the current president, the rate of people who smoke cigarettes, the death rate of a certain disease, the percent of people who have hepatitis C virus or a rare disease, the rate of contaminated fish in a river, the proportion of defective products, et cetra. All of these issues have a common property the ratio of a part to a whole. How can we estimate this proportion p? In estimating the proportion of subjects that have a certain characteristic in a population in which each individual can be classified into two categories, say with or without certain characteristic, one logical method of estimating p for the population is to use the proportion of successes meaning with that characteristic in the sample. For example, how would we estimate the true fraction of all U.S. citizens who trust the president? We can estimate the p by calculating ˆp=x/n: where x is the number of people in the sample who trust the president and n is the number of people sampled. Experimenters used this method which is called the maximum likelihood estimator (MLE) for many years. But this method has some disadvantages when p is very small. For instance, if researchers want to estimate the prevalence of human immunodeficiency virus (HIV), the ratio ˆp is extremely small and so the estimation is not stable in using the MLE method. Then we need to test every sample unit to get the estimate of ratio p. In addition, the examination of the individual members of a large sample is an expensive and tedious work. To overcome these shortcomings, Dorfman (1943) first proposed the method of group testing and applied it to testing 1

10 blood samples for the syphilis antigen. The idea of group testing is as follows. Assuming that a random sample of size n is drawn from a target population in which some units have certain characteristic that we are interested in, then we divide this sample into m groups with k units in each group. It is assumed that if at least one unit in a group has that characteristic (positive), then this group will be positive; on the other hand, if none of the units in a group is positive, then this group will be negative. The key assumption is valid for many situations in biological, environmental and medical studies. Let p denote the probability of a unit being positive in a certain population, then the probability of a group being positive, denoted as π, can be expressed in terms of p, π = 1 (1 p) k. Instead of directly estimating p from examining each unit in a sample, group testing will estimate π and then p. Instead of conducting N chemical analyses are required to check out all members of a population of size N, Dorfman (1943) applied group testing for checking the syphilis antigen. Suppose that after the blood serum is drawn for each individual, all the blood samples are pooled in groups of five, and the group rather than the individual serum are subjected to chemical analysis. If none of the five sera contains syphilis antigen, the result of this group will be negative. If the result of a group shows positive, it means that this group includes at least one individual of syphilis antigen. Then the individuals making up the pool must be retested to decide which of the members are infected. He also discussed related costs for selected prevalence rates and optimum group sizes. He suggested that the prevalence rate be sufficiently 2

11 small to make worthwhile economics possible. Afterwards, many scientists used group testing in their study fields. Gibbs and Gower (196) applied this method to estimate the frequency of success of attempts to transmit a virus disease from one plant to another. Chiang and Reeves (1962) used this approach to estimate the infection rates of mosquitoes. Bhattacharyya et al.(1979) employed group testing to give point estimates and confidence intervals for infection rates. Worlund and Taylor (1983) applied the procedure to evaluate the disease incidence in a large population of fish. Swallow (1985) used this method to estimate infection rates and probabilities of pathogen transmission by a single vector. To evaluate human immunodeficiency virus (HIV) seroprevalence in population surveys, pooled sera was used by Kline et al. (1989). Rodoni et al. (1994) applied a sequential batch testing procedure combined with Enzyme-Linked Immunosorbent Assay (ELISA) to estimate levels of virus incidence in Victoria cut-flower sim carnation. Hepworth (1996) constructed confidence intervals for proportion of infected unites in a population involving unequaled sized group. Gastwirth (2) proposed that mutations in individual patients could be tested more fruitful by being checked in pools. Xie et al. (21) used group testing to develops models and estimation procedures in order to obtain quantitative information from data in the process to discovery and development of a new drug. Katholi and Unnasch (26) discussed the suitable sampling protocol about important experimental parameters for deciding infection rates. Because of its wide applications, group testing is also known as pooled testing, or composite sampling, in different fields. In recent years, more and more scholars recognized the advantage of the group testing 3

12 experimental design. It is well known that this method is economical, time saving, and efficient particularly when it is applied to estimating small proportion of a disease or some other rare attributes. The traditional way to estimate π or p is using the method of maximum likelihood (MLE). The definition of MLE is as follows. If x 1, x 2,..., x n are independent and identically distributed observation from a population with pdf or pmf f(x; θ), where x = (x 1, x 2,..., x n ) is the fixed sample point, and θ is the unknown parameter of interest, the likelihood function is defined as the product of all the pdfs or pmfs f(x i ; θ). For the sample point x, let θ(x) be a parameter value at which likelihood function attains its maximum as a function of θ, with x held fixed. Then θ(x) is the MLE of θ. An alternative way to estimate π or p is applying the Bayes method. The Bayesian approach is totally different from the approach used in MLE analyse. In classical methods the parameter to be estimated is thought to be an unknown, but fixed quantity. The Bayesian approach treats the unknown parameter as a random variable. In many problems the investigator has some prior information about the unknown parameter. For instance, in sampling inspection, the quality engineer has some idea about the true defective rate, the unknown parameter, of a production process from past experience. The Bayesian approach assumes that this prior knowledge can be summarized in the form of a probability distribution on unknown parameter, called the prior distribution. This is a subjective distribution, based on the experimenter s belief, and is formulated before the data are seen. A sample is then taken 4

13 from a population indexed by the unknown parameter θ and the prior distribution is updated with this sample information. The updated prior distribution is named the posterior distribution. The posterior distribution is now used to make inference about θ. To use the Bayes method, we also need a loss function. The most common loss function is the square loss function. If we use this square loss function, a natural estimator for θ is the mean of the posterior distribution. If necessary, of course, we can use alternative loss function, but the square loss function is the most popular one. Usually, the posterior distribution will become very complex with each added measurement. Whereas, if we choose a conjugate prior, then the posterior distribution will be easier to obtain. A conjugate prior is defined as a prior distribution belonging to some parametric family, for which the resulted posterior distribution also belongs to the same family. This is an important property. Such prior distributions have been identified for some standard distributions f(x θ). When we want Bayes estimator for proportion θ = p, we assume the prior distribution on parameter θ is Beta (α, β). The posterior distribution is also a Beta distribution function which makes the estimation of parameter p much more convenient. Some researchers prefer to use classical Bayesian approaches because the Bayesian method gave a better point estimator of p, especially when p is very small. Gastwirth et al. (1991) suggested classical Bayesian approaches in group testing. Chaubey and Li (1995) researched the difference between MLE and Bayes method for estimation of binomial probability with group testing. They observed that the Bayes methodology 5

14 gave an alternative choices to the experimenter with possible reduction in cost as well as error. Chick (1996) has used this method to do his studies. In the current study, we want to do further study on estimating p by using Bayes methods. We will confine our inference about p to the range (,.1). We will try two Bayesian estimators corresponding to two different loss functions with assumption that p has Beta prior distribution. In addition, we will try to obtain a credible and HPD interval for p. Assuming Beta prior on the probability π of a group being positive, we will investigate two Bayes estimators corresponding to two loss functions. The criterion for comparing the accurate of estimators is its M SE (mean squared error). The performance of the proposed estimators will be studied based on heavy Monte Carlo simulation. The proposed Bayes estimators will also be compared with other estimator existed in the literature. 6

15 2 Bayes Inferences from Prior on p In this section, we will consider Bayes inferences about p. It is assumed that p has Beta prior distribution Beta(α, β) with density function f P (p) = 1 B(α, β) pα 1 (1 p) β 1, < p < 1, α >, β >. 2.1 Bayes Point Estimators of p Suppose that a random sample of size n = mk is obtained from a population of screened subjects, where m, k are positive integers. We assume that p is the probability that a randomly selected individual s test is positive. In this research our concern is the inference about small p, say < p.1. The n subjects are classified into m groups each of which includes k subjects. Throughout this research it is assumed that any one of the m groups is tested as positive if and only if it includes at least one positive subject. Let Y be the number of positive groups. It is known that Y B(m, π) where π = 1 (1 p) k, i.e., P (Y = y) = ( m y ) π y (1 π) m y. Clearly, the joint distribution of (Y, p) is f Y,P (y, p) = = ( ) 1 m B(α, β) pα 1 (1 p) β 1 π y (1 π) m y y ( m ) y B(α, β) pα 1 (1 p) β 1( 1 (1 p) k) y (1 p) k(m y) (2.1) Thus the marginal distribution of Y is given as f Y (y) = f Y,P (y, p)dp 7

16 ( m y ) = p α 1 (1 p) k(m y)+β 1( 1 (1 p) k) y dp B(α, β) ( m ) y ( ) y y = p α 1 (1 p) k(m y)+β 1 ( 1) j (1 p) kj dp B(α, β) j ( ) m (B(α, ) 1 y ( ) y 1 = β) ( 1) j p α 1 (1 p) k(m+j y)+β 1 dp y j ( ) m (B(α, ) 1 y ( ) y = β) ( 1) j B ( α, k(m + j y) + β ) (2.2) y j Therefore, the posterior distribution of p is f P Y (p y) = f Y,P (y, p) f Y (y) ( m )( ) 1p y B(α, β) α 1 (1 p) k(m y)+β 1( 1 (1 p) k) y = ( m )( ) 1 y y B(α, β) ( 1) j( ) ( ) y B α, k(m + j y) + β = pα 1 (1 p) k(m y)+β 1( 1 (1 p) k) y y ( 1) j( ) ( ) y B α, k(m + j y) + β j j (2.3) The Bayes estimator of p for the case of square loss function L 1 (p, a) = (p a) 2 is derived by Chaubey and Li ( 1995) as y ( y ) j ( 1) j B(α + 1, km + kj ky + β) ˆp B1 = y ( y ) ( 1)j B(α, km + kj ky + β) j. (2.4) Specifically, if y =, then ˆp B1 = α n + α + β. Now consider loss function L 2 (p, a) = p 1 (p a) 2 and will derive the corresponding Bayes estimator ˆp B2. 8

17 Note that if loss function has the form L(p, a) = w(p)(p a) 2, then the corresponding Bayes estimator is given by ˆP BL = w(p)pf P Y (p y)dp w(p)f P Y (p y)dp = E P Y (w(p)p y) E P Y (w(p) y) (2.5) According to Equation (2.5) we have to evaluate w(p)pf P Y (p y)dp and w(p)f P Y (p y)dp. For w(p) = p 1 in loss function L 2 (p, a) we have and w(p)pf P Y (p y)dp = p 1 pf P Y (p y)dp = 1 w(p)f P Y (p y)dp = = p 1 pα 1 (1 p) k(m y)+β 1( 1 (1 p) k) y y ( 1) j( ) ( ) dp y B α, k(m + j y) + β j pα 2 (1 p) k(m y)+β 1( 1 (1 p) k) y dp y ( 1) j( ) ( ) y B α, k(m + j y) + β j (2.6) The value of the integral of the numerator in (2.6) is = k 1 = i= p α 2 (1 p) k(m y)+β 1( 1 (1 p) k) y dp (2.7) (1 u) α 2 u k(m y)+β 1 (1 u k ) y du (y > ) u k(m y)+i+β 1 (1 u k ) y 1 (1 u) α 1 du 9

18 k 1 = i= i= ( y 1 ( ) y 1 u )( 1) k(m y)+i+β 1 j u kj (1 u) α 1 du j k 1 y 1 ( ) y 1 1 = ( 1) j j i= u k(m y)+i+β 1+kj (1 u) α 1 du k 1 y 1 ( ) y 1 = ( 1) j B(km + kj + i + β ky, α) j Therefore, if y > we have ˆp B2 = k 1 y ( 1) j( y j) B(α, km + kj + β ky) y 1 i= j=i ( 1) j( y 1 j ) B(km + kj + i + β ky, α) (2.8) In the case y =, we let α > 1 and notice that equation (2.7) becomes p α 2 (1 p) km+β 1 dp = B(α 1, km + β), (α > 1). Summarizing the above, we obtain Theorem 2.1. Let p have prior distribution B(α, β). For the loss function L 2 (p, a) = p 1 (p a) 2 the Bayes estimator of p, denoted as ˆp B2, is given as ˆp B2 = y ( 1) j ( y j)b(α,km+kj+β ky) k 1 y 1 ( 1) j ( y 1 j )B(km+kj+β+i ky,α) i= B(α,km+β) B(α 1,km+β) if y = and α > 1. if y > 1

19 2.2 Approximation to Bayes Estimators of p It is easy to show that for sufficiently large N it holds that Γ(N + a) Γ(N + b) N a b (2.9) As it was mentioned in the previous section the group test is applied for the case of large sample size n = mk. Thus we could simplify the expression of ˆp B1 and ˆp B2 First we examine the expression (2.4) for ˆp B1. We have B(α, km + kj ky + β) = Γ(α)Γ(km + kj ky + β) Γ(km + kj ky + β + α) Γ(α)(km + kj ky + β) α since y is usually a small number and n = km is large. Similarly we have B(α + 1, km + kj ky + β) Γ(α + 1)(km + kj ky + β) α 1 Hence we obtain ˆp B1 = y ( y ) j ( 1) j B(α + 1, km + kj ky + β) y ( y ) ( 1)j B(α, km + kj ky + β) j Γ(α + 1) y Γ(α) y ( y j ( y ) j ( 1) j (km + kj ky + β) α 1 ) ( 1)j (km + kj ky + β) α 11

20 = α y ( 1) j( ) y (km + kj ky + β) α 1 j y ( 1) j( ) y (km + kj ky + β) α j where ˆp B 1 is an approximation to the Bayes estimator ˆp B1. As for ˆp B2 notice that if y 1, then ˆp B 1 (2.1) Γ(α)Γ(km + kj + i + β ky) B(km + kj + i + β ky, α) = Γ(km + kj + i + β ky + α) Γ(α)(km + kj + i + β ky) α Therefore, if y 1 it holds from Theorem 2.1 that ˆp B2 = k 1 y ( 1) j( ) y Γ(α)(km + kj ky + β) α i= k 1 j y 1 ( 1) j( ) y 1 Γ(α)(km + kj + i + β ky) α j y ( 1) j( ) y (km + kj ky + β) α i= j ˆp y 1 ( 1) j( ) B 2 (2.11) y 1 (km + kj + i + β ky) α On the other hand if y = and α > 1, then we have j B(α, km + β) ˆp B2 = B(α 1, km + β) (α 1)Γ(km + α + β 1) = Γ(km + α + β) Γ(α)Γ(km + β) Γ(km + α + β 1) = Γ(km + α + β) Γ(α 1)Γ(km + β) = α 1 km + α + β 1 = α 1 n + α + β 1 if α > 1. (2.12) From (2.11) and (2.12) we see that the Bayes estimator ˆp B2 can be aproximated by 12

21 ˆp B 2 = y ( 1) j ( y j)(km+kj ky+β) α k 1 y 1 ( 1) j ( y 1 i= j )(km+kj+i+β ky) α, if y 1; α 1, if y = and α > 1. n+α+β 1 (2.13) 2.3 Comparison in the special case of Y= Note the fact that the event (Y = ) is the same as event (X = ) where X is the number of positive subjects when subjects are tested individually. So it is interesting to compare the Bayes estimators based on X = and Y =. Clearly X B(n, p) = B(mk, p). Assuming that the proportion p has Beta(α, β) as its prior distribution, it is well known that the posterior distribution of p given X = x is Beta(x + α, n x + β). Suppose that square loss function L 1 (p, a) = (p a) 2 is used, then the Bayes estimator ˆp B1 based on X = is ˆp B1 = E P X (p x = ) = α n + α + β In the mean time the Bayes estimator ˆp B1 based on Y = is exactly the same as ˆp B1.(bottom of p8) Now consider the loss function L 2 (p, a) = p 1 (p a) 2. For this loss function the Bayes estimator ˆp B2 based on X can be obtained as follows. We have w(p)pf P X (p x)dp = f P X (p x)dp = 1 and 13

22 w(p)f P X (p x)dp = p 1( B(x + α, n x + β) ) 1 p x+α 1 (1 p) n x+β 1 dp = ( B(x + α, n x + β) ) 1 p x+α 2 (1 p) n x+β 1 dp Consider two cases. If x 1 which is equivalent to y 1. In this case ( B(x + α, n x + β) w(p)f P X (p x)dp = B(x + α 1, n x + β) ( Γ(x + α)γ(n x + β) = Γ(n + α + β) ( ) 1 x + α 1 = n + α + β 1 ) 1 Γ(n + α + β 1) Γ(x + α 1)Γ(n x + β) ) 1 Thus ˆp B2 = x + α 1 n + α + β 1 However, if x =, then it must be assumed that α > 1 and so w(p)f P X (p )dp = ( B(α, n + β) ) 1 B(α 1, n + β) = B(α, n + β) p α 2 (1 p) n+β 1 dp Γ(α 1)Γ(n + β) Γ(n + α + β) = Γ(n + α + β 1) Γ(α)Γ(n + β) = n + α + β 1 α 1 Consequently, it follows that 14

23 ˆp B2 = α 1, if x = and α > 1. n + α + β 1 Therefore ˆp B2 = x+α 1 if x 1; n+α+β 1 From this and equation (2.13), i.e., α 1 if x = and α > 1. n+α+β 1 ˆp B 2 = α 1 n + α + β 1 if y = and α > 1. we see that in this case ˆp B 2 and ˆp B2 are exactly the same when there is no positive subject in the entire sample. 2.4 Credible and HPD Interval for p Suppose that p has prior distribution Beta(α, β). From equation (2.3) it is known that the posterior distribution of p given Y = y is f P Y (p y) = pα 1 (1 p) k(m y)+β 1( 1 (1 p) k) y y ) ( 1)j B(α, k(m + j y) + β) Then for any set A (, 1), the credibal probability of A is ( y j P (p A) = f P Y (p y)dp and A is a credible set for p. By numerical computation it is not difficult to obtain a 1 γ credible interval for p. A 15

24 In a special case when α = 1, i.e., p has Beta(1, β) as its prior distribution, by equation (2.2) the marginal distribution of Y has density f Y (y) = β(1 p) β 1 ( m y ) (1 (1 p) k ) y (1 p) k(m y) dp = β ( ) m B (m y + βk ) k y, y + 1 (2.14) and consequently the posterior distribution of p given Y = y has density f P Y (p y) = β( ) ( m y (1 p) k(m y)+β 1 1 (1 p) k) y ) ( B m y + β, y + 1) k βk 1( m y = k(1 p)k(m y)+β 1( 1 (1 p) k) y B ( (2.15) m y + β, y + 1) k We further let β 1. The assumption is reasonable because we are dealing with small p and the majority of the probability distribution Beta(1, β) is close to zero if β is relatively large. With the assumption it holds that m y + (β 1)/k. It can be shown easily that the posterior f P Y (p y) is unimodal. Therefore, the HPD interval for p can be obtained by {p : f P Y (p y) a} where {p:f P Y (p y) a} f P Y (p y)dp = 1 γ for any given γ (, 1). For the case < β < 1, the posterior density f P Y (p y) strictly decreases in p (, 1), so the HPD interval for p can be obtained as 16

25 {p : f P Y (p y) a} where a f P Y (p y)dp = 1 γ. 3 Bayes Estimators of p from Prior on π In the present section we will assume π 1 (1 p) k has a Beta(α, β) prior distribution. On the basis of the observation Y = y we first calculate the Bayes estimator ˆπ B, then derive the Bayes estimator of p. Following Chaubey and Li (1995) this procedure is called the indirect Bayes procedure, the produced estimators is called the indirect Bayes estimators and denoted as ˆp IB1 or ˆp IB2 depending on the loss function L 1 (p, a) = (p a) 2 and L 2 (p, a) = p 1 (p a) 2. Obviously the posterior distribution of π given Y = y is Beta(y+α, m y+β), i.e., f π Y (π y) = ( B(y + α, m y + β) ) 1 π y+α 1 (1 π) m y+β 1 From π = 1 (1 p) k = π(p) we see that p = p(π) = 1 (1 π) 1/k. Consider loss function L(p, a) = w(p)(p a) 2. It is easy to see that the corresponding indirect Bayes estimator ˆp IB is given as ˆp IB = E( pw(p) Y = y ) E ( w(p) Y = y ) For the case L 1 (p, a) = (p a) 2 the indirect Bayes estimator ˆp IB1 is given in Chaubey and Li (1995) as 17

26 Γ(m + α + β)γ(m y + β + 1/k) ˆp IB1 = 1 Γ(m y + β)γ(m + α + β + 1/k) (3.1) which can be further approximated as ˆp IB 1 = 1 ( ) 1/k m + β y (3.2) m + α + β Now let us consider loss function L 2 (p, a) = p 1 (p a) 2. Clearly E(pw(p) Y = y) = pp 1 f π Y (π y)dπ = f π Y (π y)dπ = 1 and E(w(p) Y = y) = = = p 1 f π (π y)dπ p 1( B(y + α, m y + β) ) 1 π y+α 1 (1 π) m y+β 1 dπ ( 1 (1 π) 1/k ) 1 π y+α 1 (1 π) m y+β 1 dπ B(y + α, m y + β) (3.3) To evaluate the integral on the right side of (3.3) we let u = (1 π) 1/k and have = ( 1 (1 π) 1/k ) 1 π y+α 1 (1 π) m y+β 1 dπ (1 u) 1 (1 u k ) y+α 1 u k(m y+β 1) ku k 1 du 18

27 k 1 =k i= k 1 = i= k 1 = i= k 1 = B i= u k(m y+β)+i 1 (1 u k ) y+α 2 du u k(m y+β)+i k (1 u k ) y+α 2 ku k 1 du v m y+β+i/k 1 (1 v) y+α 2 dv (m y + β + ik, y + α 1 ), here it is required that α > 1 if y =. Therefore, the indirect Bayes estimator ˆp IB2 based on observation Y = y is ˆp IB2 = B(y + α, m y + β) k 1 i= B ( (3.4) m y + β + i, y + α 1) k where α > 1 if y =. We can rewrite ˆp IB2 as ˆp IB2 = k 1 i= Γ(y + α)γ(m y + β)/γ(m + α + β) Γ ( m y + β + i k) Γ(y + α 1)/Γ ( m + α + β + i k 1) =(y + α 1) k 1 i= i= Γ(m y + β)/γ(m + α + β) Γ ( m y + β + i k) /Γ ( m + α + β + i k 1) k 1 Γ(m y + β) =(y + α 1) Γ ( ) Γ ( m + α + β + i 1) k m y + β + i Γ(m + α + β) k (3.5) From (3.5) we see that ˆp IB2 can be approximated by y + α 1 ˆp IB 2 = k 1 i= (m y + β)i/k (m + α + β) i/k+1 19

28 here α > 1 if y =. = y + α 1 m + α + β k 1 i= ( 1 m y+β m+α+β ) i/k (3.6) 2

29 4 Numerical Studies A MATLAB simulation is performed in order to analyze the behaviors of different estimators under several combinations. I chose m = 5, 1, 15, 2, 25 and k = 5, 1, 15. Fifteen cases of p between.5 and.1 are considered. First, I simulate a sample that is from an independent, identically Bernoulli distribution and divided the observations into m groups with k units in each group. I use this new grouped sample to calculate the value of M LE. Second, I use the same sample to calculate the values of Bayes estimator and approximate Bayes estimator under different βs. The process is repeated 1 times. Then I will obtain the MSE of these estimators. Using the MSE as a criterion. Table 1-3 display the optimal βs which yield the smallest M SE of Bayes estimator associated with the square loss function. From these table, we can see that the optimal β decreases with the increase of p; the optimal β increases with the increase of m and k. Figures 1-5 also indicate these trends. Tables 4-7 show how I got the optimal βs. Tables 8-19 compare the MSE of Bayes estimator under the optimal βs with the approximate Bayes estimator as well as MLE. From these tables, we can draw a conclusion that the Bayes estimator is the best one as far as MSE is concerned. MLE is not a good estimator when p is small. Moreover, the approximate Bayes estimator is also much better than the M LE. 21

30 Table 1: k = 5, α = 1, N = 1 p k = 5 m This is the optimal β based on the smallest MSE of Bayes estimator by using k = 5 with combinations of different m and p. 22

31 Table 2: k = 1, α = 1, N = 1 p k = 1 m This is the optimal β based on the smallest MSE of Bayes estimator by using k = 1 with combinations of different m and p. 23

32 Table 3: k = 15, α = 1, N = 1 p k = 15 m This is the optimal β based on the smallest MSE of Bayes estimator by using k = 15 with combinations of different m and p. 24

33 Table 4: k = 5, m = 5, p =.5, α = 1, N = 1 β MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 β MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 β MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 Table 5: k = 5, m = 5, p =.1, α = 1, N = 1 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 The comparison of M SE of Bayes estimator and approximation to the Bayes estimator under different βs. 25

34 Table 6: k = 5, m = 15, p =.5, α = 1, N = 1 β MSE(ˆp B1 ) 4.451E E E E E-6 MSE(ˆp B 1) 4.718E E E E E-6 β MSE(ˆp B1 ) 4.366E E E E E-6 MSE(ˆp B 1) 4.594E E E E E-6 β MSE(ˆp B1 ) 4.383E E E E E-6 MSE(ˆp B 1) 4.575E E E E E-6 Table 7: k = 5, m = 15, p =.1, α = 1, N = 1 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 The comparison of M SE of Bayes estimator and approximation to the Bayes estimator under different βs. 26

35 Table 8: k = 5, p =.5,.6,.7,.8, α = 1, N = 1 p =.5 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 MSE(ˆp MLE ) 2.414E E E E E-5 p =.6 MSE(p B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 MSE(ˆp MLE ) 2.972E E E E E-5 p =.7 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 MSE(ˆp MLE ) E E E E E-5 p =.8 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) 7.632E E E E E-5 MSE(ˆp MLE ) E E E E E-5 Table-8 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 5 with different combinations of p =.5,.6,.7,.8 and m = 5, 1, 15, 2,

36 Table 9: k = 5, p =.9,.1,.2,.3, α = 1, N = 1 p =.9 MSE(ˆp B1 ) 1.42E E E E E-5 MSE(ˆp B 1) 1.544E E E E E-5 MSE(ˆp MLE ) 4.357E E E E E-5 p =.1 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) E E E E E-5 p =.2 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) 7.272E E E E E-5 MSE(ˆp MLE ) E E E E E-4 p =.3 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-9 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 5 with different combinations of p =.9,.1,.2,.3 and m = 5, 1, 15, 2,

37 Table 1: k = 5, p =.4,.5,.6,.7, α = 1, N = 1 p =.4 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.5 MSE(ˆp B1 ) 4.785E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.6 MSE(ˆp B1 ) 6.619E E E E E-4 MSE(ˆp B 1) 6.974E E E E E-4 MSE(ˆp MLE ) 4.449E E E E E-4 p =.7 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-1 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 5 with different combinations of p =.4,.5,.6,.7 and m = 5, 1, 15, 2,

38 Table 11: k = 5, p =.8,.9,.1, α = 1, N = 1 p =.8 MSE(ˆp B1 ) 1.716E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) 7.835E E E E E-4 p =.9 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.1 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-11 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 5 with different combinations of p =.8,.9,.1 and m = 5, 1, 15, 2, 25. 3

39 Table 12: k = 1, p =.5,.6,.7,.8, α = 1, N = 1 p =.5 MSE(ˆp B1 ) 3.336E E E E E-6 MSE(ˆp B 1) 3.358E E E E E-6 MSE(ˆp MLE ) E E E E E-5 p =.6 MSE(ˆp B1 ) 5.462E E E E E-6 MSE(ˆp B 1) 5.498E E E E E-6 MSE(ˆp MLE ) E E E E E-5 p =.7 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) 8.65E E E E E-6 MSE(ˆp MLE ) 1.877E E-5 5.3E E E-5 p =.8 MSE(ˆp B1 ) 1.768E E E E E-5 MSE(ˆp B 1) 1.852E E E E E-5 MSE(ˆp MLE ) 2.946E E E E E-5 Table-12 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 1 with different combinations of p =.5,.6,.7,.8 and m = 5, 1, 15, 2,

40 Table 13: k = 1, p =.9,.1,.2,.3, α = 1, N = 1 p =.9 MSE(ˆp B1 ) 1.495E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) E E E E E-5 p =.1 MSE(ˆp B1 ) 1.821E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) E E E E E-5 p =.2 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) 8.445E E E E E-5 MSE(ˆp MLE ) 7.779E E E E E-5 p =.3 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) 2.751E E E E E-4 Table-13 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 1 with different combinations of p =.9,.1,.2,.3 and m = 5, 1, 15, 2,

41 Table 14: k = 1, p =.4,.5,.6,.7, α = 1, N = 1 p =.4 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) 4.797E E E E E-4 p =.5 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.6 MSE(ˆp B1 ) 5.555E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) 2.122E E E E E-4 p =.7 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-14 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 1 with different combinations of p =.4,.5,.6,.7 and m = 5, 1, 15, 2,

42 Table 15: k = 1, p =.8,.9,.1, α = 1, N = 1 p =.8 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.9 MSE(ˆp B1 ) 1.383E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.1 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-15 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 1 with different combinations of p =.8,.9,.1 and m = 5, 1, 15, 2,

43 Table 16: k = 15, p =.5,.6,.7,.8, α = 1, N = 1 p =.5 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) 4.189E E E E E-6 MSE(ˆp MLE ) E E E E E-5 p =.6 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) 6.431E E E E E-6 MSE(ˆp MLE ) 1.678E E E E E-5 p =.7 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 MSE(ˆp MLE ) E E E E E-5 p =.8 MSE(ˆp B1 ) 1.217E E E E E-5 MSE(ˆp B 1) 1.299E E E E E-5 MSE(ˆp MLE ) E E E E E-5 Table-16 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 15 with different combinations of p =.5,.6,.7,.8 and m = 5, 1, 15, 2,

44 Table 17: k = 15, p =.9,.1,.2,.3, α = 1, N = 1 p =.9 MSE(ˆp B1 ) 1.586E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) E E E E E-5 p =.1 MSE(ˆp B1 ) 1.929E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) 1.99E E E E E-5 p =.2 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) 7.822E E E E E-5 MSE(ˆp MLE ) E E E E E-5 p =.3 MSE(ˆp B1 ) E E E-5 8.8E E-5 MSE(ˆp B 1) 1.683E E E E E-5 MSE(ˆp MLE ) E E E E E-4 Table-17 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 15 with different combinations of p =.9,.1,.2,.3 and m = 5, 1, 15, 2,

45 Table 18: k = 15, p =.4,.5,.6,.7, α = 1, N = 1 p =.4 MSE(ˆp B1 ) 2.481E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.5 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.6 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) 5.21E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.7 MSE(ˆp B1 ) 6.63E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-18 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 15 with different combinations of p =.4,.5,.6,.7 and m = 5, 1, 15, 2,

46 Table 19: k = 15, p =.8,.9,.1, α = 1, N = 1 p =.8 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) 7.728E E E E E-4 MSE(ˆp MLE ) E E E E E-3 p =.9 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) 9.28E E E E E-4 MSE(ˆp MLE ) 2.59E E E E E-3 p =.1 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) 1.679E E E E E-4 MSE(ˆp MLE ) E E E E E-3 Table-19 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 15 with different combinations of p =.8,.9,.1 and m = 5, 1, 15, 2,

47 Figure 1: K = 5, m = 1 Figure 2: K = 5, m = 2 39

48 Figure 3: K = 5, m = 1 or 2 Figure 4: K = 1, m = 1 4

49 Figure 5: K = 5 or 1, m = 1 41

50 5 Conclusion According to the analytical and numerical results obtained above, We know that the M SE of Bayes estimators and approximation to Bayes estimators are both smaller than that of MLE for small p. Indirect Bayes estimators performs very well too. We could draw a conclusion that Bayes estimators are much better than MLE when p is small. 42

51 References [1] Bhattacharyya, F. K., Karandinos, M. G., Defoliart, G. R., Point estimates and confidence interval for infection rates using pooled organisms in epidemiologic studies. American Journal of Epidemiology. 19, [2] Chaubey, Y., Li, W., Comparison between maximum likelihood and Bayes methods for estimation of binomial probability with sample compositing. Journal of Official Statistics. 11, [3] Chiang, C. L., Reeves, W. C., Statistical estimation of virus infection rates in mosquito vector populations. American Journal of Hygiene. 75, [4] Chick, S., Bayesian models for limiting dilution assay and group test data. Biometrics. 52, , [5] Dorfman, R., The detection of defective members of large populations. Annals of Mathematical Statistics. 14, [6] Gastwirth, J., Johnson, W., Reneau, D., Bayesian analysis of screening data: Application to AIDS in blood donors. The Canadian Journal of Statistics. 19, [7] Gastwirth, J., 2. The efficiency of pooling in the detection of rare mutations. Americal Journal of Human Genetics. 67, [8] Gibbs, A. J., Gower, J. C., 196. The use of a multiple-transfer method in plant virus transission studies-some statistical points arising in the analysis of results. Annals of Applied Biology. 48, [9] Hepworth, G., Exact confidence intervals for proportions estimated by group testing. Biometrics. 52,

52 [1] Katholi, C. R., Unnasch, T. R., 26. Important experimental parameters for determining infection rates in arthropod vectors using pool screening approaches. The American Society of Tropical Medicine and Hygiene. 74, [11] Kline, R. L., Brothers, T. A., Brookmeyer, R., Zeger, S., Quinn, T. C., Evaluation of human immunodficiency virus seroprevalence in population surveys using pooled sera. Journal of Clinical Microbiology. 27, [12] Radoni, B. C., Hepworth, G., Richardson, C., Moran, J. R., The use of a sequential batch testing procedure and ELISA to determine the incidence of five viruses in Victorian cut-flower sim Carnations. Aust. J. Agric. Res.. 45, [13] Swallow, W. H., Group testing for estimating infection rates and probabilities of disease transmission. Phytopathology. 75, [14] Worlund, D. D., Taylor, G., Estimation of disease incidence in fish populations. Canadian Journal of Fisheries and Aquatic Science. 4, [15] Xie, M., Tatsuoka, K., Sacks, J., Young, S., 21. Group testing with blockers and synergism. J. Am. Statist. Assoc.. 96,

53 APPENDIX I list the equations in this study here for the convenience of finding them. 45

54 1 Bayes Inferences from Prior on p. Loss function L 1 (p, a) = (p a) 2. y ( y ) j ( 1) j B(α + 1, km + kj ky + β) ˆp B1 = y ( y ) ( 1)j B(α, km + kj ky + β) j. y 1 Equation 2.4 page 8 Approximation to the Bayes estimator ˆp B1. ˆp B 1 = α y ( 1) j( ) y (km + kj ky + β) α 1 j y ( 1) j( ) y (km + kj ky + β) α j. y 1 Equation 2.1 page 12 Loss function L 2 (p, a) = p 1 (p a) 2. ˆp B2 = y ( 1) j ( y j)b(α,km+kj+β ky) k 1 y 1 ( 1) j ( y 1 j )B(km+kj+β+i k,α) i= y ( 1) j ( y j)b(α,km+β+kj ky) if y 1 B(α 1,km+β) if y = and α > 1. Approximation to the Bayes estimator ˆp B2. ˆp B 2 = y ( 1) j ( y j)(km+kj ky+β) α k 1 y 1 ( 1) j ( y 1 i= j )(km+kj+i+β ky) α, if y 1; α 1, if y = and α > 1. n+α+β 1 Theorem 1 page 1 Equation 2.13 page 13 46

55 2 Bayes Estimators of p from Prior on π. Loss function L 1 (p, a) = (p a) 2. ˆp IB1 = 1 Γ(m + α + β)γ(m y + β + 1/k) Γ(m y + β)γ(m + α + β + 1/k) Equation 3.1 page 18 Approximation to the Bayes estimator ˆp IB1. ˆp IB 1 = 1 ( ) 1/k m + β y Equation 3.2 page 18 m + α + β Loss function L 2 (p, a) = p 1 (p a) 2. k 1 Γ(m y + β)γ ( m + α + β + i k ˆp IB2 = (y + α 1) 1) Γ ( ) Equation 3.5 page 19 m y + β + i k Γ(m + α + β) i= Approximation to the Bayes estimator ˆp IB2. ˆp IB 2 = y + α 1 m + α + β k 1 i= ( 1 m y+β m+α+β ) i/k Equation 3.6 page 2 47

Inferences about Parameters of Trivariate Normal Distribution with Missing Data

Inferences about Parameters of Trivariate Normal Distribution with Missing Data Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 7-5-3 Inferences about Parameters of Trivariate Normal Distribution with Missing

More information

Bayes Factors for Grouped Data

Bayes Factors for Grouped Data Bayes Factors for Grouped Data Lizanne Raubenheimer and Abrie J. van der Merwe 2 Department of Statistics, Rhodes University, Grahamstown, South Africa, L.Raubenheimer@ru.ac.za 2 Department of Mathematical

More information

Inference for a Population Proportion

Inference for a Population Proportion Al Nosedal. University of Toronto. November 11, 2015 Statistical inference is drawing conclusions about an entire population based on data in a sample drawn from that population. From both frequentist

More information

Consecutively Choosing between Three Experiments for Estimating Prevalence Rate of a Trait with Imperfect Tests

Consecutively Choosing between Three Experiments for Estimating Prevalence Rate of a Trait with Imperfect Tests International Journal of Statistics and Applications 07, 7(): 9-06 DOI: 0.9/j.statistics.07070.04 Consecutively Choosing between Three periments for stimating revalence Rate of a Trait with Imperfect Tests

More information

Comparison of Some Improved Estimators for Linear Regression Model under Different Conditions

Comparison of Some Improved Estimators for Linear Regression Model under Different Conditions Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-24-2015 Comparison of Some Improved Estimators for Linear Regression Model under

More information

A Very Brief Summary of Bayesian Inference, and Examples

A Very Brief Summary of Bayesian Inference, and Examples A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X

More information

Bayesian Inference for Binomial Proportion

Bayesian Inference for Binomial Proportion 8 Bayesian Inference for Binomial Proportion Frequently there is a large population where π, a proportion of the population, has some attribute. For instance, the population could be registered voters

More information

Inference about Reliability Parameter with Underlying Gamma and Exponential Distribution

Inference about Reliability Parameter with Underlying Gamma and Exponential Distribution Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 9-3-211 Inference about Reliability Parameter with Underlying Gamma and Exponential

More information

Using Probability to do Statistics.

Using Probability to do Statistics. Al Nosedal. University of Toronto. November 5, 2015 Milk and honey and hemoglobin Animal experiments suggested that honey in a diet might raise hemoglobin level. A researcher designed a study involving

More information

DIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS

DIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS DIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS THUY ANH NGO 1. Introduction Statistics are easily come across in our daily life. Statements such as the average

More information

HPD Intervals / Regions

HPD Intervals / Regions HPD Intervals / Regions The HPD region will be an interval when the posterior is unimodal. If the posterior is multimodal, the HPD region might be a discontiguous set. Picture: The set {θ : θ (1.5, 3.9)

More information

Study on a Hierarchy Model

Study on a Hierarchy Model Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-23-212 Study on a Hierarchy Model Suisui Che Florida International University,

More information

1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D.

1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D. probabilities, we ll use Bayes formula. We can easily compute the reverse probabilities A short introduction to Bayesian statistics, part I Math 17 Probability and Statistics Prof. D. Joyce, Fall 014 I

More information

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017 Bayesian inference Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 10, 2017 1 / 22 Outline for today A genetic example Bayes theorem Examples Priors Posterior summaries

More information

A Study on the Correlation of Bivariate And Trivariate Normal Models

A Study on the Correlation of Bivariate And Trivariate Normal Models Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 11-1-2013 A Study on the Correlation of Bivariate And Trivariate Normal Models Maria

More information

Statistical Inference Using Progressively Type-II Censored Data with Random Scheme

Statistical Inference Using Progressively Type-II Censored Data with Random Scheme International Mathematical Forum, 3, 28, no. 35, 1713-1725 Statistical Inference Using Progressively Type-II Censored Data with Random Scheme Ammar M. Sarhan 1 and A. Abuammoh Department of Statistics

More information

Bayesian Meta-analysis with Hierarchical Modeling Brian P. Hobbs 1

Bayesian Meta-analysis with Hierarchical Modeling Brian P. Hobbs 1 Bayesian Meta-analysis with Hierarchical Modeling Brian P. Hobbs 1 Division of Biostatistics, School of Public Health, University of Minnesota, Mayo Mail Code 303, Minneapolis, Minnesota 55455 0392, U.S.A.

More information

BAYESIAN ANALYSIS OF DOSE-RESPONSE CALIBRATION CURVES

BAYESIAN ANALYSIS OF DOSE-RESPONSE CALIBRATION CURVES Libraries Annual Conference on Applied Statistics in Agriculture 2005-17th Annual Conference Proceedings BAYESIAN ANALYSIS OF DOSE-RESPONSE CALIBRATION CURVES William J. Price Bahman Shafii Follow this

More information

Human or Cylon? Group testing on the Battlestar Galactica

Human or Cylon? Group testing on the Battlestar Galactica Human or Cylon? Group testing on the Statistics and The story so far Video Christopher R. Bilder Department of Statistics University of Nebraska-Lincoln chris@chrisbilder.com Slide 1 of 37 Slide 2 of 37

More information

Probability and Probability Distributions. Dr. Mohammed Alahmed

Probability and Probability Distributions. Dr. Mohammed Alahmed Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about

More information

Experimental designs for multiple responses with different models

Experimental designs for multiple responses with different models Graduate Theses and Dissertations Graduate College 2015 Experimental designs for multiple responses with different models Wilmina Mary Marget Iowa State University Follow this and additional works at:

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

9 Bayesian inference. 9.1 Subjective probability

9 Bayesian inference. 9.1 Subjective probability 9 Bayesian inference 1702-1761 9.1 Subjective probability This is probability regarded as degree of belief. A subjective probability of an event A is assessed as p if you are prepared to stake pm to win

More information

PS 203 Spring 2002 Homework One - Answer Key

PS 203 Spring 2002 Homework One - Answer Key PS 203 Spring 2002 Homework One - Answer Key 1. If you have a home or office computer, download and install WinBUGS. If you don t have your own computer, try running WinBUGS in the Department lab. 2. The

More information

Assessing the Effect of Prior Distribution Assumption on the Variance Parameters in Evaluating Bioequivalence Trials

Assessing the Effect of Prior Distribution Assumption on the Variance Parameters in Evaluating Bioequivalence Trials Georgia State University ScholarWorks @ Georgia State University Mathematics Theses Department of Mathematics and Statistics 8--006 Assessing the Effect of Prior Distribution Assumption on the Variance

More information

Lecture 10 and 11: Text and Discrete Distributions

Lecture 10 and 11: Text and Discrete Distributions Lecture 10 and 11: Text and Discrete Distributions Machine Learning 4F13, Spring 2014 Carl Edward Rasmussen and Zoubin Ghahramani CUED http://mlg.eng.cam.ac.uk/teaching/4f13/ Rasmussen and Ghahramani Lecture

More information

Introduction to Bayesian Methods

Introduction to Bayesian Methods Introduction to Bayesian Methods Jessi Cisewski Department of Statistics Yale University Sagan Summer Workshop 2016 Our goal: introduction to Bayesian methods Likelihoods Priors: conjugate priors, non-informative

More information

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture

More information

Bayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017

Bayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017 Chalmers April 6, 2017 Bayesian philosophy Bayesian philosophy Bayesian statistics versus classical statistics: War or co-existence? Classical statistics: Models have variables and parameters; these are

More information

Objective Bayesian Hypothesis Testing and Estimation for the Risk Ratio in a Correlated 2x2 Table with Structural Zero

Objective Bayesian Hypothesis Testing and Estimation for the Risk Ratio in a Correlated 2x2 Table with Structural Zero Clemson University TigerPrints All Theses Theses 8-2013 Objective Bayesian Hypothesis Testing and Estimation for the Risk Ratio in a Correlated 2x2 Table with Structural Zero Xiaohua Bai Clemson University,

More information

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007 Bayesian inference Fredrik Ronquist and Peter Beerli October 3, 2007 1 Introduction The last few decades has seen a growing interest in Bayesian inference, an alternative approach to statistical inference.

More information

Probability Review - Bayes Introduction

Probability Review - Bayes Introduction Probability Review - Bayes Introduction Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Advantages of Bayesian Analysis Answers the questions that researchers are usually interested in, What

More information

New Bivariate Lifetime Distributions Based on Bath-Tub Shaped Failure Rate

New Bivariate Lifetime Distributions Based on Bath-Tub Shaped Failure Rate Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 10-30-2014 New Bivariate Lifetime Distributions Based on Bath-Tub Shaped Failure

More information

Part 2: One-parameter models

Part 2: One-parameter models Part 2: One-parameter models 1 Bernoulli/binomial models Return to iid Y 1,...,Y n Bin(1, ). The sampling model/likelihood is p(y 1,...,y n ) = P y i (1 ) n P y i When combined with a prior p( ), Bayes

More information

Chapter 5. Bayesian Statistics

Chapter 5. Bayesian Statistics Chapter 5. Bayesian Statistics Principles of Bayesian Statistics Anything unknown is given a probability distribution, representing degrees of belief [subjective probability]. Degrees of belief [subjective

More information

Bayesian Econometrics

Bayesian Econometrics Bayesian Econometrics Christopher A. Sims Princeton University sims@princeton.edu September 20, 2016 Outline I. The difference between Bayesian and non-bayesian inference. II. Confidence sets and confidence

More information

Numerical computation of an optimal control problem with homogenization in one-dimensional case

Numerical computation of an optimal control problem with homogenization in one-dimensional case Retrospective Theses and Dissertations Iowa State University Capstones, Theses and Dissertations 28 Numerical computation of an optimal control problem with homogenization in one-dimensional case Zhen

More information

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian

More information

PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.

PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation. PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.. Beta Distribution We ll start by learning about the Beta distribution, since we end up using

More information

AN OPTIMAL STRATEGY FOR SEQUENTIAL CLASSIFICATION ON PARTIALLY ORDERED SETS. Abstract

AN OPTIMAL STRATEGY FOR SEQUENTIAL CLASSIFICATION ON PARTIALLY ORDERED SETS. Abstract AN OPTIMAL STRATEGY FOR SEQUENTIAL CLASSIFICATION ON PARTIALLY ORDERED SETS Thomas S. Ferguson Department of Mathematics UCLA Los Angeles, CA 90095, USA Curtis Tatsuoka 1,2 Department of Statistics The

More information

Bayesian Inference for Normal Mean

Bayesian Inference for Normal Mean Al Nosedal. University of Toronto. November 18, 2015 Likelihood of Single Observation The conditional observation distribution of y µ is Normal with mean µ and variance σ 2, which is known. Its density

More information

Introduction to Bayesian Methods. Introduction to Bayesian Methods p.1/??

Introduction to Bayesian Methods. Introduction to Bayesian Methods p.1/?? to Bayesian Methods Introduction to Bayesian Methods p.1/?? We develop the Bayesian paradigm for parametric inference. To this end, suppose we conduct (or wish to design) a study, in which the parameter

More information

Data Analysis and Uncertainty Part 2: Estimation

Data Analysis and Uncertainty Part 2: Estimation Data Analysis and Uncertainty Part 2: Estimation Instructor: Sargur N. University at Buffalo The State University of New York srihari@cedar.buffalo.edu 1 Topics in Estimation 1. Estimation 2. Desirable

More information

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012 Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood

More information

ESTIMATING STATISTICAL CHARACTERISTICS UNDER INTERVAL UNCERTAINTY AND CONSTRAINTS: MEAN, VARIANCE, COVARIANCE, AND CORRELATION ALI JALAL-KAMALI

ESTIMATING STATISTICAL CHARACTERISTICS UNDER INTERVAL UNCERTAINTY AND CONSTRAINTS: MEAN, VARIANCE, COVARIANCE, AND CORRELATION ALI JALAL-KAMALI ESTIMATING STATISTICAL CHARACTERISTICS UNDER INTERVAL UNCERTAINTY AND CONSTRAINTS: MEAN, VARIANCE, COVARIANCE, AND CORRELATION ALI JALAL-KAMALI Department of Computer Science APPROVED: Vladik Kreinovich,

More information

Single Sample Means. SOCY601 Alan Neustadtl

Single Sample Means. SOCY601 Alan Neustadtl Single Sample Means SOCY601 Alan Neustadtl The Central Limit Theorem If we have a population measured by a variable with a mean µ and a standard deviation σ, and if all possible random samples of size

More information

Bayesian course - problem set 5 (lecture 6)

Bayesian course - problem set 5 (lecture 6) Bayesian course - problem set 5 (lecture 6) Ben Lambert November 30, 2016 1 Stan entry level: discoveries data The file prob5 discoveries.csv contains data on the numbers of great inventions and scientific

More information

INFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS. Tao Jiang. A Thesis

INFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS. Tao Jiang. A Thesis INFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS Tao Jiang A Thesis Submitted to the Graduate College of Bowling Green State University in partial fulfillment of the

More information

Foundations of Statistical Inference

Foundations of Statistical Inference Foundations of Statistical Inference Julien Berestycki Department of Statistics University of Oxford MT 2016 Julien Berestycki (University of Oxford) SB2a MT 2016 1 / 20 Lecture 6 : Bayesian Inference

More information

Unobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior:

Unobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior: Pi Priors Unobservable Parameter population proportion, p prior: π ( p) Conjugate prior π ( p) ~ Beta( a, b) same PDF family exponential family only Posterior π ( p y) ~ Beta( a + y, b + n y) Observed

More information

Noninformative Priors for the Ratio of the Scale Parameters in the Inverted Exponential Distributions

Noninformative Priors for the Ratio of the Scale Parameters in the Inverted Exponential Distributions Communications for Statistical Applications and Methods 03, Vol. 0, No. 5, 387 394 DOI: http://dx.doi.org/0.535/csam.03.0.5.387 Noninformative Priors for the Ratio of the Scale Parameters in the Inverted

More information

Confidence Intervals for the Ratio of Two Exponential Means with Applications to Quality Control

Confidence Intervals for the Ratio of Two Exponential Means with Applications to Quality Control Western Kentucky University TopSCHOLAR Student Research Conference Select Presentations Student Research Conference 6-009 Confidence Intervals for the Ratio of Two Exponential Means with Applications to

More information

General Bayesian Inference I

General Bayesian Inference I General Bayesian Inference I Outline: Basic concepts, One-parameter models, Noninformative priors. Reading: Chapters 10 and 11 in Kay-I. (Occasional) Simplified Notation. When there is no potential for

More information

MTH U481 : SPRING 2009: PRACTICE PROBLEMS FOR FINAL

MTH U481 : SPRING 2009: PRACTICE PROBLEMS FOR FINAL MTH U481 : SPRING 2009: PRACTICE PROBLEMS FOR FINAL 1). Two urns are provided as follows: urn 1 contains 2 white chips and 4 red chips, while urn 2 contains 5 white chips and 3 red chips. One chip is chosen

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

On the Performance of some Poisson Ridge Regression Estimators

On the Performance of some Poisson Ridge Regression Estimators Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-28-2018 On the Performance of some Poisson Ridge Regression Estimators Cynthia

More information

1 Introduction. P (n = 1 red ball drawn) =

1 Introduction. P (n = 1 red ball drawn) = Introduction Exercises and outline solutions. Y has a pack of 4 cards (Ace and Queen of clubs, Ace and Queen of Hearts) from which he deals a random of selection 2 to player X. What is the probability

More information

Lecture 2: Conjugate priors

Lecture 2: Conjugate priors (Spring ʼ) Lecture : Conjugate priors Julia Hockenmaier juliahmr@illinois.edu Siebel Center http://www.cs.uiuc.edu/class/sp/cs98jhm The binomial distribution If p is the probability of heads, the probability

More information

Midterm Examination. Mth 136 = Sta 114. Wednesday, 2000 March 8, 2:20 3:35 pm

Midterm Examination. Mth 136 = Sta 114. Wednesday, 2000 March 8, 2:20 3:35 pm Midterm Examination Mth 136 = Sta 114 Wednesday, 2000 March 8, 2:20 3:35 pm This is a closed-book examination so please do not refer to your notes, the text, or to any other books. You may use a two-sided

More information

Bayesian Inference: Probit and Linear Probability Models

Bayesian Inference: Probit and Linear Probability Models Utah State University DigitalCommons@USU All Graduate Plan B and other Reports Graduate Studies 5-1-2014 Bayesian Inference: Probit and Linear Probability Models Nate Rex Reasch Utah State University Follow

More information

Inverse Sampling for McNemar s Test

Inverse Sampling for McNemar s Test International Journal of Statistics and Probability; Vol. 6, No. 1; January 27 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education Inverse Sampling for McNemar s Test

More information

PubH 5450 Biostatistics I Prof. Carlin. Lecture 13

PubH 5450 Biostatistics I Prof. Carlin. Lecture 13 PubH 5450 Biostatistics I Prof. Carlin Lecture 13 Outline Outline Sample Size Counts, Rates and Proportions Part I Sample Size Type I Error and Power Type I error rate: probability of rejecting the null

More information

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics March 14, 2018 CS 361: Probability & Statistics Inference The prior From Bayes rule, we know that we can express our function of interest as Likelihood Prior Posterior The right hand side contains the

More information

Compositions, Bijections, and Enumerations

Compositions, Bijections, and Enumerations Georgia Southern University Digital Commons@Georgia Southern Electronic Theses & Dissertations COGS- Jack N. Averitt College of Graduate Studies Fall 2012 Compositions, Bijections, and Enumerations Charles

More information

Downloaded from:

Downloaded from: Camacho, A; Kucharski, AJ; Funk, S; Breman, J; Piot, P; Edmunds, WJ (2014) Potential for large outbreaks of Ebola virus disease. Epidemics, 9. pp. 70-8. ISSN 1755-4365 DOI: https://doi.org/10.1016/j.epidem.2014.09.003

More information

Classical and Bayesian inference

Classical and Bayesian inference Classical and Bayesian inference AMS 132 January 18, 2018 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 1 / 9 Sampling from a Bernoulli Distribution Theorem (Beta-Bernoulli

More information

Examining the accuracy of the normal approximation to the poisson random variable

Examining the accuracy of the normal approximation to the poisson random variable Eastern Michigan University DigitalCommons@EMU Master's Theses and Doctoral Dissertations Master's Theses, and Doctoral Dissertations, and Graduate Capstone Projects 2009 Examining the accuracy of the

More information

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio Estimation of reliability parameters from Experimental data (Parte 2) This lecture Life test (t 1,t 2,...,t n ) Estimate θ of f T t θ For example: λ of f T (t)= λe - λt Classical approach (frequentist

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics October 17, 2017 CS 361: Probability & Statistics Inference Maximum likelihood: drawbacks A couple of things might trip up max likelihood estimation: 1) Finding the maximum of some functions can be quite

More information

2 Inference for Multinomial Distribution

2 Inference for Multinomial Distribution Markov Chain Monte Carlo Methods Part III: Statistical Concepts By K.B.Athreya, Mohan Delampady and T.Krishnan 1 Introduction In parts I and II of this series it was shown how Markov chain Monte Carlo

More information

MAT 2379, Introduction to Biostatistics, Sample Calculator Questions 1. MAT 2379, Introduction to Biostatistics

MAT 2379, Introduction to Biostatistics, Sample Calculator Questions 1. MAT 2379, Introduction to Biostatistics MAT 2379, Introduction to Biostatistics, Sample Calculator Questions 1 MAT 2379, Introduction to Biostatistics Sample Calculator Problems for the Final Exam Note: The exam will also contain some problems

More information

Bayesian RL Seminar. Chris Mansley September 9, 2008

Bayesian RL Seminar. Chris Mansley September 9, 2008 Bayesian RL Seminar Chris Mansley September 9, 2008 Bayes Basic Probability One of the basic principles of probability theory, the chain rule, will allow us to derive most of the background material in

More information

Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units

Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units Sahar Z Zangeneh Robert W. Keener Roderick J.A. Little Abstract In Probability proportional

More information

Estimation of Quantiles

Estimation of Quantiles 9 Estimation of Quantiles The notion of quantiles was introduced in Section 3.2: recall that a quantile x α for an r.v. X is a constant such that P(X x α )=1 α. (9.1) In this chapter we examine quantiles

More information

Bayesian Inference. STA 121: Regression Analysis Artin Armagan

Bayesian Inference. STA 121: Regression Analysis Artin Armagan Bayesian Inference STA 121: Regression Analysis Artin Armagan Bayes Rule...s! Reverend Thomas Bayes Posterior Prior p(θ y) = p(y θ)p(θ)/p(y) Likelihood - Sampling Distribution Normalizing Constant: p(y

More information

Bayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification

Bayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification Bayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification Michael Anderson, PhD Hélène Carabin, DVM, PhD Department of Biostatistics and Epidemiology The University of Oklahoma

More information

Introduction to Machine Learning. Lecture 2

Introduction to Machine Learning. Lecture 2 Introduction to Machine Learning Lecturer: Eran Halperin Lecture 2 Fall Semester Scribe: Yishay Mansour Some of the material was not presented in class (and is marked with a side line) and is given for

More information

Introduction to Bayesian Statistics 1

Introduction to Bayesian Statistics 1 Introduction to Bayesian Statistics 1 STA 442/2101 Fall 2018 1 This slide show is an open-source document. See last slide for copyright information. 1 / 42 Thomas Bayes (1701-1761) Image from the Wikipedia

More information

Bernoulli and Poisson models

Bernoulli and Poisson models Bernoulli and Poisson models Bernoulli/binomial models Return to iid Y 1,...,Y n Bin(1, ). The sampling model/likelihood is p(y 1,...,y n ) = P y i (1 ) n P y i When combined with a prior p( ), Bayes rule

More information

The extreme points of symmetric norms on R^2

The extreme points of symmetric norms on R^2 Graduate Theses and Dissertations Iowa State University Capstones, Theses and Dissertations 2008 The extreme points of symmetric norms on R^2 Anchalee Khemphet Iowa State University Follow this and additional

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

10. Exchangeability and hierarchical models Objective. Recommended reading

10. Exchangeability and hierarchical models Objective. Recommended reading 10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.

More information

Bayesian Estimation An Informal Introduction

Bayesian Estimation An Informal Introduction Mary Parker, Bayesian Estimation An Informal Introduction page 1 of 8 Bayesian Estimation An Informal Introduction Example: I take a coin out of my pocket and I want to estimate the probability of heads

More information

Bayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida

Bayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida Bayesian Statistical Methods Jeff Gill Department of Political Science, University of Florida 234 Anderson Hall, PO Box 117325, Gainesville, FL 32611-7325 Voice: 352-392-0262x272, Fax: 352-392-8127, Email:

More information

Joint, Conditional, & Marginal Probabilities

Joint, Conditional, & Marginal Probabilities Joint, Conditional, & Marginal Probabilities The three axioms for probability don t discuss how to create probabilities for combined events such as P [A B] or for the likelihood of an event A given that

More information

Bayesian Statistics. Debdeep Pati Florida State University. February 11, 2016

Bayesian Statistics. Debdeep Pati Florida State University. February 11, 2016 Bayesian Statistics Debdeep Pati Florida State University February 11, 2016 Historical Background Historical Background Historical Background Brief History of Bayesian Statistics 1764-1838: called probability

More information

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm Midterm Examination STA 215: Statistical Inference Due Wednesday, 2006 Mar 8, 1:15 pm This is an open-book take-home examination. You may work on it during any consecutive 24-hour period you like; please

More information

NIH Public Access Author Manuscript Commun Stat Theory Methods. Author manuscript; available in PMC 2011 January 7.

NIH Public Access Author Manuscript Commun Stat Theory Methods. Author manuscript; available in PMC 2011 January 7. NIH Public Access Author Manuscript Published in final edited form as: Commun Stat Theory Methods. 2011 January 1; 40(3): 436 448. doi:10.1080/03610920903391303. Optimal Configuration of a Square Array

More information

ECE521 W17 Tutorial 6. Min Bai and Yuhuai (Tony) Wu

ECE521 W17 Tutorial 6. Min Bai and Yuhuai (Tony) Wu ECE521 W17 Tutorial 6 Min Bai and Yuhuai (Tony) Wu Agenda knn and PCA Bayesian Inference k-means Technique for clustering Unsupervised pattern and grouping discovery Class prediction Outlier detection

More information

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b) LECTURE 5 NOTES 1. Bayesian point estimators. In the conventional (frequentist) approach to statistical inference, the parameter θ Θ is considered a fixed quantity. In the Bayesian approach, it is considered

More information

Joint, Conditional, & Marginal Probabilities

Joint, Conditional, & Marginal Probabilities Joint, Conditional, & Marginal Probabilities Statistics 110 Summer 2006 Copyright c 2006 by Mark E. Irwin Joint, Conditional, & Marginal Probabilities The three axioms for probability don t discuss how

More information

A CONDITION TO OBTAIN THE SAME DECISION IN THE HOMOGENEITY TEST- ING PROBLEM FROM THE FREQUENTIST AND BAYESIAN POINT OF VIEW

A CONDITION TO OBTAIN THE SAME DECISION IN THE HOMOGENEITY TEST- ING PROBLEM FROM THE FREQUENTIST AND BAYESIAN POINT OF VIEW A CONDITION TO OBTAIN THE SAME DECISION IN THE HOMOGENEITY TEST- ING PROBLEM FROM THE FREQUENTIST AND BAYESIAN POINT OF VIEW Miguel A Gómez-Villegas and Beatriz González-Pérez Departamento de Estadística

More information

The Bayesian Approach to Multi-equation Econometric Model Estimation

The Bayesian Approach to Multi-equation Econometric Model Estimation Journal of Statistical and Econometric Methods, vol.3, no.1, 2014, 85-96 ISSN: 2241-0384 (print), 2241-0376 (online) Scienpress Ltd, 2014 The Bayesian Approach to Multi-equation Econometric Model Estimation

More information

Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion

Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Glenn Heller and Jing Qin Department of Epidemiology and Biostatistics Memorial

More information

Dropping the Lowest Score: A Mathematical Analysis of a Common Grading Practice

Dropping the Lowest Score: A Mathematical Analysis of a Common Grading Practice Western Oregon University Digital Commons@WOU Honors Senior Theses/Projects Student Scholarship - 6-1-2013 Dropping the Lowest Score: A Mathematical Analysis of a Common Grading Practice Rosie Brown Western

More information

Discrete Binary Distributions

Discrete Binary Distributions Discrete Binary Distributions Carl Edward Rasmussen November th, 26 Carl Edward Rasmussen Discrete Binary Distributions November th, 26 / 5 Key concepts Bernoulli: probabilities over binary variables Binomial:

More information

Weakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0

More information

CS 340 Fall 2007: Homework 3

CS 340 Fall 2007: Homework 3 CS 34 Fall 27: Homework 3 1 Marginal likelihood for the Beta-Bernoulli model We showed that the marginal likelihood is the ratio of the normalizing constants: p(d) = B(α 1 + N 1, α + N ) B(α 1, α ) = Γ(α

More information

STAT:5100 (22S:193) Statistical Inference I

STAT:5100 (22S:193) Statistical Inference I STAT:5100 (22S:193) Statistical Inference I Week 10 Luke Tierney University of Iowa Fall 2015 Luke Tierney (U Iowa) STAT:5100 (22S:193) Statistical Inference I Fall 2015 1 Monday, October 26, 2015 Recap

More information