Bayesian Estimation of Small Proportions Using Binomial Group Test
|
|
- Branden Doyle
- 5 years ago
- Views:
Transcription
1 Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School Bayesian Estimation of Small Proportions Using Binomial Group Test Shihua Luo Florida International University, DOI: /etd.FI Follow this and additional works at: Recommended Citation Luo, Shihua, "Bayesian Estimation of Small Proportions Using Binomial Group Test" (212). FIU Electronic Theses and Dissertations This work is brought to you for free and open access by the University Graduate School at FIU Digital Commons. It has been accepted for inclusion in FIU Electronic Theses and Dissertations by an authorized administrator of FIU Digital Commons. For more information, please contact
2 FLORIDA INTERNATIONAL UNIVERSITY Miami, Florida BAYESIAN ESTIMATION OF SMALL PROPORTIONS USING BINOMIAL GROUP TEST A thesis submitted in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE in STATISTICS by Shihua Luo 212
3 To: Dean Kenneth Furton College of Arts and Sciences This thesis, written by Shihua Luo, and entitled Bayesian Estimation of Small Proportions Using Binomial Group Test, having been approved in respect to style and intellectual content, is referred to you for judgement. We have read this thesis and recommend that it be approved. Florence George Kai Huang, Co-Major Professor Jie Mi, Co-Major Professor Date of Defense: November 9, 212 The thesis of Shihua Luo is approved. Dean Kenneth Furton College of Arts and Sciences Dean Lakshmi N. Reddi University Graduate School Florida International University, 212 ii
4 ACKNOWLEDGMENTS First of all, I would like to express my sincere thanks to my major professor, Dr. Jie Mi for his patient guidance, enthusiasm, encouragement and friendship throughout this whole study. I couldn t finish my research without his great support. I would like to express deepest appreciation to my co-major professor, Dr. Kai Huang, for his technical support and patient guidance in Latex and Matlab. I would also like to thank the member of my committee, Dr. Florence George for her time, valuable advice and great encouragement. In addition, I would like to thank the Department of Mathematics and Statistics, all the professors who supported and encouraged me throughout my life in FIU. I wish to thank aunt Miao, Dr. Jie Mi s wife. She gave me substantial big help in my life. I also thank my classmates, friends. I am so happy I have them around me. Finally, I wish to thank my family for their strong support and their absolute confidence in me. iii
5 ABSTRACT OF THE THESIS BAYSIAN ESTIMATION OF SMALL PROPORTIONS USING BINOMIAL GROUP TEST by Shihua Luo Florida International University, 212 Miami, Florida Professor Jie Mi, Co-Major Professor Professor Kai Huang, Co-Major Professor Group testing has long been considered as a safe and sensible relative to oneat-a-time testing in applications where the prevalence rate p is small. In this thesis, we applied Bayes approach to estimate p using Beta-type prior distribution. First, we showed two Bayes estimators of p from prior on p derived from two different loss functions. Second, we presented two more Bayes estimators of p from prior on π according to two loss functions. We also displayed credible and HPD interval for p. In addition, we did intensive numerical studies. All results showed that the Bayes estimator was preferred over the usual maximum likelihood estimator (MLE) for small p. We also presented the optimal β for different p,m, and k. Keywords: Group Test, Bayes Estimator, Beta Distribution, MLE, MSE. iv
6 TABLE OF CONTENTS CHAPTER PAGE 1 Introduction Bayes Inferences from Prior on p Bayes Point Estimators of p Approximation to Bayes Estimators of p Comparison in the Special Case of Y= Credible and HPD Interval for p Bayes Estimators of p from Prior on π Numerical Studies Conclusion REFERENCES APPENDIX v
7 LIST OF TABLES TABLE PAGE 1 Optimal β For k = 5, α = 1, N = Optimal β For k = 1, α = 1, N = Optimal β For k = 15, α = 1, N = MSE For p =.5, k = 5, m = 5, α = 1, N = MSE For p =.1, k = 5, m = 5, α = 1, N = MSE For p =.5, k = 5, m = 15, α = 1, N = MSE For p =.1, k = 5, m = 15, α = 1, N = MSE For p =.5,.6,.7,.8, k = 5, α = 1, N = MSE For p =.9,.1,.2,.3, k = 5, α = 1, N = MSE For p =.4,.5,.6,.7, k = 5, α = 1, N = MSE For p =.8,.9,.1, k = 5, α = 1, N = MSE For p =.5,.6,.7,.8, k = 1, α = 1, N = MSE For p =.9,.1,.2,.3, k = 1, α = 1, N = MSE For p =.4,.5,.6,.7, k = 1, α = 1, N = MSE For p =.8,.9,.1, k = 1, α = 1, N = MSE For p =.5,.6,.7,.8, k = 15, α = 1, N = MSE For p =.9,.1,.2,.3, k = 15, α = 1, N = MSE For p =.4,.5,.6,.7, k = 15, α = 1, N = MSE For p =.8,.9,.1, k = 15, α = 1, N = vi
8 LIST OF FIGURES FIGURE PAGE 1 Optimal β for k = 5, m = Optimal β for k = 5, m = Optimal β for k = 5, m = 1 or Optimal β for k = 1, m = Optimal β for k = 5 or 1, m = vii
9 1 Introduction Almost daily, the news media report the results of some poll. Pollsters often give some data, for example, the percentage of people in favor of the current president, the rate of people who smoke cigarettes, the death rate of a certain disease, the percent of people who have hepatitis C virus or a rare disease, the rate of contaminated fish in a river, the proportion of defective products, et cetra. All of these issues have a common property the ratio of a part to a whole. How can we estimate this proportion p? In estimating the proportion of subjects that have a certain characteristic in a population in which each individual can be classified into two categories, say with or without certain characteristic, one logical method of estimating p for the population is to use the proportion of successes meaning with that characteristic in the sample. For example, how would we estimate the true fraction of all U.S. citizens who trust the president? We can estimate the p by calculating ˆp=x/n: where x is the number of people in the sample who trust the president and n is the number of people sampled. Experimenters used this method which is called the maximum likelihood estimator (MLE) for many years. But this method has some disadvantages when p is very small. For instance, if researchers want to estimate the prevalence of human immunodeficiency virus (HIV), the ratio ˆp is extremely small and so the estimation is not stable in using the MLE method. Then we need to test every sample unit to get the estimate of ratio p. In addition, the examination of the individual members of a large sample is an expensive and tedious work. To overcome these shortcomings, Dorfman (1943) first proposed the method of group testing and applied it to testing 1
10 blood samples for the syphilis antigen. The idea of group testing is as follows. Assuming that a random sample of size n is drawn from a target population in which some units have certain characteristic that we are interested in, then we divide this sample into m groups with k units in each group. It is assumed that if at least one unit in a group has that characteristic (positive), then this group will be positive; on the other hand, if none of the units in a group is positive, then this group will be negative. The key assumption is valid for many situations in biological, environmental and medical studies. Let p denote the probability of a unit being positive in a certain population, then the probability of a group being positive, denoted as π, can be expressed in terms of p, π = 1 (1 p) k. Instead of directly estimating p from examining each unit in a sample, group testing will estimate π and then p. Instead of conducting N chemical analyses are required to check out all members of a population of size N, Dorfman (1943) applied group testing for checking the syphilis antigen. Suppose that after the blood serum is drawn for each individual, all the blood samples are pooled in groups of five, and the group rather than the individual serum are subjected to chemical analysis. If none of the five sera contains syphilis antigen, the result of this group will be negative. If the result of a group shows positive, it means that this group includes at least one individual of syphilis antigen. Then the individuals making up the pool must be retested to decide which of the members are infected. He also discussed related costs for selected prevalence rates and optimum group sizes. He suggested that the prevalence rate be sufficiently 2
11 small to make worthwhile economics possible. Afterwards, many scientists used group testing in their study fields. Gibbs and Gower (196) applied this method to estimate the frequency of success of attempts to transmit a virus disease from one plant to another. Chiang and Reeves (1962) used this approach to estimate the infection rates of mosquitoes. Bhattacharyya et al.(1979) employed group testing to give point estimates and confidence intervals for infection rates. Worlund and Taylor (1983) applied the procedure to evaluate the disease incidence in a large population of fish. Swallow (1985) used this method to estimate infection rates and probabilities of pathogen transmission by a single vector. To evaluate human immunodeficiency virus (HIV) seroprevalence in population surveys, pooled sera was used by Kline et al. (1989). Rodoni et al. (1994) applied a sequential batch testing procedure combined with Enzyme-Linked Immunosorbent Assay (ELISA) to estimate levels of virus incidence in Victoria cut-flower sim carnation. Hepworth (1996) constructed confidence intervals for proportion of infected unites in a population involving unequaled sized group. Gastwirth (2) proposed that mutations in individual patients could be tested more fruitful by being checked in pools. Xie et al. (21) used group testing to develops models and estimation procedures in order to obtain quantitative information from data in the process to discovery and development of a new drug. Katholi and Unnasch (26) discussed the suitable sampling protocol about important experimental parameters for deciding infection rates. Because of its wide applications, group testing is also known as pooled testing, or composite sampling, in different fields. In recent years, more and more scholars recognized the advantage of the group testing 3
12 experimental design. It is well known that this method is economical, time saving, and efficient particularly when it is applied to estimating small proportion of a disease or some other rare attributes. The traditional way to estimate π or p is using the method of maximum likelihood (MLE). The definition of MLE is as follows. If x 1, x 2,..., x n are independent and identically distributed observation from a population with pdf or pmf f(x; θ), where x = (x 1, x 2,..., x n ) is the fixed sample point, and θ is the unknown parameter of interest, the likelihood function is defined as the product of all the pdfs or pmfs f(x i ; θ). For the sample point x, let θ(x) be a parameter value at which likelihood function attains its maximum as a function of θ, with x held fixed. Then θ(x) is the MLE of θ. An alternative way to estimate π or p is applying the Bayes method. The Bayesian approach is totally different from the approach used in MLE analyse. In classical methods the parameter to be estimated is thought to be an unknown, but fixed quantity. The Bayesian approach treats the unknown parameter as a random variable. In many problems the investigator has some prior information about the unknown parameter. For instance, in sampling inspection, the quality engineer has some idea about the true defective rate, the unknown parameter, of a production process from past experience. The Bayesian approach assumes that this prior knowledge can be summarized in the form of a probability distribution on unknown parameter, called the prior distribution. This is a subjective distribution, based on the experimenter s belief, and is formulated before the data are seen. A sample is then taken 4
13 from a population indexed by the unknown parameter θ and the prior distribution is updated with this sample information. The updated prior distribution is named the posterior distribution. The posterior distribution is now used to make inference about θ. To use the Bayes method, we also need a loss function. The most common loss function is the square loss function. If we use this square loss function, a natural estimator for θ is the mean of the posterior distribution. If necessary, of course, we can use alternative loss function, but the square loss function is the most popular one. Usually, the posterior distribution will become very complex with each added measurement. Whereas, if we choose a conjugate prior, then the posterior distribution will be easier to obtain. A conjugate prior is defined as a prior distribution belonging to some parametric family, for which the resulted posterior distribution also belongs to the same family. This is an important property. Such prior distributions have been identified for some standard distributions f(x θ). When we want Bayes estimator for proportion θ = p, we assume the prior distribution on parameter θ is Beta (α, β). The posterior distribution is also a Beta distribution function which makes the estimation of parameter p much more convenient. Some researchers prefer to use classical Bayesian approaches because the Bayesian method gave a better point estimator of p, especially when p is very small. Gastwirth et al. (1991) suggested classical Bayesian approaches in group testing. Chaubey and Li (1995) researched the difference between MLE and Bayes method for estimation of binomial probability with group testing. They observed that the Bayes methodology 5
14 gave an alternative choices to the experimenter with possible reduction in cost as well as error. Chick (1996) has used this method to do his studies. In the current study, we want to do further study on estimating p by using Bayes methods. We will confine our inference about p to the range (,.1). We will try two Bayesian estimators corresponding to two different loss functions with assumption that p has Beta prior distribution. In addition, we will try to obtain a credible and HPD interval for p. Assuming Beta prior on the probability π of a group being positive, we will investigate two Bayes estimators corresponding to two loss functions. The criterion for comparing the accurate of estimators is its M SE (mean squared error). The performance of the proposed estimators will be studied based on heavy Monte Carlo simulation. The proposed Bayes estimators will also be compared with other estimator existed in the literature. 6
15 2 Bayes Inferences from Prior on p In this section, we will consider Bayes inferences about p. It is assumed that p has Beta prior distribution Beta(α, β) with density function f P (p) = 1 B(α, β) pα 1 (1 p) β 1, < p < 1, α >, β >. 2.1 Bayes Point Estimators of p Suppose that a random sample of size n = mk is obtained from a population of screened subjects, where m, k are positive integers. We assume that p is the probability that a randomly selected individual s test is positive. In this research our concern is the inference about small p, say < p.1. The n subjects are classified into m groups each of which includes k subjects. Throughout this research it is assumed that any one of the m groups is tested as positive if and only if it includes at least one positive subject. Let Y be the number of positive groups. It is known that Y B(m, π) where π = 1 (1 p) k, i.e., P (Y = y) = ( m y ) π y (1 π) m y. Clearly, the joint distribution of (Y, p) is f Y,P (y, p) = = ( ) 1 m B(α, β) pα 1 (1 p) β 1 π y (1 π) m y y ( m ) y B(α, β) pα 1 (1 p) β 1( 1 (1 p) k) y (1 p) k(m y) (2.1) Thus the marginal distribution of Y is given as f Y (y) = f Y,P (y, p)dp 7
16 ( m y ) = p α 1 (1 p) k(m y)+β 1( 1 (1 p) k) y dp B(α, β) ( m ) y ( ) y y = p α 1 (1 p) k(m y)+β 1 ( 1) j (1 p) kj dp B(α, β) j ( ) m (B(α, ) 1 y ( ) y 1 = β) ( 1) j p α 1 (1 p) k(m+j y)+β 1 dp y j ( ) m (B(α, ) 1 y ( ) y = β) ( 1) j B ( α, k(m + j y) + β ) (2.2) y j Therefore, the posterior distribution of p is f P Y (p y) = f Y,P (y, p) f Y (y) ( m )( ) 1p y B(α, β) α 1 (1 p) k(m y)+β 1( 1 (1 p) k) y = ( m )( ) 1 y y B(α, β) ( 1) j( ) ( ) y B α, k(m + j y) + β = pα 1 (1 p) k(m y)+β 1( 1 (1 p) k) y y ( 1) j( ) ( ) y B α, k(m + j y) + β j j (2.3) The Bayes estimator of p for the case of square loss function L 1 (p, a) = (p a) 2 is derived by Chaubey and Li ( 1995) as y ( y ) j ( 1) j B(α + 1, km + kj ky + β) ˆp B1 = y ( y ) ( 1)j B(α, km + kj ky + β) j. (2.4) Specifically, if y =, then ˆp B1 = α n + α + β. Now consider loss function L 2 (p, a) = p 1 (p a) 2 and will derive the corresponding Bayes estimator ˆp B2. 8
17 Note that if loss function has the form L(p, a) = w(p)(p a) 2, then the corresponding Bayes estimator is given by ˆP BL = w(p)pf P Y (p y)dp w(p)f P Y (p y)dp = E P Y (w(p)p y) E P Y (w(p) y) (2.5) According to Equation (2.5) we have to evaluate w(p)pf P Y (p y)dp and w(p)f P Y (p y)dp. For w(p) = p 1 in loss function L 2 (p, a) we have and w(p)pf P Y (p y)dp = p 1 pf P Y (p y)dp = 1 w(p)f P Y (p y)dp = = p 1 pα 1 (1 p) k(m y)+β 1( 1 (1 p) k) y y ( 1) j( ) ( ) dp y B α, k(m + j y) + β j pα 2 (1 p) k(m y)+β 1( 1 (1 p) k) y dp y ( 1) j( ) ( ) y B α, k(m + j y) + β j (2.6) The value of the integral of the numerator in (2.6) is = k 1 = i= p α 2 (1 p) k(m y)+β 1( 1 (1 p) k) y dp (2.7) (1 u) α 2 u k(m y)+β 1 (1 u k ) y du (y > ) u k(m y)+i+β 1 (1 u k ) y 1 (1 u) α 1 du 9
18 k 1 = i= i= ( y 1 ( ) y 1 u )( 1) k(m y)+i+β 1 j u kj (1 u) α 1 du j k 1 y 1 ( ) y 1 1 = ( 1) j j i= u k(m y)+i+β 1+kj (1 u) α 1 du k 1 y 1 ( ) y 1 = ( 1) j B(km + kj + i + β ky, α) j Therefore, if y > we have ˆp B2 = k 1 y ( 1) j( y j) B(α, km + kj + β ky) y 1 i= j=i ( 1) j( y 1 j ) B(km + kj + i + β ky, α) (2.8) In the case y =, we let α > 1 and notice that equation (2.7) becomes p α 2 (1 p) km+β 1 dp = B(α 1, km + β), (α > 1). Summarizing the above, we obtain Theorem 2.1. Let p have prior distribution B(α, β). For the loss function L 2 (p, a) = p 1 (p a) 2 the Bayes estimator of p, denoted as ˆp B2, is given as ˆp B2 = y ( 1) j ( y j)b(α,km+kj+β ky) k 1 y 1 ( 1) j ( y 1 j )B(km+kj+β+i ky,α) i= B(α,km+β) B(α 1,km+β) if y = and α > 1. if y > 1
19 2.2 Approximation to Bayes Estimators of p It is easy to show that for sufficiently large N it holds that Γ(N + a) Γ(N + b) N a b (2.9) As it was mentioned in the previous section the group test is applied for the case of large sample size n = mk. Thus we could simplify the expression of ˆp B1 and ˆp B2 First we examine the expression (2.4) for ˆp B1. We have B(α, km + kj ky + β) = Γ(α)Γ(km + kj ky + β) Γ(km + kj ky + β + α) Γ(α)(km + kj ky + β) α since y is usually a small number and n = km is large. Similarly we have B(α + 1, km + kj ky + β) Γ(α + 1)(km + kj ky + β) α 1 Hence we obtain ˆp B1 = y ( y ) j ( 1) j B(α + 1, km + kj ky + β) y ( y ) ( 1)j B(α, km + kj ky + β) j Γ(α + 1) y Γ(α) y ( y j ( y ) j ( 1) j (km + kj ky + β) α 1 ) ( 1)j (km + kj ky + β) α 11
20 = α y ( 1) j( ) y (km + kj ky + β) α 1 j y ( 1) j( ) y (km + kj ky + β) α j where ˆp B 1 is an approximation to the Bayes estimator ˆp B1. As for ˆp B2 notice that if y 1, then ˆp B 1 (2.1) Γ(α)Γ(km + kj + i + β ky) B(km + kj + i + β ky, α) = Γ(km + kj + i + β ky + α) Γ(α)(km + kj + i + β ky) α Therefore, if y 1 it holds from Theorem 2.1 that ˆp B2 = k 1 y ( 1) j( ) y Γ(α)(km + kj ky + β) α i= k 1 j y 1 ( 1) j( ) y 1 Γ(α)(km + kj + i + β ky) α j y ( 1) j( ) y (km + kj ky + β) α i= j ˆp y 1 ( 1) j( ) B 2 (2.11) y 1 (km + kj + i + β ky) α On the other hand if y = and α > 1, then we have j B(α, km + β) ˆp B2 = B(α 1, km + β) (α 1)Γ(km + α + β 1) = Γ(km + α + β) Γ(α)Γ(km + β) Γ(km + α + β 1) = Γ(km + α + β) Γ(α 1)Γ(km + β) = α 1 km + α + β 1 = α 1 n + α + β 1 if α > 1. (2.12) From (2.11) and (2.12) we see that the Bayes estimator ˆp B2 can be aproximated by 12
21 ˆp B 2 = y ( 1) j ( y j)(km+kj ky+β) α k 1 y 1 ( 1) j ( y 1 i= j )(km+kj+i+β ky) α, if y 1; α 1, if y = and α > 1. n+α+β 1 (2.13) 2.3 Comparison in the special case of Y= Note the fact that the event (Y = ) is the same as event (X = ) where X is the number of positive subjects when subjects are tested individually. So it is interesting to compare the Bayes estimators based on X = and Y =. Clearly X B(n, p) = B(mk, p). Assuming that the proportion p has Beta(α, β) as its prior distribution, it is well known that the posterior distribution of p given X = x is Beta(x + α, n x + β). Suppose that square loss function L 1 (p, a) = (p a) 2 is used, then the Bayes estimator ˆp B1 based on X = is ˆp B1 = E P X (p x = ) = α n + α + β In the mean time the Bayes estimator ˆp B1 based on Y = is exactly the same as ˆp B1.(bottom of p8) Now consider the loss function L 2 (p, a) = p 1 (p a) 2. For this loss function the Bayes estimator ˆp B2 based on X can be obtained as follows. We have w(p)pf P X (p x)dp = f P X (p x)dp = 1 and 13
22 w(p)f P X (p x)dp = p 1( B(x + α, n x + β) ) 1 p x+α 1 (1 p) n x+β 1 dp = ( B(x + α, n x + β) ) 1 p x+α 2 (1 p) n x+β 1 dp Consider two cases. If x 1 which is equivalent to y 1. In this case ( B(x + α, n x + β) w(p)f P X (p x)dp = B(x + α 1, n x + β) ( Γ(x + α)γ(n x + β) = Γ(n + α + β) ( ) 1 x + α 1 = n + α + β 1 ) 1 Γ(n + α + β 1) Γ(x + α 1)Γ(n x + β) ) 1 Thus ˆp B2 = x + α 1 n + α + β 1 However, if x =, then it must be assumed that α > 1 and so w(p)f P X (p )dp = ( B(α, n + β) ) 1 B(α 1, n + β) = B(α, n + β) p α 2 (1 p) n+β 1 dp Γ(α 1)Γ(n + β) Γ(n + α + β) = Γ(n + α + β 1) Γ(α)Γ(n + β) = n + α + β 1 α 1 Consequently, it follows that 14
23 ˆp B2 = α 1, if x = and α > 1. n + α + β 1 Therefore ˆp B2 = x+α 1 if x 1; n+α+β 1 From this and equation (2.13), i.e., α 1 if x = and α > 1. n+α+β 1 ˆp B 2 = α 1 n + α + β 1 if y = and α > 1. we see that in this case ˆp B 2 and ˆp B2 are exactly the same when there is no positive subject in the entire sample. 2.4 Credible and HPD Interval for p Suppose that p has prior distribution Beta(α, β). From equation (2.3) it is known that the posterior distribution of p given Y = y is f P Y (p y) = pα 1 (1 p) k(m y)+β 1( 1 (1 p) k) y y ) ( 1)j B(α, k(m + j y) + β) Then for any set A (, 1), the credibal probability of A is ( y j P (p A) = f P Y (p y)dp and A is a credible set for p. By numerical computation it is not difficult to obtain a 1 γ credible interval for p. A 15
24 In a special case when α = 1, i.e., p has Beta(1, β) as its prior distribution, by equation (2.2) the marginal distribution of Y has density f Y (y) = β(1 p) β 1 ( m y ) (1 (1 p) k ) y (1 p) k(m y) dp = β ( ) m B (m y + βk ) k y, y + 1 (2.14) and consequently the posterior distribution of p given Y = y has density f P Y (p y) = β( ) ( m y (1 p) k(m y)+β 1 1 (1 p) k) y ) ( B m y + β, y + 1) k βk 1( m y = k(1 p)k(m y)+β 1( 1 (1 p) k) y B ( (2.15) m y + β, y + 1) k We further let β 1. The assumption is reasonable because we are dealing with small p and the majority of the probability distribution Beta(1, β) is close to zero if β is relatively large. With the assumption it holds that m y + (β 1)/k. It can be shown easily that the posterior f P Y (p y) is unimodal. Therefore, the HPD interval for p can be obtained by {p : f P Y (p y) a} where {p:f P Y (p y) a} f P Y (p y)dp = 1 γ for any given γ (, 1). For the case < β < 1, the posterior density f P Y (p y) strictly decreases in p (, 1), so the HPD interval for p can be obtained as 16
25 {p : f P Y (p y) a} where a f P Y (p y)dp = 1 γ. 3 Bayes Estimators of p from Prior on π In the present section we will assume π 1 (1 p) k has a Beta(α, β) prior distribution. On the basis of the observation Y = y we first calculate the Bayes estimator ˆπ B, then derive the Bayes estimator of p. Following Chaubey and Li (1995) this procedure is called the indirect Bayes procedure, the produced estimators is called the indirect Bayes estimators and denoted as ˆp IB1 or ˆp IB2 depending on the loss function L 1 (p, a) = (p a) 2 and L 2 (p, a) = p 1 (p a) 2. Obviously the posterior distribution of π given Y = y is Beta(y+α, m y+β), i.e., f π Y (π y) = ( B(y + α, m y + β) ) 1 π y+α 1 (1 π) m y+β 1 From π = 1 (1 p) k = π(p) we see that p = p(π) = 1 (1 π) 1/k. Consider loss function L(p, a) = w(p)(p a) 2. It is easy to see that the corresponding indirect Bayes estimator ˆp IB is given as ˆp IB = E( pw(p) Y = y ) E ( w(p) Y = y ) For the case L 1 (p, a) = (p a) 2 the indirect Bayes estimator ˆp IB1 is given in Chaubey and Li (1995) as 17
26 Γ(m + α + β)γ(m y + β + 1/k) ˆp IB1 = 1 Γ(m y + β)γ(m + α + β + 1/k) (3.1) which can be further approximated as ˆp IB 1 = 1 ( ) 1/k m + β y (3.2) m + α + β Now let us consider loss function L 2 (p, a) = p 1 (p a) 2. Clearly E(pw(p) Y = y) = pp 1 f π Y (π y)dπ = f π Y (π y)dπ = 1 and E(w(p) Y = y) = = = p 1 f π (π y)dπ p 1( B(y + α, m y + β) ) 1 π y+α 1 (1 π) m y+β 1 dπ ( 1 (1 π) 1/k ) 1 π y+α 1 (1 π) m y+β 1 dπ B(y + α, m y + β) (3.3) To evaluate the integral on the right side of (3.3) we let u = (1 π) 1/k and have = ( 1 (1 π) 1/k ) 1 π y+α 1 (1 π) m y+β 1 dπ (1 u) 1 (1 u k ) y+α 1 u k(m y+β 1) ku k 1 du 18
27 k 1 =k i= k 1 = i= k 1 = i= k 1 = B i= u k(m y+β)+i 1 (1 u k ) y+α 2 du u k(m y+β)+i k (1 u k ) y+α 2 ku k 1 du v m y+β+i/k 1 (1 v) y+α 2 dv (m y + β + ik, y + α 1 ), here it is required that α > 1 if y =. Therefore, the indirect Bayes estimator ˆp IB2 based on observation Y = y is ˆp IB2 = B(y + α, m y + β) k 1 i= B ( (3.4) m y + β + i, y + α 1) k where α > 1 if y =. We can rewrite ˆp IB2 as ˆp IB2 = k 1 i= Γ(y + α)γ(m y + β)/γ(m + α + β) Γ ( m y + β + i k) Γ(y + α 1)/Γ ( m + α + β + i k 1) =(y + α 1) k 1 i= i= Γ(m y + β)/γ(m + α + β) Γ ( m y + β + i k) /Γ ( m + α + β + i k 1) k 1 Γ(m y + β) =(y + α 1) Γ ( ) Γ ( m + α + β + i 1) k m y + β + i Γ(m + α + β) k (3.5) From (3.5) we see that ˆp IB2 can be approximated by y + α 1 ˆp IB 2 = k 1 i= (m y + β)i/k (m + α + β) i/k+1 19
28 here α > 1 if y =. = y + α 1 m + α + β k 1 i= ( 1 m y+β m+α+β ) i/k (3.6) 2
29 4 Numerical Studies A MATLAB simulation is performed in order to analyze the behaviors of different estimators under several combinations. I chose m = 5, 1, 15, 2, 25 and k = 5, 1, 15. Fifteen cases of p between.5 and.1 are considered. First, I simulate a sample that is from an independent, identically Bernoulli distribution and divided the observations into m groups with k units in each group. I use this new grouped sample to calculate the value of M LE. Second, I use the same sample to calculate the values of Bayes estimator and approximate Bayes estimator under different βs. The process is repeated 1 times. Then I will obtain the MSE of these estimators. Using the MSE as a criterion. Table 1-3 display the optimal βs which yield the smallest M SE of Bayes estimator associated with the square loss function. From these table, we can see that the optimal β decreases with the increase of p; the optimal β increases with the increase of m and k. Figures 1-5 also indicate these trends. Tables 4-7 show how I got the optimal βs. Tables 8-19 compare the MSE of Bayes estimator under the optimal βs with the approximate Bayes estimator as well as MLE. From these tables, we can draw a conclusion that the Bayes estimator is the best one as far as MSE is concerned. MLE is not a good estimator when p is small. Moreover, the approximate Bayes estimator is also much better than the M LE. 21
30 Table 1: k = 5, α = 1, N = 1 p k = 5 m This is the optimal β based on the smallest MSE of Bayes estimator by using k = 5 with combinations of different m and p. 22
31 Table 2: k = 1, α = 1, N = 1 p k = 1 m This is the optimal β based on the smallest MSE of Bayes estimator by using k = 1 with combinations of different m and p. 23
32 Table 3: k = 15, α = 1, N = 1 p k = 15 m This is the optimal β based on the smallest MSE of Bayes estimator by using k = 15 with combinations of different m and p. 24
33 Table 4: k = 5, m = 5, p =.5, α = 1, N = 1 β MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 β MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 β MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 Table 5: k = 5, m = 5, p =.1, α = 1, N = 1 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 The comparison of M SE of Bayes estimator and approximation to the Bayes estimator under different βs. 25
34 Table 6: k = 5, m = 15, p =.5, α = 1, N = 1 β MSE(ˆp B1 ) 4.451E E E E E-6 MSE(ˆp B 1) 4.718E E E E E-6 β MSE(ˆp B1 ) 4.366E E E E E-6 MSE(ˆp B 1) 4.594E E E E E-6 β MSE(ˆp B1 ) 4.383E E E E E-6 MSE(ˆp B 1) 4.575E E E E E-6 Table 7: k = 5, m = 15, p =.1, α = 1, N = 1 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 β MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 The comparison of M SE of Bayes estimator and approximation to the Bayes estimator under different βs. 26
35 Table 8: k = 5, p =.5,.6,.7,.8, α = 1, N = 1 p =.5 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 MSE(ˆp MLE ) 2.414E E E E E-5 p =.6 MSE(p B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 MSE(ˆp MLE ) 2.972E E E E E-5 p =.7 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 MSE(ˆp MLE ) E E E E E-5 p =.8 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) 7.632E E E E E-5 MSE(ˆp MLE ) E E E E E-5 Table-8 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 5 with different combinations of p =.5,.6,.7,.8 and m = 5, 1, 15, 2,
36 Table 9: k = 5, p =.9,.1,.2,.3, α = 1, N = 1 p =.9 MSE(ˆp B1 ) 1.42E E E E E-5 MSE(ˆp B 1) 1.544E E E E E-5 MSE(ˆp MLE ) 4.357E E E E E-5 p =.1 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) E E E E E-5 p =.2 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) 7.272E E E E E-5 MSE(ˆp MLE ) E E E E E-4 p =.3 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-9 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 5 with different combinations of p =.9,.1,.2,.3 and m = 5, 1, 15, 2,
37 Table 1: k = 5, p =.4,.5,.6,.7, α = 1, N = 1 p =.4 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.5 MSE(ˆp B1 ) 4.785E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.6 MSE(ˆp B1 ) 6.619E E E E E-4 MSE(ˆp B 1) 6.974E E E E E-4 MSE(ˆp MLE ) 4.449E E E E E-4 p =.7 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-1 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 5 with different combinations of p =.4,.5,.6,.7 and m = 5, 1, 15, 2,
38 Table 11: k = 5, p =.8,.9,.1, α = 1, N = 1 p =.8 MSE(ˆp B1 ) 1.716E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) 7.835E E E E E-4 p =.9 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.1 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-11 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 5 with different combinations of p =.8,.9,.1 and m = 5, 1, 15, 2, 25. 3
39 Table 12: k = 1, p =.5,.6,.7,.8, α = 1, N = 1 p =.5 MSE(ˆp B1 ) 3.336E E E E E-6 MSE(ˆp B 1) 3.358E E E E E-6 MSE(ˆp MLE ) E E E E E-5 p =.6 MSE(ˆp B1 ) 5.462E E E E E-6 MSE(ˆp B 1) 5.498E E E E E-6 MSE(ˆp MLE ) E E E E E-5 p =.7 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) 8.65E E E E E-6 MSE(ˆp MLE ) 1.877E E-5 5.3E E E-5 p =.8 MSE(ˆp B1 ) 1.768E E E E E-5 MSE(ˆp B 1) 1.852E E E E E-5 MSE(ˆp MLE ) 2.946E E E E E-5 Table-12 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 1 with different combinations of p =.5,.6,.7,.8 and m = 5, 1, 15, 2,
40 Table 13: k = 1, p =.9,.1,.2,.3, α = 1, N = 1 p =.9 MSE(ˆp B1 ) 1.495E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) E E E E E-5 p =.1 MSE(ˆp B1 ) 1.821E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) E E E E E-5 p =.2 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) 8.445E E E E E-5 MSE(ˆp MLE ) 7.779E E E E E-5 p =.3 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) 2.751E E E E E-4 Table-13 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 1 with different combinations of p =.9,.1,.2,.3 and m = 5, 1, 15, 2,
41 Table 14: k = 1, p =.4,.5,.6,.7, α = 1, N = 1 p =.4 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) 4.797E E E E E-4 p =.5 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.6 MSE(ˆp B1 ) 5.555E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) 2.122E E E E E-4 p =.7 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-14 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 1 with different combinations of p =.4,.5,.6,.7 and m = 5, 1, 15, 2,
42 Table 15: k = 1, p =.8,.9,.1, α = 1, N = 1 p =.8 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.9 MSE(ˆp B1 ) 1.383E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.1 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-15 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 1 with different combinations of p =.8,.9,.1 and m = 5, 1, 15, 2,
43 Table 16: k = 15, p =.5,.6,.7,.8, α = 1, N = 1 p =.5 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) 4.189E E E E E-6 MSE(ˆp MLE ) E E E E E-5 p =.6 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) 6.431E E E E E-6 MSE(ˆp MLE ) 1.678E E E E E-5 p =.7 MSE(ˆp B1 ) E E E E E-6 MSE(ˆp B 1) E E E E E-6 MSE(ˆp MLE ) E E E E E-5 p =.8 MSE(ˆp B1 ) 1.217E E E E E-5 MSE(ˆp B 1) 1.299E E E E E-5 MSE(ˆp MLE ) E E E E E-5 Table-16 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 15 with different combinations of p =.5,.6,.7,.8 and m = 5, 1, 15, 2,
44 Table 17: k = 15, p =.9,.1,.2,.3, α = 1, N = 1 p =.9 MSE(ˆp B1 ) 1.586E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) E E E E E-5 p =.1 MSE(ˆp B1 ) 1.929E E E E E-5 MSE(ˆp B 1) E E E E E-5 MSE(ˆp MLE ) 1.99E E E E E-5 p =.2 MSE(ˆp B1 ) E E E E E-5 MSE(ˆp B 1) 7.822E E E E E-5 MSE(ˆp MLE ) E E E E E-5 p =.3 MSE(ˆp B1 ) E E E-5 8.8E E-5 MSE(ˆp B 1) 1.683E E E E E-5 MSE(ˆp MLE ) E E E E E-4 Table-17 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 15 with different combinations of p =.9,.1,.2,.3 and m = 5, 1, 15, 2,
45 Table 18: k = 15, p =.4,.5,.6,.7, α = 1, N = 1 p =.4 MSE(ˆp B1 ) 2.481E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.5 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.6 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) 5.21E E E E E-4 MSE(ˆp MLE ) E E E E E-4 p =.7 MSE(ˆp B1 ) 6.63E E E E E-4 MSE(ˆp B 1) E E E E E-4 MSE(ˆp MLE ) E E E E E-4 Table-18 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 15 with different combinations of p =.4,.5,.6,.7 and m = 5, 1, 15, 2,
46 Table 19: k = 15, p =.8,.9,.1, α = 1, N = 1 p =.8 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) 7.728E E E E E-4 MSE(ˆp MLE ) E E E E E-3 p =.9 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) 9.28E E E E E-4 MSE(ˆp MLE ) 2.59E E E E E-3 p =.1 MSE(ˆp B1 ) E E E E E-4 MSE(ˆp B 1) 1.679E E E E E-4 MSE(ˆp MLE ) E E E E E-3 Table-19 compares the M SE of Bayes estimator and approximation to the Bayes estimator with the MSE of the MLE by using k = 15 with different combinations of p =.8,.9,.1 and m = 5, 1, 15, 2,
47 Figure 1: K = 5, m = 1 Figure 2: K = 5, m = 2 39
48 Figure 3: K = 5, m = 1 or 2 Figure 4: K = 1, m = 1 4
49 Figure 5: K = 5 or 1, m = 1 41
50 5 Conclusion According to the analytical and numerical results obtained above, We know that the M SE of Bayes estimators and approximation to Bayes estimators are both smaller than that of MLE for small p. Indirect Bayes estimators performs very well too. We could draw a conclusion that Bayes estimators are much better than MLE when p is small. 42
51 References [1] Bhattacharyya, F. K., Karandinos, M. G., Defoliart, G. R., Point estimates and confidence interval for infection rates using pooled organisms in epidemiologic studies. American Journal of Epidemiology. 19, [2] Chaubey, Y., Li, W., Comparison between maximum likelihood and Bayes methods for estimation of binomial probability with sample compositing. Journal of Official Statistics. 11, [3] Chiang, C. L., Reeves, W. C., Statistical estimation of virus infection rates in mosquito vector populations. American Journal of Hygiene. 75, [4] Chick, S., Bayesian models for limiting dilution assay and group test data. Biometrics. 52, , [5] Dorfman, R., The detection of defective members of large populations. Annals of Mathematical Statistics. 14, [6] Gastwirth, J., Johnson, W., Reneau, D., Bayesian analysis of screening data: Application to AIDS in blood donors. The Canadian Journal of Statistics. 19, [7] Gastwirth, J., 2. The efficiency of pooling in the detection of rare mutations. Americal Journal of Human Genetics. 67, [8] Gibbs, A. J., Gower, J. C., 196. The use of a multiple-transfer method in plant virus transission studies-some statistical points arising in the analysis of results. Annals of Applied Biology. 48, [9] Hepworth, G., Exact confidence intervals for proportions estimated by group testing. Biometrics. 52,
52 [1] Katholi, C. R., Unnasch, T. R., 26. Important experimental parameters for determining infection rates in arthropod vectors using pool screening approaches. The American Society of Tropical Medicine and Hygiene. 74, [11] Kline, R. L., Brothers, T. A., Brookmeyer, R., Zeger, S., Quinn, T. C., Evaluation of human immunodficiency virus seroprevalence in population surveys using pooled sera. Journal of Clinical Microbiology. 27, [12] Radoni, B. C., Hepworth, G., Richardson, C., Moran, J. R., The use of a sequential batch testing procedure and ELISA to determine the incidence of five viruses in Victorian cut-flower sim Carnations. Aust. J. Agric. Res.. 45, [13] Swallow, W. H., Group testing for estimating infection rates and probabilities of disease transmission. Phytopathology. 75, [14] Worlund, D. D., Taylor, G., Estimation of disease incidence in fish populations. Canadian Journal of Fisheries and Aquatic Science. 4, [15] Xie, M., Tatsuoka, K., Sacks, J., Young, S., 21. Group testing with blockers and synergism. J. Am. Statist. Assoc.. 96,
53 APPENDIX I list the equations in this study here for the convenience of finding them. 45
54 1 Bayes Inferences from Prior on p. Loss function L 1 (p, a) = (p a) 2. y ( y ) j ( 1) j B(α + 1, km + kj ky + β) ˆp B1 = y ( y ) ( 1)j B(α, km + kj ky + β) j. y 1 Equation 2.4 page 8 Approximation to the Bayes estimator ˆp B1. ˆp B 1 = α y ( 1) j( ) y (km + kj ky + β) α 1 j y ( 1) j( ) y (km + kj ky + β) α j. y 1 Equation 2.1 page 12 Loss function L 2 (p, a) = p 1 (p a) 2. ˆp B2 = y ( 1) j ( y j)b(α,km+kj+β ky) k 1 y 1 ( 1) j ( y 1 j )B(km+kj+β+i k,α) i= y ( 1) j ( y j)b(α,km+β+kj ky) if y 1 B(α 1,km+β) if y = and α > 1. Approximation to the Bayes estimator ˆp B2. ˆp B 2 = y ( 1) j ( y j)(km+kj ky+β) α k 1 y 1 ( 1) j ( y 1 i= j )(km+kj+i+β ky) α, if y 1; α 1, if y = and α > 1. n+α+β 1 Theorem 1 page 1 Equation 2.13 page 13 46
55 2 Bayes Estimators of p from Prior on π. Loss function L 1 (p, a) = (p a) 2. ˆp IB1 = 1 Γ(m + α + β)γ(m y + β + 1/k) Γ(m y + β)γ(m + α + β + 1/k) Equation 3.1 page 18 Approximation to the Bayes estimator ˆp IB1. ˆp IB 1 = 1 ( ) 1/k m + β y Equation 3.2 page 18 m + α + β Loss function L 2 (p, a) = p 1 (p a) 2. k 1 Γ(m y + β)γ ( m + α + β + i k ˆp IB2 = (y + α 1) 1) Γ ( ) Equation 3.5 page 19 m y + β + i k Γ(m + α + β) i= Approximation to the Bayes estimator ˆp IB2. ˆp IB 2 = y + α 1 m + α + β k 1 i= ( 1 m y+β m+α+β ) i/k Equation 3.6 page 2 47
Inferences about Parameters of Trivariate Normal Distribution with Missing Data
Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 7-5-3 Inferences about Parameters of Trivariate Normal Distribution with Missing
More informationBayes Factors for Grouped Data
Bayes Factors for Grouped Data Lizanne Raubenheimer and Abrie J. van der Merwe 2 Department of Statistics, Rhodes University, Grahamstown, South Africa, L.Raubenheimer@ru.ac.za 2 Department of Mathematical
More informationInference for a Population Proportion
Al Nosedal. University of Toronto. November 11, 2015 Statistical inference is drawing conclusions about an entire population based on data in a sample drawn from that population. From both frequentist
More informationConsecutively Choosing between Three Experiments for Estimating Prevalence Rate of a Trait with Imperfect Tests
International Journal of Statistics and Applications 07, 7(): 9-06 DOI: 0.9/j.statistics.07070.04 Consecutively Choosing between Three periments for stimating revalence Rate of a Trait with Imperfect Tests
More informationComparison of Some Improved Estimators for Linear Regression Model under Different Conditions
Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-24-2015 Comparison of Some Improved Estimators for Linear Regression Model under
More informationA Very Brief Summary of Bayesian Inference, and Examples
A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X
More informationBayesian Inference for Binomial Proportion
8 Bayesian Inference for Binomial Proportion Frequently there is a large population where π, a proportion of the population, has some attribute. For instance, the population could be registered voters
More informationInference about Reliability Parameter with Underlying Gamma and Exponential Distribution
Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 9-3-211 Inference about Reliability Parameter with Underlying Gamma and Exponential
More informationUsing Probability to do Statistics.
Al Nosedal. University of Toronto. November 5, 2015 Milk and honey and hemoglobin Animal experiments suggested that honey in a diet might raise hemoglobin level. A researcher designed a study involving
More informationDIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS
DIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS THUY ANH NGO 1. Introduction Statistics are easily come across in our daily life. Statements such as the average
More informationHPD Intervals / Regions
HPD Intervals / Regions The HPD region will be an interval when the posterior is unimodal. If the posterior is multimodal, the HPD region might be a discontiguous set. Picture: The set {θ : θ (1.5, 3.9)
More informationStudy on a Hierarchy Model
Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-23-212 Study on a Hierarchy Model Suisui Che Florida International University,
More information1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D.
probabilities, we ll use Bayes formula. We can easily compute the reverse probabilities A short introduction to Bayesian statistics, part I Math 17 Probability and Statistics Prof. D. Joyce, Fall 014 I
More informationBayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017
Bayesian inference Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 10, 2017 1 / 22 Outline for today A genetic example Bayes theorem Examples Priors Posterior summaries
More informationA Study on the Correlation of Bivariate And Trivariate Normal Models
Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 11-1-2013 A Study on the Correlation of Bivariate And Trivariate Normal Models Maria
More informationStatistical Inference Using Progressively Type-II Censored Data with Random Scheme
International Mathematical Forum, 3, 28, no. 35, 1713-1725 Statistical Inference Using Progressively Type-II Censored Data with Random Scheme Ammar M. Sarhan 1 and A. Abuammoh Department of Statistics
More informationBayesian Meta-analysis with Hierarchical Modeling Brian P. Hobbs 1
Bayesian Meta-analysis with Hierarchical Modeling Brian P. Hobbs 1 Division of Biostatistics, School of Public Health, University of Minnesota, Mayo Mail Code 303, Minneapolis, Minnesota 55455 0392, U.S.A.
More informationBAYESIAN ANALYSIS OF DOSE-RESPONSE CALIBRATION CURVES
Libraries Annual Conference on Applied Statistics in Agriculture 2005-17th Annual Conference Proceedings BAYESIAN ANALYSIS OF DOSE-RESPONSE CALIBRATION CURVES William J. Price Bahman Shafii Follow this
More informationHuman or Cylon? Group testing on the Battlestar Galactica
Human or Cylon? Group testing on the Statistics and The story so far Video Christopher R. Bilder Department of Statistics University of Nebraska-Lincoln chris@chrisbilder.com Slide 1 of 37 Slide 2 of 37
More informationProbability and Probability Distributions. Dr. Mohammed Alahmed
Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about
More informationExperimental designs for multiple responses with different models
Graduate Theses and Dissertations Graduate College 2015 Experimental designs for multiple responses with different models Wilmina Mary Marget Iowa State University Follow this and additional works at:
More informationBayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework
HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for
More information9 Bayesian inference. 9.1 Subjective probability
9 Bayesian inference 1702-1761 9.1 Subjective probability This is probability regarded as degree of belief. A subjective probability of an event A is assessed as p if you are prepared to stake pm to win
More informationPS 203 Spring 2002 Homework One - Answer Key
PS 203 Spring 2002 Homework One - Answer Key 1. If you have a home or office computer, download and install WinBUGS. If you don t have your own computer, try running WinBUGS in the Department lab. 2. The
More informationAssessing the Effect of Prior Distribution Assumption on the Variance Parameters in Evaluating Bioequivalence Trials
Georgia State University ScholarWorks @ Georgia State University Mathematics Theses Department of Mathematics and Statistics 8--006 Assessing the Effect of Prior Distribution Assumption on the Variance
More informationLecture 10 and 11: Text and Discrete Distributions
Lecture 10 and 11: Text and Discrete Distributions Machine Learning 4F13, Spring 2014 Carl Edward Rasmussen and Zoubin Ghahramani CUED http://mlg.eng.cam.ac.uk/teaching/4f13/ Rasmussen and Ghahramani Lecture
More informationIntroduction to Bayesian Methods
Introduction to Bayesian Methods Jessi Cisewski Department of Statistics Yale University Sagan Summer Workshop 2016 Our goal: introduction to Bayesian methods Likelihoods Priors: conjugate priors, non-informative
More informationThe Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations
The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture
More informationBayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017
Chalmers April 6, 2017 Bayesian philosophy Bayesian philosophy Bayesian statistics versus classical statistics: War or co-existence? Classical statistics: Models have variables and parameters; these are
More informationObjective Bayesian Hypothesis Testing and Estimation for the Risk Ratio in a Correlated 2x2 Table with Structural Zero
Clemson University TigerPrints All Theses Theses 8-2013 Objective Bayesian Hypothesis Testing and Estimation for the Risk Ratio in a Correlated 2x2 Table with Structural Zero Xiaohua Bai Clemson University,
More informationBayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007
Bayesian inference Fredrik Ronquist and Peter Beerli October 3, 2007 1 Introduction The last few decades has seen a growing interest in Bayesian inference, an alternative approach to statistical inference.
More informationProbability Review - Bayes Introduction
Probability Review - Bayes Introduction Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Advantages of Bayesian Analysis Answers the questions that researchers are usually interested in, What
More informationNew Bivariate Lifetime Distributions Based on Bath-Tub Shaped Failure Rate
Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 10-30-2014 New Bivariate Lifetime Distributions Based on Bath-Tub Shaped Failure
More informationPart 2: One-parameter models
Part 2: One-parameter models 1 Bernoulli/binomial models Return to iid Y 1,...,Y n Bin(1, ). The sampling model/likelihood is p(y 1,...,y n ) = P y i (1 ) n P y i When combined with a prior p( ), Bayes
More informationChapter 5. Bayesian Statistics
Chapter 5. Bayesian Statistics Principles of Bayesian Statistics Anything unknown is given a probability distribution, representing degrees of belief [subjective probability]. Degrees of belief [subjective
More informationBayesian Econometrics
Bayesian Econometrics Christopher A. Sims Princeton University sims@princeton.edu September 20, 2016 Outline I. The difference between Bayesian and non-bayesian inference. II. Confidence sets and confidence
More informationNumerical computation of an optimal control problem with homogenization in one-dimensional case
Retrospective Theses and Dissertations Iowa State University Capstones, Theses and Dissertations 28 Numerical computation of an optimal control problem with homogenization in one-dimensional case Zhen
More informationDS-GA 1002 Lecture notes 11 Fall Bayesian statistics
DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian
More informationPARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.
PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.. Beta Distribution We ll start by learning about the Beta distribution, since we end up using
More informationAN OPTIMAL STRATEGY FOR SEQUENTIAL CLASSIFICATION ON PARTIALLY ORDERED SETS. Abstract
AN OPTIMAL STRATEGY FOR SEQUENTIAL CLASSIFICATION ON PARTIALLY ORDERED SETS Thomas S. Ferguson Department of Mathematics UCLA Los Angeles, CA 90095, USA Curtis Tatsuoka 1,2 Department of Statistics The
More informationBayesian Inference for Normal Mean
Al Nosedal. University of Toronto. November 18, 2015 Likelihood of Single Observation The conditional observation distribution of y µ is Normal with mean µ and variance σ 2, which is known. Its density
More informationIntroduction to Bayesian Methods. Introduction to Bayesian Methods p.1/??
to Bayesian Methods Introduction to Bayesian Methods p.1/?? We develop the Bayesian paradigm for parametric inference. To this end, suppose we conduct (or wish to design) a study, in which the parameter
More informationData Analysis and Uncertainty Part 2: Estimation
Data Analysis and Uncertainty Part 2: Estimation Instructor: Sargur N. University at Buffalo The State University of New York srihari@cedar.buffalo.edu 1 Topics in Estimation 1. Estimation 2. Desirable
More informationParametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012
Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood
More informationESTIMATING STATISTICAL CHARACTERISTICS UNDER INTERVAL UNCERTAINTY AND CONSTRAINTS: MEAN, VARIANCE, COVARIANCE, AND CORRELATION ALI JALAL-KAMALI
ESTIMATING STATISTICAL CHARACTERISTICS UNDER INTERVAL UNCERTAINTY AND CONSTRAINTS: MEAN, VARIANCE, COVARIANCE, AND CORRELATION ALI JALAL-KAMALI Department of Computer Science APPROVED: Vladik Kreinovich,
More informationSingle Sample Means. SOCY601 Alan Neustadtl
Single Sample Means SOCY601 Alan Neustadtl The Central Limit Theorem If we have a population measured by a variable with a mean µ and a standard deviation σ, and if all possible random samples of size
More informationBayesian course - problem set 5 (lecture 6)
Bayesian course - problem set 5 (lecture 6) Ben Lambert November 30, 2016 1 Stan entry level: discoveries data The file prob5 discoveries.csv contains data on the numbers of great inventions and scientific
More informationINFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS. Tao Jiang. A Thesis
INFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS Tao Jiang A Thesis Submitted to the Graduate College of Bowling Green State University in partial fulfillment of the
More informationFoundations of Statistical Inference
Foundations of Statistical Inference Julien Berestycki Department of Statistics University of Oxford MT 2016 Julien Berestycki (University of Oxford) SB2a MT 2016 1 / 20 Lecture 6 : Bayesian Inference
More informationUnobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior:
Pi Priors Unobservable Parameter population proportion, p prior: π ( p) Conjugate prior π ( p) ~ Beta( a, b) same PDF family exponential family only Posterior π ( p y) ~ Beta( a + y, b + n y) Observed
More informationNoninformative Priors for the Ratio of the Scale Parameters in the Inverted Exponential Distributions
Communications for Statistical Applications and Methods 03, Vol. 0, No. 5, 387 394 DOI: http://dx.doi.org/0.535/csam.03.0.5.387 Noninformative Priors for the Ratio of the Scale Parameters in the Inverted
More informationConfidence Intervals for the Ratio of Two Exponential Means with Applications to Quality Control
Western Kentucky University TopSCHOLAR Student Research Conference Select Presentations Student Research Conference 6-009 Confidence Intervals for the Ratio of Two Exponential Means with Applications to
More informationGeneral Bayesian Inference I
General Bayesian Inference I Outline: Basic concepts, One-parameter models, Noninformative priors. Reading: Chapters 10 and 11 in Kay-I. (Occasional) Simplified Notation. When there is no potential for
More informationMTH U481 : SPRING 2009: PRACTICE PROBLEMS FOR FINAL
MTH U481 : SPRING 2009: PRACTICE PROBLEMS FOR FINAL 1). Two urns are provided as follows: urn 1 contains 2 white chips and 4 red chips, while urn 2 contains 5 white chips and 3 red chips. One chip is chosen
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationOn the Performance of some Poisson Ridge Regression Estimators
Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-28-2018 On the Performance of some Poisson Ridge Regression Estimators Cynthia
More information1 Introduction. P (n = 1 red ball drawn) =
Introduction Exercises and outline solutions. Y has a pack of 4 cards (Ace and Queen of clubs, Ace and Queen of Hearts) from which he deals a random of selection 2 to player X. What is the probability
More informationLecture 2: Conjugate priors
(Spring ʼ) Lecture : Conjugate priors Julia Hockenmaier juliahmr@illinois.edu Siebel Center http://www.cs.uiuc.edu/class/sp/cs98jhm The binomial distribution If p is the probability of heads, the probability
More informationMidterm Examination. Mth 136 = Sta 114. Wednesday, 2000 March 8, 2:20 3:35 pm
Midterm Examination Mth 136 = Sta 114 Wednesday, 2000 March 8, 2:20 3:35 pm This is a closed-book examination so please do not refer to your notes, the text, or to any other books. You may use a two-sided
More informationBayesian Inference: Probit and Linear Probability Models
Utah State University DigitalCommons@USU All Graduate Plan B and other Reports Graduate Studies 5-1-2014 Bayesian Inference: Probit and Linear Probability Models Nate Rex Reasch Utah State University Follow
More informationInverse Sampling for McNemar s Test
International Journal of Statistics and Probability; Vol. 6, No. 1; January 27 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education Inverse Sampling for McNemar s Test
More informationPubH 5450 Biostatistics I Prof. Carlin. Lecture 13
PubH 5450 Biostatistics I Prof. Carlin Lecture 13 Outline Outline Sample Size Counts, Rates and Proportions Part I Sample Size Type I Error and Power Type I error rate: probability of rejecting the null
More informationOutline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution
Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model
More informationCS 361: Probability & Statistics
March 14, 2018 CS 361: Probability & Statistics Inference The prior From Bayes rule, we know that we can express our function of interest as Likelihood Prior Posterior The right hand side contains the
More informationCompositions, Bijections, and Enumerations
Georgia Southern University Digital Commons@Georgia Southern Electronic Theses & Dissertations COGS- Jack N. Averitt College of Graduate Studies Fall 2012 Compositions, Bijections, and Enumerations Charles
More informationDownloaded from:
Camacho, A; Kucharski, AJ; Funk, S; Breman, J; Piot, P; Edmunds, WJ (2014) Potential for large outbreaks of Ebola virus disease. Epidemics, 9. pp. 70-8. ISSN 1755-4365 DOI: https://doi.org/10.1016/j.epidem.2014.09.003
More informationClassical and Bayesian inference
Classical and Bayesian inference AMS 132 January 18, 2018 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 1 / 9 Sampling from a Bernoulli Distribution Theorem (Beta-Bernoulli
More informationExamining the accuracy of the normal approximation to the poisson random variable
Eastern Michigan University DigitalCommons@EMU Master's Theses and Doctoral Dissertations Master's Theses, and Doctoral Dissertations, and Graduate Capstone Projects 2009 Examining the accuracy of the
More informationEstimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio
Estimation of reliability parameters from Experimental data (Parte 2) This lecture Life test (t 1,t 2,...,t n ) Estimate θ of f T t θ For example: λ of f T (t)= λe - λt Classical approach (frequentist
More informationCS 361: Probability & Statistics
October 17, 2017 CS 361: Probability & Statistics Inference Maximum likelihood: drawbacks A couple of things might trip up max likelihood estimation: 1) Finding the maximum of some functions can be quite
More information2 Inference for Multinomial Distribution
Markov Chain Monte Carlo Methods Part III: Statistical Concepts By K.B.Athreya, Mohan Delampady and T.Krishnan 1 Introduction In parts I and II of this series it was shown how Markov chain Monte Carlo
More informationMAT 2379, Introduction to Biostatistics, Sample Calculator Questions 1. MAT 2379, Introduction to Biostatistics
MAT 2379, Introduction to Biostatistics, Sample Calculator Questions 1 MAT 2379, Introduction to Biostatistics Sample Calculator Problems for the Final Exam Note: The exam will also contain some problems
More informationBayesian RL Seminar. Chris Mansley September 9, 2008
Bayesian RL Seminar Chris Mansley September 9, 2008 Bayes Basic Probability One of the basic principles of probability theory, the chain rule, will allow us to derive most of the background material in
More informationBayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units
Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units Sahar Z Zangeneh Robert W. Keener Roderick J.A. Little Abstract In Probability proportional
More informationEstimation of Quantiles
9 Estimation of Quantiles The notion of quantiles was introduced in Section 3.2: recall that a quantile x α for an r.v. X is a constant such that P(X x α )=1 α. (9.1) In this chapter we examine quantiles
More informationBayesian Inference. STA 121: Regression Analysis Artin Armagan
Bayesian Inference STA 121: Regression Analysis Artin Armagan Bayes Rule...s! Reverend Thomas Bayes Posterior Prior p(θ y) = p(y θ)p(θ)/p(y) Likelihood - Sampling Distribution Normalizing Constant: p(y
More informationBayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification
Bayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification Michael Anderson, PhD Hélène Carabin, DVM, PhD Department of Biostatistics and Epidemiology The University of Oklahoma
More informationIntroduction to Machine Learning. Lecture 2
Introduction to Machine Learning Lecturer: Eran Halperin Lecture 2 Fall Semester Scribe: Yishay Mansour Some of the material was not presented in class (and is marked with a side line) and is given for
More informationIntroduction to Bayesian Statistics 1
Introduction to Bayesian Statistics 1 STA 442/2101 Fall 2018 1 This slide show is an open-source document. See last slide for copyright information. 1 / 42 Thomas Bayes (1701-1761) Image from the Wikipedia
More informationBernoulli and Poisson models
Bernoulli and Poisson models Bernoulli/binomial models Return to iid Y 1,...,Y n Bin(1, ). The sampling model/likelihood is p(y 1,...,y n ) = P y i (1 ) n P y i When combined with a prior p( ), Bayes rule
More informationThe extreme points of symmetric norms on R^2
Graduate Theses and Dissertations Iowa State University Capstones, Theses and Dissertations 2008 The extreme points of symmetric norms on R^2 Anchalee Khemphet Iowa State University Follow this and additional
More informationMultivariate Survival Analysis
Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in
More information10. Exchangeability and hierarchical models Objective. Recommended reading
10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.
More informationBayesian Estimation An Informal Introduction
Mary Parker, Bayesian Estimation An Informal Introduction page 1 of 8 Bayesian Estimation An Informal Introduction Example: I take a coin out of my pocket and I want to estimate the probability of heads
More informationBayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida
Bayesian Statistical Methods Jeff Gill Department of Political Science, University of Florida 234 Anderson Hall, PO Box 117325, Gainesville, FL 32611-7325 Voice: 352-392-0262x272, Fax: 352-392-8127, Email:
More informationJoint, Conditional, & Marginal Probabilities
Joint, Conditional, & Marginal Probabilities The three axioms for probability don t discuss how to create probabilities for combined events such as P [A B] or for the likelihood of an event A given that
More informationBayesian Statistics. Debdeep Pati Florida State University. February 11, 2016
Bayesian Statistics Debdeep Pati Florida State University February 11, 2016 Historical Background Historical Background Historical Background Brief History of Bayesian Statistics 1764-1838: called probability
More informationMidterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm
Midterm Examination STA 215: Statistical Inference Due Wednesday, 2006 Mar 8, 1:15 pm This is an open-book take-home examination. You may work on it during any consecutive 24-hour period you like; please
More informationNIH Public Access Author Manuscript Commun Stat Theory Methods. Author manuscript; available in PMC 2011 January 7.
NIH Public Access Author Manuscript Published in final edited form as: Commun Stat Theory Methods. 2011 January 1; 40(3): 436 448. doi:10.1080/03610920903391303. Optimal Configuration of a Square Array
More informationECE521 W17 Tutorial 6. Min Bai and Yuhuai (Tony) Wu
ECE521 W17 Tutorial 6 Min Bai and Yuhuai (Tony) Wu Agenda knn and PCA Bayesian Inference k-means Technique for clustering Unsupervised pattern and grouping discovery Class prediction Outlier detection
More informationLECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)
LECTURE 5 NOTES 1. Bayesian point estimators. In the conventional (frequentist) approach to statistical inference, the parameter θ Θ is considered a fixed quantity. In the Bayesian approach, it is considered
More informationJoint, Conditional, & Marginal Probabilities
Joint, Conditional, & Marginal Probabilities Statistics 110 Summer 2006 Copyright c 2006 by Mark E. Irwin Joint, Conditional, & Marginal Probabilities The three axioms for probability don t discuss how
More informationA CONDITION TO OBTAIN THE SAME DECISION IN THE HOMOGENEITY TEST- ING PROBLEM FROM THE FREQUENTIST AND BAYESIAN POINT OF VIEW
A CONDITION TO OBTAIN THE SAME DECISION IN THE HOMOGENEITY TEST- ING PROBLEM FROM THE FREQUENTIST AND BAYESIAN POINT OF VIEW Miguel A Gómez-Villegas and Beatriz González-Pérez Departamento de Estadística
More informationThe Bayesian Approach to Multi-equation Econometric Model Estimation
Journal of Statistical and Econometric Methods, vol.3, no.1, 2014, 85-96 ISSN: 2241-0384 (print), 2241-0376 (online) Scienpress Ltd, 2014 The Bayesian Approach to Multi-equation Econometric Model Estimation
More informationPairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion
Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Glenn Heller and Jing Qin Department of Epidemiology and Biostatistics Memorial
More informationDropping the Lowest Score: A Mathematical Analysis of a Common Grading Practice
Western Oregon University Digital Commons@WOU Honors Senior Theses/Projects Student Scholarship - 6-1-2013 Dropping the Lowest Score: A Mathematical Analysis of a Common Grading Practice Rosie Brown Western
More informationDiscrete Binary Distributions
Discrete Binary Distributions Carl Edward Rasmussen November th, 26 Carl Edward Rasmussen Discrete Binary Distributions November th, 26 / 5 Key concepts Bernoulli: probabilities over binary variables Binomial:
More informationWeakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0
More informationCS 340 Fall 2007: Homework 3
CS 34 Fall 27: Homework 3 1 Marginal likelihood for the Beta-Bernoulli model We showed that the marginal likelihood is the ratio of the normalizing constants: p(d) = B(α 1 + N 1, α + N ) B(α 1, α ) = Γ(α
More informationSTAT:5100 (22S:193) Statistical Inference I
STAT:5100 (22S:193) Statistical Inference I Week 10 Luke Tierney University of Iowa Fall 2015 Luke Tierney (U Iowa) STAT:5100 (22S:193) Statistical Inference I Fall 2015 1 Monday, October 26, 2015 Recap
More information