Inferences about Parameters of Trivariate Normal Distribution with Missing Data

Size: px
Start display at page:

Download "Inferences about Parameters of Trivariate Normal Distribution with Missing Data"

Transcription

1 Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School Inferences about Parameters of Trivariate Normal Distribution with Missing Data Xing Wang Florida International University, DOI:.548/etd.FI3899 Follow this and additional works at: Part of the Statistical Models Commons, and the Statistical Theory Commons Recommended Citation Wang, Xing, "Inferences about Parameters of Trivariate Normal Distribution with Missing Data" (3). FIU Electronic Theses and Dissertations This work is brought to you for free and open access by the University Graduate School at FIU Digital Commons. It has been accepted for inclusion in FIU Electronic Theses and Dissertations by an authorized administrator of FIU Digital Commons. For more information, please contact

2 FLORIDA INTERNATIONAL UNIVERSITY Miami, Florida INFERENCES ABOUT PARAMETERS OF TRIVARIATE NORMAL DISTRIBUTION WITH MISSING DATA A thesis submitted in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE in STATISTICS by Xing Wang 3

3 To: Dean Kenneth Furton College of Arts and Sciences This thesis, written by Xing Wang, and entitled Inferences about Parameters of Trivariate Normal Distribution with Missing Data, having been approved in respect to style and intellectual content, is referred to you for judgment. We have read this thesis and recommend that it be approved. Florence George Kai Huang, Co-Major Professor Jie Mi, Co-Major Professor Date of Defense: July 5, 3 The thesis of Xing Wang is approved. Dean Kenneth Furton College of Arts and Sciences Dean Lakshmi N. Reddi University Graduate School Florida International University, 3 ii

4 ACKNOWLEDGMENTS First of all, I would like to express my sincere thanks to my major professor, Dr. Jie Mi for his patient guidance, enthusiasm, encouragement and friendship throughout this whole study. I couldn t finish my research without his great support. I would like to express deepest appreciation to my co-major professor, Dr. Kai Huang, for his technical support and patient guidance in Latex and Matlab. I would also like to thank the member of my committee, Dr. Florence George for her time, valuable advice and great encouragement. In addition, I would like to thank the Department of Mathematics and Statistics, all the professors who supported and encouraged me throughout my life in FIU. I wish to thank auntie Miao, who gave me substantial help in my life. I also thank my classmate Maria, my technical support of Matlab. I am really grateful to have them with me, and may God bless them all. iii

5 ABSTRACT OF THE THESIS INFERENCES ABOUT PARAMETERS OF TRIVARIATE NORMAL DISTRIBUTION WITH MISSING DATA by Xing Wang Florida International University, 3 Miami, Florida Professor Jie Mi, Co-Major Professor Professor Kai Huang, Co-Major Professor Multivariate normal distribution is commonly encountered in any field, a frequent issue is the missing values in practice. The purpose of this research was to estimate the parameters in three-dimensional covariance permutation-symmetric normal distribution with complete data and all possible patterns of incomplete data. In this study, MLE with missing data were derived, and the properties of the MLE as well as the sampling distributions were obtained. A Monte Carlo simulation study was used to evaluate the performance of the considered estimators for both cases when was known and unknown. All results indicated that, compared to estimators in the case of omitting observations with missing data, the estimators derived in this article led to better performance. Furthermore, when was unknown, using the estimate of would lead to the same conclusion. Keywords: Trivariate Normal Distribution, Permutation-Symmetric Covariance, Missing Data, MLE. iv

6 TABLE OF CONTENTS CHAPTER PAGE I. Introduction II. Review of MLE with complete data III. The Maximum Likelihood Estimates with Missing Data Notation and Assumption Likelihood Function Derivation of ˆµ and ˆ IV. Properties of MLEs Mean of ˆµ Variance of ˆµ Variance of ˆµ ˆµ V. Sampling distributions The distribution of ˆµ µ The distribution of ˆµ ˆµ VI. Simulation Study with Known VII. Simulation Study with Unknown VIII. Conclusion REFERENCES v

7 LIST OF TABLES TABLE PAGE. Coverage Probability of confidence regions of µ The Probability of Type I error The Power of testing H : µ = vs.h a : µ The Coverage probability of confidence intervals of µ The average width of confidence intervals of µ The average variance of ˆµ The coverage probability of confidence intervals of µ µ The average width of confidence intervals of µ µ The average variance of ˆµ ˆµ Coverage Probability of confidence regions of µ The Probability of Type I error The Power of testing H : µ = vs.h a : µ The Coverage probability of confidence intervals of µ The average width of confidence intervals of µ The average variance of ˆµ The coverage probability of confidence intervals of µ µ The average width of confidence intervals of µ µ The average variance of ˆµ ˆµ vi

8 LIST OF FIGURES FIGURE PAGE. Comparison of Coverage Probabilities of Confidence Region of µ Absolute Value of the Difference Between the Coverage Probabilities and Comparison of Type I Error Absolute Value of the Difference Between the Type I Error and α = Compare the Power of Testing Compare the Coverage Probabilities of the confidence intervals of µ Absolute Value of the Difference Between the Coverage Probabilities and Compare the Average Widths of the Confidence Intervals of µ Compare the average variance of ˆµ Compare the Coverage Probabilities of the confidence intervals of µ µ Absolute value of the difference between Coverage Probabilities and Compare Average Widths of Confidence Intervals of µ µ Compare the average variance of ˆµ ˆµ Compare the Coverage Probabilities of the Confidence Regions of µ Absolute value of the difference between Coverage Probabilities and Compare the Type I Error Absolute value of difference between Type I Error and α = Compare the Power of Testing Compare the Coverage Probabilities of the Confidence Intervals of µ = Absolute value of the difference between Coverage Probabilities and Compare the Average Widths of the Confidence Intervals of µ = vii

9 . Compare the average variance of ˆµ Compare the Coverage Probabilities of the Confidence Intervals of µ µ Absolute value of the difference between Coverage Probabilities and Compare the Average Widths of the Confidence Intervals of µ µ Compare the average variance of ˆµ ˆµ viii

10 I. Introduction In practice, the normal distribution is commonly encountered, since the sampling distribution of many multivariate statistics are proximately normal because of the a central limit effect. Early application of the multivariate normal distribution was with regard to biological studies. For instance, we often think that height and weight, when observed on the same individual, approximately follow bivariate normal distribution. We could extend this to the foot size, or any other variable of related physical characteristics, and these measurements together could follow the multivariate normal distribution. Moreover, we often have to analyze data that contains missing values in practice. Incomplete normal data could arise in any number of scientific investigations: early detection of diseases, wildlife survey research, mental health research, and so on. Missing data are one of the most persuasive problems for analysis of data which can occur for a variety of reasons. For example, a participant may refuse to answer some of the questions; the use of new instruments results in incomplete historical data; some information may be purposely excised to protect confidentiality. Estimation of parameters of a multivariate normal distribution when data are incomplete has been discussed by many authors. A systematic approach to missing values problem was derived using likelihoods of observed values. Wilks (93) considered MLEs for a bivariate normal population with missing data in both variables. Srivastava and Zaatar (973) used Monte Carlo simulation to compare four estimators of the covariance matrix of a bivariate normal distribution. Edgett (956) gave maximum likelihood estimates of parameters of a trivariate normal distribution when observations on only one variable are missing. Lord (955) and Matthai (95) also found estimates of parameters

11 of a trivariate normal distribution in other special cases. Anderson (957) indicated how one could obtain the Maximum Likelihood Estimates when the sample was monotone. Hocking and Smith (968) developed a method of estimating parameters of a p-variate normal distribution with zero mean vector in which the missing observations are not required to follow certain patterns. Their estimation technique could be summarized as follows: The data were divided into groups according to which variates were missing. Initial estimates of the parameters were obtained from that group of observations with no missing variates. (c) These initial estimators were modified by adjoining, optimally, the information in the remaining groups in a sequential manner until all data was used. Hocking and Marx (979) used the same method as Hocking and Smith (968) to derive estimates, but their use of matrices simplified the notation and gave the estimates in a form that was easily implemented on a computer. They also gave exact small sample moments of the estimators for the case of two data groups. The purpose of my research is to estimate the parameters in three-dimensional covariance permutation-symmetric normal distribution with complete data and all possible patterns of incomplete data, and then to study the properties of the estimators. It is assumed that all correlation coefficients and variances are equal, which means that we focus on the covariance permutation-symmetric trivariate normal distribution. The special case of the covariance permutation-symmetric model are the exchangeable normal variables, which make their appearance in many statistical applications. For instance, in the Bayes theory concerning normal observations, it is generally assumed that the prior distribution

12 of θ, the population mean, has an N(µ, ω ) distribution for some µ and ω >. Thus, for given θ, if the random variables are (conditionally) i.i.d normal variables, then the marginal distribution of the data variable X = (X,..., X n ) is a mixture. In this case X,..., X n are exchangeable normal variables. The thesis is organized as follows. The Maximum Likelihood Estimates of the parameters of trivariate normal distribution with complete data are reviewed in Section. Maximum Likelihood Estimates with missing data will be derived in Section 3. In Section 4, the properties of the MLEs will be obtained, followed by the sampling distributions in Section 5. The numerical studies based on Monte Carlo simulations are considered in Section 6, which include comparison of confidence regions of µ; the probability of Type I error, and power of testing H : µ = vs H a : µ ; (c) coverage probability and average width of confidence intervals with regard to µ ; (d) coverage probability and average width of confidence intervals with regard to µ µ using i) only three-dimensional observations and ii) both three-dimensional observations and all possible incomplete observations. Finally, in Section 7, the simulation will follow the same procedure in Section 6 but based on unknown. 3

13 II. Review of MLE with complete data In order to obtain the trivariate normal likelihood function, let us assume that the 3 vectors X, X,..., X n represent a random sample from a three-dimensional multivariate normal population with mean vector µ and covariance matrix Σ, where X i = (x i, x i, x i3 ), i =,,..., n, µ = (µ, µ, µ 3 ) and Σ = Since X, X,..., X n are mutually independent and each has a distribution N 3 (µ,σ), the joint density function of all the observations is the product of the marginal normal densities: f(x, X,..., X n ) = = n { (π) 3/ Σ /exp[ (X i µ) Σ (X i µ) ] } n (π) 3n/ Σ n/exp[ (X i µ) Σ (X i µ) ] It has been derived that ˆµ = X () and ˆΣ = n (X i X)(X i X) n = (n )S n () are the maximum likelihood estimates of µ and Σ respectively, where X = n n X i, i =,,..., n 4

14 and S = n (X i X)(X i X). n 5

15 III. The Maximum Likelihood Estimates with Missing Data. Notation and Assumption In the present paper, we consider the covariance permutation-symmetric trivariate normal distribution, which means that = = 3 and all correlation coefficients are equal: = 3 = 3. By denoting τ =, the marginal density function can be expressed as: where f(x, x, x 3 ) = e w/[τ( )] π 3/ τ 3/ (3) w = ( + ) [ (x µ ) + (x µ ) + (x 3 µ 3 ) ] + [ (x µ )(x µ ) + (x µ )(x 3 µ 3 ) + (x µ )(x 3 µ 3 ) ] The data in our study are divided into groups according to which variables are missing, and initial estimates of the parameters are obtained from that group of observations with no missing variables. Specifically, it is assumed that a sample of size n + n + n 3 + n 3 + n + n + n 3 is taken from a three-dimensional normal distribution but some of the observations have randomly occurring missing entries. In this case, the observations can be placed in 7 groups. Specifically, it means that the data consist of n complete observations {(x i, x i, x i3 ), i =,, n } on X = (X, X, X 3 ), n observations {(x () i, x() i ), i =,, n } on (X, X ), n 3 observations {(x (3) i, x(3) i3 ), i =,, n 3 } on (X, X 3 ), n 3 observations {(x (3) i, x(3) i3 ), i =,, n 3 } on (X, X 3 ), n observations {x () i, i =,, n } on X, n observations {x () i, i =,, n } on X, n 3 observations {x (3) i3, i =,, n 3 } on X 3. 6

16 . Likelihood Function Following Hocking and Smith(968), the likelihood function can be written as: L(µ, τ) = L (µ, τ) L (µ, µ, τ) L 3 (µ, µ 3, τ) L 3 (µ, µ 3, τ) L (µ, τ) = L (µ, τ) L 3 (µ 3, τ) n n n 3 () (3) f(x i; µ, τ) i ; µ, µ, τ) i ; µ, µ 3, τ) f(x f(x n 3 n n n 3 i ; µ, µ 3, τ) f(x () i ; µ, τ) f(x () i ; µ, τ) (3) f(x f(x (3) i ; µ 3, τ) (4) Where x i = (x i, x i, x i3 ) (), i =,, n ; i = (x x () i, x() i ), i =,, n ; (3) xi = (x (3) i, x (3) i3 ) (3), i =,, n 3 ; i = (x x (3) i, x (3) i3 ), i =,, n 3 ; x () i = x () i, i =,, n ; x () i = x () i, i =,, n ; x (3) i = x (3) i3, i =,, n 3 L is used to denote the likelihood function for the three-dimensional complete observations, where L (µ, τ) = e P n w i/[τ( )] ( π 3/ τ 3/ ) n with w i = ( + ) [ (x i µ ) + (x i µ ) + (x i3 µ 3 ) ] + [ (x i µ )(x i µ ) + (x i µ )(x i3 µ 3 ) + (x i µ )(x i3 µ 3 ) ] Similarly, other notations are defined as follows: P n L (µ, µ, τ) = e [ (x() i µ ) P n (x() i µ )(x () i µ )+ P n (x() i µ ) ]/[τ( )] (π) n τ n ( ) n / 7

17 P n3 L 3 (µ, µ 3, τ) = e [ (x(3) i µ ) P n 3 (x(3) i µ )(x (3) i3 µ 3)+ P n 3 (x(3) i3 µ 3) ]/[τ( )] (π) n 3 τ n 3( ) n 3/ P n3 L 3 (µ, µ 3, τ) = e [ (x(3) i µ ) P n 3 (x(3) i µ )(x (3) i3 µ 3)+ P n 3 (x(3) i3 µ 3) ]/[τ( )] (π) n 3 τ n 3( ) n 3/ P n L (µ, τ) = e (x() i µ ) /(τ) (π) n / τ n / P n L (µ, τ) = e (x() i µ ) /(τ) (π) n / τ n / 3. Derivation of ˆµ and ˆ P n3 L 3 (µ 3, τ) = e (x(3) i3 µ 3) /(τ) (π) n 3/ τ n 3/ In the rest of this study it is assumed that ( /, ) is known. In order to apply likelihood-based method to obtain the MLEs ˆµ and ˆ of µ and, the natural logarithms of the above likelihood functions are taken: lnl (µ, τ) = 3n ln(π) 3n lnτ n ln( ) n + + τ( ) [ n (x i µ ) + (x i µ ) + n (x i3 µ 3 ) ] n τ( ) [ n (x i µ )(x i µ ) + (x i µ )(x i3 µ 3 ) n + (x i µ )(x i3 µ 3 )] lnl (µ, µ, τ) = n ln(π) n lnτ n ln( ) 8

18 n τ( ) [ n + (x () i µ ) ] (x () i n µ ) (x () i µ )(x () i µ ) lnl 3 (µ, µ 3, τ) = n 3 ln(π) n 3 lnτ n 3 ln( ) n3 τ( ) [ n 3 (x (3) i µ ) (x (3) i µ )(x (3) i3 µ 3 ) n 3 + (x (3) i3 µ 3 ) ] lnl 3 (µ, µ 3, τ) = n 3 ln(π) n 3 lnτ n 3 ln( ) n3 τ( ) [ n 3 (x (3) i µ ) (x (3) i µ )(x (3) i3 µ 3 ) n 3 + (x (3) i3 µ 3 ) ] lnl (µ, τ) = n ln(π) n lnτ τ lnl (µ, τ) = n ln(π) n lnτ τ lnl 3 (µ 3, τ) = n 3 ln(π) n 3 lnτ τ n n n 3 (x () i µ ) (x () i µ ) (x (3) i3 µ 3 ) Combining all the above expressions together yields the log-likelihood function (5): lnl(µ, τ) = 3n ln(π) 3n lnτ n ln( ) n ln(π) n lnτ n ln( ) n 3 ln(π) n 3 lnτ n 3 ln( ) n 3 ln(π) n 3 lnτ n 3 ln( ) n ln(π) n lnτ n ln(π) n lnτ n 3 ln(π) n 3 lnτ 9

19 n + + τ( ) [ n n (x i µ ) + (x i µ ) + (x i3 µ 3 ) ] n τ( ) [ n (x i µ )(x i µ ) + (x i µ )(x i3 µ 3 ) n + (x i µ )(x i3 µ 3 )] n (x () i µ )(x () n3 τ( ) [ n 3 + n 3 (x (3) i3 (x (3) i µ 3 ) ] (x (3) i µ )(x (3) n (x () n τ( ) [ n i µ ) + n 3 µ ) (x () (x () i µ ) i µ ) ] (x (3) i µ )(x (3) n3 τ( ) [ (x (3) i µ ) i3 µ 3 ) + n (x () n 3 (x (3) i3 µ 3 ) ] i3 µ 3 ) i µ ) i µ ) n 3 (x (3) i3 µ 3 ) τ τ τ = c ( 3n + n + n 3 + n 3 + n + n + n 3) lnτ n + + τ( ) [ n n (x i µ ) + (x i µ ) + (x i3 µ 3 ) ] n τ( ) [ n3 τ( ) [ n3 τ( ) [ τ n (x () n i µ ) + (x () (x (3) i µ ) + (x (3) i µ ) + i µ ) τ n (x () (x () n 3 (x (3) n 3 (x (3) i µ ) ] i3 µ 3 ) ] i3 µ 3 ) ] i µ ) τ n 3 (x (3) i3 µ 3) n τ( ) [ n (x i µ )(x i µ ) + (x i µ )(x i3 µ 3 ) n + (x i µ )(x i3 µ 3 )] + n 3 + (x (3) i µ )(x (3) n 3 i3 µ 3 ) + [ n τ( ) (x (3) i µ )(x (3) (x () i µ )(x () i3 µ 3 ) ] i µ )

20 (5) where c = 3n ln(π) n ln( ) n ln(π) n ln( ) n 3 ln(π) n 3 ln( ) n 3 ln(π) n 3 ln( ) n ln(π) n ln(π) n 3 ln(π) In order to find the Maximum Likelihood Estimates of µ = (µ, µ, µ 3 ) and τ, the partial derivatives of lnl with respect to µ and τ are taken, and are set to be zero: lnl µ = lnl τ = which are equivalent to the following expressions: n lnl + n = (x µ τ( i µ ) + ) τ( ) [ n 3 + n (x (3) i µ )] + τ + (x i3 µ 3 )] n (x () n τ( ) [ i µ ) + (x () (x () i µ ) n τ( ) [ (x i µ ) n 3 i µ ) + (x (3) i3 µ 3 )] = (6) n lnl + n = (x µ τ( i µ ) + ) τ( ) [ n 3 + n (x (3) i µ )] + τ + (x i3 µ 3 )] n (x () n τ( ) [ i µ ) + (x () (x () i µ ) n τ( ) [ (x i µ ) n 3 i µ ) + (x (3) i3 µ 3 )] = (7)

21 n lnl + n3 = (x µ 3 τ( i3 µ 3 ) + ) τ( ) [ n 3 + n (x (3) i3 µ 3 )] + τ + (x i µ )] n 3 (x (3) n3 τ( ) [ i3 µ 3) + (x (3) (x (3) i3 µ 3 ) n τ( ) [ (x i µ ) n 3 i µ ) + (x (3) i µ )] = (8) lnl τ = (3n τ + n + n 3 + n 3 + n + n + n 3) n + τ ( ) [ n (x i µ ) + (x i µ ) + n (x i3 µ 3 ) ] n + τ ( ) [ n (x i µ )(x i µ ) + (x i µ )(x i3 µ 3 ) n + (x i µ )(x i3 µ 3 )] + n n 3 n 3 + n τ [ (x () i µ )(x () (x (3) i µ )(x (3) (x (3) i µ )(x (3) (x () n τ ( ) [ n i µ ) + i3 µ 3 ) + i3 µ 3 ) + n i µ ) + (x () (x () n 3 (x (3) n 3 (x (3) (x () i µ ) n 3 i µ ) + (x (3) i µ ) n 3 i3 µ 3 ) + i3 µ 3 ) ] n 3 i µ ) + (x (3) i3 µ 3 ) ] = (x (3) i µ ) (9) The system of linear equations (6) (7) (8) can be rewritten as: n n ( + )( ) (x i µ ) + ( )[ n + ( )( ) n ( )[ (x () (x () n n 3 i µ ) + (x () n (x (3) i µ )] i µ ) + ( )[ (x i µ ) + (x i3 µ 3 )] n 3 i µ ) + (x (3) i3 µ 3 )] = ()

22 n n ( + )( ) (x i µ ) + ( )[ n + ( )( ) n ( )[ (x () (x () n n 3 i µ ) + (x () n (x (3) i µ )] i µ ) + ( )[ (x i µ ) + (x i3 µ 3 )] n 3 i µ ) + (x (3) i3 µ 3 )] = () n n 3 ( + )( ) (x i3 µ 3 ) + ( )[ n 3 + ( )( ) n 3 ( )[ (x (3) (x (3) n n 3 i3 µ 3 ) + (x (3) n (x (3) i3 µ 3 )] i3 µ 3) + ( )[ (x i µ ) + (x i µ )] n 3 i µ ) + (x (3) They could be further expressed as follows: i µ )] = () [n ( + ) + (n + n 3 )( + ) + n ( + )( + )( )]µ [n ( + ) + n ( + )]µ [n 3 ( + ) + n ( + )]µ 3 n ( + ) n x i ( + )( x () i + n 3 x (3) i ) n n ( + )( + )( ) x () i + ( + ) (x i + x i3 ) n + ( + )( x () i + n 3 x (3) i3 ) = (3) [n ( + ) + n ( + )]µ + [n ( + ) + (n + n 3 )( + ) + n ( + )( + )( )]µ [n 3 ( + ) + n ( + )]µ 3 n n ( + ) x i ( + )( x () i + 3 n 3 x (3) i )

23 n n ( + )( + )( ) x () i + ( + ) (x i + x i3 ) n + ( + )( x () i + n 3 x (3) i3 ) = (4) [n 3 ( + ) + n ( + )]µ [n 3 ( + ) + n ( + )]µ + [n ( + ) + (n 3 + n 3 )( + ) + n 3 ( + )( + )( )]µ 3 n ( + ) n 3 x i3 ( + )( x (3) i3 + n 3 x (3) i3 ) n 3 n ( + )( + )( ) x (3) i3 + ( + ) (x i + x i ) n 3 + ( + )( x (3) i + n 3 x (3) i ) = (5) For the sake of convenience, the above system of equations (3) (4) (5) could be denoted by matrix and vectors: Aµ = b (6) where Matrix A= a (n γ 3 + n γ ) (n 3 γ 3 + n γ ) (n γ 3 + n γ ) a (n 3 γ 3 + n γ ) (n 3 γ 3 + n γ ) (n 3 γ 3 + n γ ) a 33 with (7) a = n γ + (n + n 3 )γ 3 + n γ 3 γ γ, a = n γ + (n + n 3 )γ 3 + n γ 3 γ γ, a 33 = n γ + (n 3 + n 3 )γ 3 + n 3 γ 3 γ γ, γ = +, γ =, γ 3 = + 4

24 µ = (µ, µ, µ 3 ), b = (b, b, b 3 ) (8) b = γ n x. + γ 3 (n x (). + n 3 x (3). ) + γ 3 γ γ n x (). γ n ( x. + x.3 ) γ 3 (n x (). + n 3 x (3).3 ) b = γ n x. + γ 3 (n x (). + n 3 x (3). ) + γ 3 γ γ n x (). γ n ( x. + x.3 ) γ 3 (n x (). + n 3 x (3).3 ) where x.j = b 3 = γ n x.3 + γ 3 (n 3 x (3).3 + n 3 x (3).3 ) + γ 3 γ γ n 3 x (3).3 γ n ( x. + x. ) γ 3 (n 3 x (3). + n 3 x (3). ) P n x ij n, x ().j = x () P n.j = x() ij n, x ().j = P n x() ij n, x (3).j = P n x() ij n, x (3) P n3.j = x(3) ij n 3, j =,, 3. P n3 x(3) ij n 3, x (3) P n3.j = x(3) ij n 3, If A is a positive definite matrix, then the equation (6) can be solved and thus allow us to obtain the matrix expression of Maximum Likelihood Estimates of µ. To prove the positive definiteness of A, matrix A is firstly denoted as the sum of two matrices: B and B. The next step is to show that B is positive semi-definite and B is positive definite, where n + n 3 n n 3 B = γ 3 n n + n 3 n 3 n 3 n 3 n 3 + n 3 n γ + n γ γ 3 n n B = γ n n γ + n γ γ 3 n n n n γ + n 3 γ γ 3 (9). () 5

25 It can be proved that matrix B is positive semi-definite by rewriting B as the sum of C, C and C 3, where n n C = γ 3 n n () C = γ 3 C 3 = γ 3 n 3 n 3 () n 3 n 3 n 3 n 3. (3) n 3 n 3 Proof. A matrix is positive semi-definite if and only if every principal minor of the matrix is nonnegative. Since n, (, ), γ 3 = + >, the first-order principal minors of matrix C are nonnegative; the second-order principal minors of matrix C are γ 3 n ( ) and ; the third-order principal minor of matrix C is. Therefore, C is a positive semi-definite matrix. Similarly, since n 3 and n 3, all the principal minors of C and C 3 are nonnegative, so that C and C 3 are also positive semi-definite matrices. Based on the results above, matrix B = C + C + C 3 is a positive semi-definite matrix. Applying the similar argument to matrix B, we could write it as the sum of C 4 and C 5 given below. By showing that C 4 is positive definite and C 5 is positive semi-definite, B can be proved as a positive definite matrix. 6

26 n γ n n C 4 = γ n n γ n n n n γ n γ γ 3 C 5 = γ n γ γ 3 n 3 γ γ 3 (4) (5) Proof. Since n >, (, ), γ = + >, γ = >, γ 3 = + >, the first-order principal minor of matrix C 4 is n γ > ; the second-order principal minor is n γ (γ ) > ; the third-order principal minor is n 3 ( + ) 3 ( )( + ) >. Therefore, C 4 is a positive definite matrix. Since n, n, n 3, the first-order principal minors of matrix C 5 are n γ γ γ 3, n γ γ γ 3 and n 3 γ γ γ 3. Similarly, it can be checked that all the second and third-order principal minors are nonnegative. Therefore C 5 is a positive semi-definite matrix. As a result, matrix B = C 4 + C 5 is a positive definite matrix. B is shown to be positive semi-definite and B is positive definite, therefore A = B + B is a positive definite matrix. Note that A is positive definite, thus it is a nonsingular matrix and its inverse matrix exists. It could be derived from expression (6) that ˆµ = A b where ˆµ = (ˆµ, ˆµ, ˆµ 3 ). (6) Moreover, the Maximum Likelihood Estimator of τ could be obtained by plugging ˆµ into equation (9): 7

27 ˆτ = 3n + n + n 3 + n 3 + n + n + n { n + 3 ( + )( ) [ (x i ˆµ ) n n n + (x i ˆµ ) + (x i3 ˆµ 3 ) ] ( + )( ) [ (x i ˆµ )(x i ˆµ ) n n + (x i ˆµ )(x i3 ˆµ 3 ) + (x i ˆµ )(x i3 ˆµ 3 )] n + ( + )( ) [ n + + (x () n 3 (x (3) n 3 + n 3 (x () i i ˆµ ) + n 3 i3 ˆµ 3 ) + (x (3) i3 (x (3) i (x (3) i ˆµ 3 ) ] + n [ n ˆµ ) (x () n 3 ˆµ ) (x () i ˆµ )(x () n 3 ˆµ ) i ˆµ ) + i ˆµ ) (x (3) i ˆµ )(x (3) (x (3) i ˆµ )(x (3) n (x () i ˆµ n 3 ) + i3 ˆµ 3 ) i3 ˆµ 3 ) (x (3) i3 ˆµ 3) ] } (7) By denoting c = 3n + n + n 3 + n 3 + n + n + n 3, γ = +, γ =, γ 3 = +, the equation above can be further simplified as: ˆτ = { γ n n n [ (x i ˆµ ) + (x i ˆµ ) + (x i3 ˆµ 3 ) ] c γ 3 γ n [ (x i ˆµ )(x i ˆµ ) + γ 3 γ n + n n (x i ˆµ )(x i3 ˆµ 3 )] + γ γ [ n 3 n 3 + n [ (x () i ˆµ )(x () (x (3) i ˆµ )(x (3) (x (3) i ˆµ )(x (3) i ˆµ n ) + (x () n (x i ˆµ )(x i3 ˆµ 3 ) i ˆµ ) + i3 ˆµ 3 ) + i3 ˆµ 3 ) + n (x () n 3 (x (3) n 3 (x (3) i ˆµ n 3 ) + (x () 8 (x () i ˆµ ) n 3 i ˆµ ) + n 3 i3 ˆµ 3 ) + i3 ˆµ 3 ) ] (x (3) i3 ˆµ 3) ] } (x (3) i ˆµ ) (x (3) i ˆµ )

28 (8) IV. Properties of MLEs. Mean of ˆµ The estimate of µ in our model with missing data is an unbiased estimator. This property could be easily proved as follows. Proof. Since E( x.j ) = E( x ().j ) = E( x (3).j ) = E( x (3).j ) = E( x ().j ) = E( x().j ) = E( x(3).j ) = µ j, j =,, 3. E(b ) E = E(b ) E(b 3 ) = γ n µ + γ 3 (n µ + n 3 µ ) + γ 3 γ γ n µ γ n (µ + µ 3 ) γ 3 (n µ + n 3 µ 3 ) γ n µ + γ 3 (n µ + n 3 µ ) + γ 3 γ γ n µ γ n (µ + µ 3 ) γ 3 (n µ + n 3 µ 3 ) γn µ 3 + γ 3 (n 3 µ 3 + n 3 µ 3 ) + γ 3 γ γ n 3 µ 3 γ n (µ + µ ) γ 3 (n 3 µ + n 3 µ ) = [n γ + (n + n 3 )γ 3 + n γ 3 γ γ ]µ (n γ 3 + n γ )µ (n 3 γ 3 + n γ )µ 3 (n γ 3 + n γ )µ + [n γ + (n + n 3 )γ 3 + n γ 3 γ γ ]µ (n 3 γ 3 + n γ )µ 3 (n 3 γ 3 + n γ )µ (n 3 γ 3 + n γ )µ + [n γ + (n 3 + n 3 )γ 3 + n 3 γ 3 γ γ ]µ 3 = Aµ Consequently E(ˆµ) = E(A b) = A E = A Aµ = µ 9

29 . Variance of ˆµ With matrix notation, the variance of ˆµ can be easily denoted as V ar(ˆµ) = V ar(a b) = A V ar(a ) (9) where b is the same vector as in (8). For the sake of convenience, b could be denoted as b = DX where and D = n γ n γ n γ n γ n γ n γ n γ n γ n γ n γ 3 n γ 3 n γ 3 n γ 3 n 3 γ 3 n 3 γ 3 n 3 γ 3 n 3 γ 3 n 3 γ 3 n 3 γ 3 n 3 γ 3 n 3 γ 3 n γ 3 γ γ n γ 3 γ γ n 3 γ 3 γ γ X = ( x., x., x.3, x ()., x ()., x (3)., x (3).3, x (3)., x (3).3, x ()., x ()., x (3).3 ). Since the V ar could be further expressed as: V ar = V ar(dx ) = DV ar(x )D, (3) combining (9) and (3), we get: V ar(ˆµ) = (A D)V ar(x )(A D) (3)

30 where the variance of X can be obtained by V ar(x ) = n n n n n n n n n n n n n n 3 n 3 n 3 n 3 n 3 n 3 n 3 n 3 n n n Variance of ˆµ ˆµ The variance of ˆµ is a 3 3 matrix which has the format as follows: V ar(ˆµ ) Cov(ˆµ, ˆµ ) Cov(ˆµ, ˆµ 3 ) V ar(ˆµ) = Cov(ˆµ, ˆµ ) V ar(ˆµ ) Cov(ˆµ, ˆµ 3 ). (3) Cov(ˆµ, ˆµ 3 ) Cov(ˆµ, ˆµ 3 ) V ar(ˆµ 3 ) Consequently, the variance of ˆµ ˆµ can be obtained as: V ar(ˆµ ˆµ ) = V ar(ˆµ ) + V ar(ˆµ ) Cov(ˆµ, ˆµ ) V. Sampling distributions. The distribution of ˆµ µ Since vector X = (X, X, X 3 ) has a joint normal distribution, ˆµ is a linear combination of b, which is also the linear combination of the components of X. As a

31 result, ˆµ follows the normal distribution: N 3 (E(ˆµ µ), V ar(ˆµ µ)), where E(ˆµ µ) and V ar(ˆµ µ) could be simply calculated as: E(ˆµ µ) = E(ˆµ) E(µ) = µ µ = V ar(ˆµ µ) = V ar(ˆµ) = (A D)V ar(x )(A D).. The distribution of ˆµ ˆµ Since ˆµ = (ˆµ, ˆµ, ˆµ 3 ) has a joint normal distribution, ˆµ ˆµ follows normal distribution with mean E(ˆµ ˆµ ) = E(ˆµ ) E(ˆµ ) = µ µ and variance V ar(ˆµ ˆµ ) = V ar(ˆµ ) + V ar(ˆµ ) Cov(ˆµ, ˆµ ). VI. Simulation Study with Known In this section, a simulation study is conducted to compare the performance of estimates in case only three-dimensional observations and both three-dimensional observations and all possible incomplete observations. The efficiency of this estimator is evaluated by means of coverage probabilities and average width of the confidence intervals, type I error and power of testing H : µ = vs.h a : µ. The sample size is chosen as n =, n = n 3 = n 3 = 5, n = n = n 3 =, and the following five levels of standard deviations are considered: =.5,.5,,, 4. It is assumed that is known in this section, so the real value of is used, which ranges from. to.9. The population mean µ is chosen to be µ = (,, 3) when calculating the power of testing, coverage probabilities and average width of the confidence intervals. The number of replication is r =.

32 . Compare the confidence regions of µ Table : Coverage Probabilities of 95% confidence regions of µ = (,, 3)

33 Coverage Probability Coverage Probability Coverage Probability = =.94 = Coverage Probability Coverage Probability =.5.94 = Confidence level=.95 Figure : Comparison of Coverage Probabilities of Confidence Region of µ Absolute value of the difference Absolute value of the difference Absolute value of the difference 4 x 3 = x 3 = 4 6 x 3 =4 4 Absolute value of the difference Absolute value of the difference 8 x 3 = x 3 = 3 Figure : Absolute Value of the Difference Between the Coverage Probabilities and.95 It is evident from Figure that, in both case and, the coverage probabilities of the 95% confidence regions are close to.95. The absolute values of difference between the coverage probabilities and confidence level.95, as shown in Figure, are also small in both cases. 4

34 . Compare the probability of type I error and power of testing H : µ = vs.h a : µ. Compare the probability of type I error Table : The Probability of Type I error

35 Probability of Type I Error Probability of Type I Error Probability of Type I Error = = =.4 Probability of Type I Error Probability of Type I Error =.5.4 = α=.5 Figure 3: Comparison of Type I Error Absolute value of the difference Absolute value of the difference Absolute value of the difference 8 x 3 = x 3 = 3 6 x 3 =4 4 Absolute value of the difference Absolute value of the difference 6 x 3 = x 3 = 4 Figure 4: Absolute Value of the Difference Between the Type I Error and α =.5 In the case of α =.5, type I errors in both and are close to.5 at each level of. It can be noted more clearly from Figure 4 that the absolute values of difference between Type I Errors and α =.5 are small in both cases. 6

36 . Compare the power of testing H : µ = vs.h a : µ Table 3: The Power of testing H : µ = (,, ) vs.h a : µ = (,, 3)

37 =6 =8 Power of testing Power of testing = =.8 Power of testing.5 Power of testing.6.4. =4.8 Power of testing.6.4. Figure 5: Compare the Power of Testing In order to compare the power of testing, our levels of standard deviation have been chosen as = 6, 8,,, 4 to make the results more observable. It can be observed from Figure 5 that, at each level of, the power of testing is much higher in case than that in. 3. Compare the confidence intervals of µ 3. Compare the coverage probabilities of confidence intervals of µ It can be noted from Figure 6 and Figure 7 that, in both case and, the coverage probabilities of the 95% confidence intervals of µ are close to.95. The absolute values of differences between the coverage probabilities and confidence level.95 are small in both cases. 8

38 Table 4: The Coverage probabilities of confidence intervals of µ =

39 Coverage Probability Coverage Probability Coverage Probability = =.94 = Coverage Probability Coverage Probability = = Confidence level=.95 Figure 6: Compare the Coverage Probabilities of the confidence intervals of µ Absolute value of the difference Absolute value of the difference Absolute value of the difference 4 x 3 = x 3 = x 3 =4 6 4 Absolute value of the difference Absolute value of the difference 4 x 3 = x 3 = 4 Figure 7: Absolute Value of the Difference Between the Coverage Probabilities and.95 3

40 3. Compare the average width of confidence intervals of µ Table 5: The average width of confidence intervals of µ =

41 . =.5.4 =.5 Average Width.5..5 Average Width.3.. =.8 =.5 Average Width.6.4. Average Width.5 =4.5 Average Width.5.5 Figure 8: Compare the Average Widths of the Confidence Intervals of µ As presented in Figure 8, at each level of, the average width of the confidence intervals is smaller in case than that in. 3.3 Compare the average variance of ˆµ As presented in Figure 9, at each level of, the average variance of ˆµ is smaller in case than that in. This result explains our finding in section 3.: the average width of the confidence intervals is smaller in case than that in. 3

42 Table 6: The average variance of ˆµ e e e e e e e

43 Average Variance Average Variance Average Variance 8 x 3 = = =4.5 Average Variance Average Variance =.5 = Figure 9: Compare the average variance of ˆµ 4. Compare the confidence intervals of µ µ 4. The coverage probabilities of confidence intervals of µ µ (µ =, µ = ) It can be observed from Figure and Figure that, in both case and, the coverage probabilities of the 95% confidence intervals of µ µ are close to.95. The absolute values of differences between the coverage probabilities and confidence level.95 are small in both cases. 34

44 Table 7: The coverage probability of confidence intervals of µ µ (µ =, µ = )

45 Coverage Probability Coverage Probability Coverage Probability = = =.94 Coverage Probability Coverage Probability =.5.94 = Confidence level=.95 Figure : Compare the Coverage Probabilities of the confidence intervals of µ µ Absolute value of the difference Absolute value of the difference Absolute value of the difference 6 x 3 = x 3 = 4 6 x 3 =4 4 Absolute value of the difference Absolute value of the difference 6 x 3 = x 3 = 4 Figure : Absolute value of the difference between Coverage Probabilities and.95 36

46 4. Average widths of confidence intervals of µ µ (µ =, µ = ) Table 8: Average widths of confidence intervals of µ µ (µ =, µ = )

47 .4 =.5.8 =.5 Average Width.3.. Average Width.6.4. = = Average Width.5 Average Width.5.5 =4 4 Average Width 3 Figure : Compare Average Widths of Confidence Intervals of µ µ Figure shows that the average width of confidence intervals of µ µ is decreasing as the is increasing. The average widths in case are always smaller than those in at each level of. 4.3 Compare the average variance of confidence intervals of µ µ As presented in Figure 3, at each level of, the average variance of ˆµ ˆµ is smaller in case than that in. This result explains our finding in section 4.: the average width of the confidence intervals is smaller in case than that in. 38

48 Table 9: The average variance of ˆµ ˆµ e e e

49 .5 =.5.6 =.5 Average Variance..5 Average Variance.4. =. =.8 Average Variance.5..5 Average Variance.6.4. =4 3 Average Variance Figure 3: Compare the average variance of ˆµ ˆµ VII. Simulation Study with Unknown It is assumed that is unknown in this section, so the estimate of is used. The formula to estimate is shown by Orjuela (3) using MLE method with threedimensional complete data: ˆ = n [(x i x. )(x i x. ) + (x i x. )(x i3 x.3 ) + (x i x. )(x i3 x.3 )] n [(x. i x. ) + (x i x. ) + (x i3 x.3 ) ] Everything else is the same as that in Section VI, specifically, simulation study in this part is conducted to compare the performance of estimates in case only three-dimensional observations and both three-dimensional observations and all possible incomplete observations. 4

50 . Compare the confidence regions of µ Table : Coverage Probabilities of confidence regions of µ = (,, 3)

51 Coverage Probability Coverage Probability Coverage Probability =.5.9 =.95.9 = Coverage Probability Coverage Probability =.5.9 =.95.9 Confidence level=.95 Figure 4: Compare the Coverage Probabilities of the Confidence Regions of µ Absolute value of the difference Absolute value of the difference Absolute value of the differen.3. =.5. = =4 Absolute value of the difference Absolute value of the differen.4.3. =.5. =.4. Figure 5: Absolute value of the difference between Coverage Probabilities and.95 Using the estimates of would make coverage probabilities a little lower than the confidence level.95 for both case and. It can be noted from Figure 5 that the absolute values of the difference between Coverage Probabilities and.95 are around. for each level of. 4

52 . Compare the probability of type I error and power of testing H : µ = vs.h a : µ. Compare the probability of type I error Table : The Probability of Type I error

53 Type I Error.8.6 =.5.4 = Type I Error.8.6 =.5.4 = Type I Error =4 Type I Error Type I Error α=.5 Figure 6: Compare the Type I Error Absolute value of the difference Absolute value of the difference Absolute value of the differen.3.. =.5 = =4 Absolute value of the differen Absolute value of the difference.5..5 =.5. =.3.. Figure 7: Absolute value of difference between Type I Error and α =.5 When using the estimate of, type I errors in both and are a little higher than α =.5 at each level of. It can be observed more clearly from Figure 7 that the absolute values of difference between Type I Errors and α =.5 are around. in both cases. 44

54 . Compare the power of testing H : µ = vs.h a : µ Table : The Power of testing H : µ = vs.h a : µ = (,, 3)

55 Power of testing Power of testing Power of testing.5 =6 =.5 =4.5 Power of testing Power of testing.5 =8 =.5 Figure 8: Compare the Power of Testing H : µ = vs.h a : µ = (,, 3) With regard to the power of testing, it can be observed that, when the level of is fixed, case has higher power than that of case. 3. Compare the confidence intervals of µ 3. Compare the coverage probabilities of confidence intervals of µ The coverage probabilities of the 95% confidence intervals are close to.95 for all cases, and the differences between coverage probabilities and confidence level.95 are small. 46

56 Table 3: The Coverage probabilities of confidence intervals of µ =

57 Coverage Probability Coverage Probability Coverage Probability.95 =.5.94 = = Coverage Probability Coverage Probability.95 =.5.94 = Confidence level=.95 Figure 9: Compare the Coverage Probabilities of the Confidence Intervals of µ = Absolute value of the difference Absolute value of the difference Absolute value of the differen..5 =.5 =..5 =4..5 Absolute value of the difference Absolute value of the differen..5 =.5 =..5 Figure : Absolute value of the difference between Coverage Probabilities and.95 48

58 3. Compare the average width of confidence intervals of µ Table 4: The average width of confidence intervals of µ =

59 Average Width Average Width Average Width.. =.5 =.5 =4 4 Average Width Average Width.4. =.5 = Figure : Compare the Average Widths of the Confidence Intervals of µ = With regard to the average width of the confidence intervals of µ =, the confidence intervals in are always narrower than those in. 3.3 Compare the average variance of ˆµ As presented in Figure, at each level of, the average variance of ˆµ is smaller in case than that in. This result explains our finding in section 3.: the average width of the confidence intervals is smaller in case than that in. 5

60 Table 5: The average variance of ˆµ e e e e e e e

61 Average Variance Average Variance..5 =.5 =..5 Average Variance =4 Average Variance Average Variance.4. =.5 =.5 Figure : Compare the average variance of ˆµ 4. Compare the confidence intervals of µ µ 4. The coverage probabilities of confidence intervals of µ µ (µ =, µ = ) Using the estimates of would make coverage probabilities a little lower than the confidence level.95. It can be noted from Figure 4 that the absolute values of the difference are around. for each level of. 5

62 Table 6: The coverage probabilities of confidence intervals of µ µ (µ =, µ = )

63 Coverage Probability Coverage Probability Coverage Probability = =.94.9 = Coverage Probability Coverage Probability =.5 = Confidence level=.95 Figure 3: Compare the Coverage Probabilities of the Confidence Intervals of µ µ Absolute value of the difference Absolute value of the difference Absolute value of the differen.3. =.5. = =4 Absolute value of the difference Absolute value of the differen.4.3. =.5. =.4. Figure 4: Absolute value of the difference between Coverage Probabilities and.95 54

64 4. The average width of confidence intervals of µ µ (µ =, µ = ) Table 7: The average width of confidence intervals of µ µ (µ =, µ = )

65 Average Width Average Width Average Width.4. =.5 =.5 =4 4 Average Width Average Width.5 =.5 = Figure 5: Compare the Average Widths of the Confidence Intervals of µ µ With regard to the average width of the confidence intervals of µ µ, the confidence intervals in are always narrower than those in. 4.3 Compare the average variance of ˆµ ˆµ As presented in Figure 6, at each level of, the average variance of ˆµ is smaller in case than that in. This result explains our finding in section 4.: the average width of the confidence intervals is smaller in case than that in. 56

66 Table 8: The average variance of ˆµ e e

67 Average Variance Average Variance.. =.5 =. Average Variance. =4 4 Average Variance Average Variance.5 =.5 =.5 Figure 6: Compare the average variance of ˆµ VIII. Conclusion It can be concluded from the numerical study that, compared to estimators in the case of omitting observations with missing data, using the considered estimators in this thesis would generally lead to better performance in the following aspects: higher power of testing and shorter confidence intervals, while keeping the coverage probabilities and type I errors close to the desired α level. Furthermore, when is unknown, using the estimate of would lead to the same conclusion. 58

A Study on the Correlation of Bivariate And Trivariate Normal Models

A Study on the Correlation of Bivariate And Trivariate Normal Models Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 11-1-2013 A Study on the Correlation of Bivariate And Trivariate Normal Models Maria

More information

Comparison of Some Improved Estimators for Linear Regression Model under Different Conditions

Comparison of Some Improved Estimators for Linear Regression Model under Different Conditions Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-24-2015 Comparison of Some Improved Estimators for Linear Regression Model under

More information

Inference about Reliability Parameter with Underlying Gamma and Exponential Distribution

Inference about Reliability Parameter with Underlying Gamma and Exponential Distribution Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 9-3-211 Inference about Reliability Parameter with Underlying Gamma and Exponential

More information

Study on a Hierarchy Model

Study on a Hierarchy Model Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-23-212 Study on a Hierarchy Model Suisui Che Florida International University,

More information

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data Journal of Multivariate Analysis 78, 6282 (2001) doi:10.1006jmva.2000.1939, available online at http:www.idealibrary.com on Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Statistical Inference with Monotone Incomplete Multivariate Normal Data

Statistical Inference with Monotone Incomplete Multivariate Normal Data Statistical Inference with Monotone Incomplete Multivariate Normal Data p. 1/4 Statistical Inference with Monotone Incomplete Multivariate Normal Data This talk is based on joint work with my wonderful

More information

STA 2201/442 Assignment 2

STA 2201/442 Assignment 2 STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

Exam 2. Jeremy Morris. March 23, 2006

Exam 2. Jeremy Morris. March 23, 2006 Exam Jeremy Morris March 3, 006 4. Consider a bivariate normal population with µ 0, µ, σ, σ and ρ.5. a Write out the bivariate normal density. The multivariate normal density is defined by the following

More information

Primer on statistics:

Primer on statistics: Primer on statistics: MLE, Confidence Intervals, and Hypothesis Testing ryan.reece@gmail.com http://rreece.github.io/ Insight Data Science - AI Fellows Workshop Feb 16, 018 Outline 1. Maximum likelihood

More information

On the Performance of some Poisson Ridge Regression Estimators

On the Performance of some Poisson Ridge Regression Estimators Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-28-2018 On the Performance of some Poisson Ridge Regression Estimators Cynthia

More information

Bayesian Estimation of Small Proportions Using Binomial Group Test

Bayesian Estimation of Small Proportions Using Binomial Group Test Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 11-9-212 Bayesian Estimation of Small Proportions Using Binomial Group Test Shihua

More information

More on nuisance parameters

More on nuisance parameters BS2 Statistical Inference, Lecture 3, Hilary Term 2009 January 30, 2009 Suppose that there is a minimal sufficient statistic T = t(x ) partitioned as T = (S, C) = (s(x ), c(x )) where: C1: the distribution

More information

Assessing the Effect of Prior Distribution Assumption on the Variance Parameters in Evaluating Bioequivalence Trials

Assessing the Effect of Prior Distribution Assumption on the Variance Parameters in Evaluating Bioequivalence Trials Georgia State University ScholarWorks @ Georgia State University Mathematics Theses Department of Mathematics and Statistics 8--006 Assessing the Effect of Prior Distribution Assumption on the Variance

More information

Experimental designs for multiple responses with different models

Experimental designs for multiple responses with different models Graduate Theses and Dissertations Graduate College 2015 Experimental designs for multiple responses with different models Wilmina Mary Marget Iowa State University Follow this and additional works at:

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

The Multivariate Normal Distribution 1

The Multivariate Normal Distribution 1 The Multivariate Normal Distribution 1 STA 302 Fall 2014 1 See last slide for copyright information. 1 / 37 Overview 1 Moment-generating Functions 2 Definition 3 Properties 4 χ 2 and t distributions 2

More information

New Bivariate Lifetime Distributions Based on Bath-Tub Shaped Failure Rate

New Bivariate Lifetime Distributions Based on Bath-Tub Shaped Failure Rate Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 10-30-2014 New Bivariate Lifetime Distributions Based on Bath-Tub Shaped Failure

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information

On the Existence and Uniqueness of the Maximum Likelihood Estimators of Normal and Lognormal Population Parameters with Grouped Data

On the Existence and Uniqueness of the Maximum Likelihood Estimators of Normal and Lognormal Population Parameters with Grouped Data Florida International University FIU Digital Commons Department of Mathematics and Statistics College of Arts, Sciences & Education 6-16-2009 On the Existence and Uniqueness of the Maximum Likelihood Estimators

More information

Multivariate Statistics

Multivariate Statistics Multivariate Statistics Chapter 2: Multivariate distributions and inference Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2016/2017 Master in Mathematical

More information

Numerical computation of an optimal control problem with homogenization in one-dimensional case

Numerical computation of an optimal control problem with homogenization in one-dimensional case Retrospective Theses and Dissertations Iowa State University Capstones, Theses and Dissertations 28 Numerical computation of an optimal control problem with homogenization in one-dimensional case Zhen

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

Random Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30

Random Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30 Random Vectors 1 STA442/2101 Fall 2017 1 See last slide for copyright information. 1 / 30 Background Reading: Renscher and Schaalje s Linear models in statistics Chapter 3 on Random Vectors and Matrices

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

Bayesian Inference. Chapter 9. Linear models and regression

Bayesian Inference. Chapter 9. Linear models and regression Bayesian Inference Chapter 9. Linear models and regression M. Concepcion Ausin Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering

More information

Math 152. Rumbos Fall Solutions to Assignment #12

Math 152. Rumbos Fall Solutions to Assignment #12 Math 52. umbos Fall 2009 Solutions to Assignment #2. Suppose that you observe n iid Bernoulli(p) random variables, denoted by X, X 2,..., X n. Find the LT rejection region for the test of H o : p p o versus

More information

Statistical Inference with Monotone Incomplete Multivariate Normal Data

Statistical Inference with Monotone Incomplete Multivariate Normal Data Statistical Inference with Monotone Incomplete Multivariate Normal Data p. 1/4 Statistical Inference with Monotone Incomplete Multivariate Normal Data This talk is based on joint work with my wonderful

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida First Year Examination Department of Statistics, University of Florida August 19, 010, 8:00 am - 1:00 noon Instructions: 1. You have four hours to answer questions in this examination.. You must show your

More information

Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples

Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples Nobumichi Shutoh, Masashi Hyodo and Takashi Seo 2 Department of Mathematics, Graduate

More information

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics The candidates for the research course in Statistics will have to take two shortanswer type tests

More information

Parametric Techniques Lecture 3

Parametric Techniques Lecture 3 Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to

More information

SOLUTION FOR HOMEWORK 6, STAT 6331

SOLUTION FOR HOMEWORK 6, STAT 6331 SOLUTION FOR HOMEWORK 6, STAT 633. Exerc.7.. It is given that X,...,X n is a sample from N(θ, σ ), and the Bayesian approach is used with Θ N(µ, τ ). The parameters σ, µ and τ are given. (a) Find the joinf

More information

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ). .8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics

More information

Compositions, Bijections, and Enumerations

Compositions, Bijections, and Enumerations Georgia Southern University Digital Commons@Georgia Southern Electronic Theses & Dissertations COGS- Jack N. Averitt College of Graduate Studies Fall 2012 Compositions, Bijections, and Enumerations Charles

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

Associated Hypotheses in Linear Models for Unbalanced Data

Associated Hypotheses in Linear Models for Unbalanced Data University of Wisconsin Milwaukee UWM Digital Commons Theses and Dissertations May 5 Associated Hypotheses in Linear Models for Unbalanced Data Carlos J. Soto University of Wisconsin-Milwaukee Follow this

More information

Statistics and Econometrics I

Statistics and Econometrics I Statistics and Econometrics I Point Estimation Shiu-Sheng Chen Department of Economics National Taiwan University September 13, 2016 Shiu-Sheng Chen (NTU Econ) Statistics and Econometrics I September 13,

More information

Introduction to Normal Distribution

Introduction to Normal Distribution Introduction to Normal Distribution Nathaniel E. Helwig Assistant Professor of Psychology and Statistics University of Minnesota (Twin Cities) Updated 17-Jan-2017 Nathaniel E. Helwig (U of Minnesota) Introduction

More information

Parametric Techniques

Parametric Techniques Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure

More information

Random Vectors and Multivariate Normal Distributions

Random Vectors and Multivariate Normal Distributions Chapter 3 Random Vectors and Multivariate Normal Distributions 3.1 Random vectors Definition 3.1.1. Random vector. Random vectors are vectors of random 75 variables. For instance, X = X 1 X 2., where each

More information

Part 4: Multi-parameter and normal models

Part 4: Multi-parameter and normal models Part 4: Multi-parameter and normal models 1 The normal model Perhaps the most useful (or utilized) probability model for data analysis is the normal distribution There are several reasons for this, e.g.,

More information

Problem Selected Scores

Problem Selected Scores Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected

More information

Comparing Group Means When Nonresponse Rates Differ

Comparing Group Means When Nonresponse Rates Differ UNF Digital Commons UNF Theses and Dissertations Student Scholarship 2015 Comparing Group Means When Nonresponse Rates Differ Gabriela M. Stegmann University of North Florida Suggested Citation Stegmann,

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An

More information

Lifetime prediction and confidence bounds in accelerated degradation testing for lognormal response distributions with an Arrhenius rate relationship

Lifetime prediction and confidence bounds in accelerated degradation testing for lognormal response distributions with an Arrhenius rate relationship Scholars' Mine Doctoral Dissertations Student Research & Creative Works Spring 01 Lifetime prediction and confidence bounds in accelerated degradation testing for lognormal response distributions with

More information

STATISTICAL INFERENCE IN ACCELERATED LIFE TESTING WITH GEOMETRIC PROCESS MODEL. A Thesis. Presented to the. Faculty of. San Diego State University

STATISTICAL INFERENCE IN ACCELERATED LIFE TESTING WITH GEOMETRIC PROCESS MODEL. A Thesis. Presented to the. Faculty of. San Diego State University STATISTICAL INFERENCE IN ACCELERATED LIFE TESTING WITH GEOMETRIC PROCESS MODEL A Thesis Presented to the Faculty of San Diego State University In Partial Fulfillment of the Requirements for the Degree

More information

Examining the accuracy of the normal approximation to the poisson random variable

Examining the accuracy of the normal approximation to the poisson random variable Eastern Michigan University DigitalCommons@EMU Master's Theses and Doctoral Dissertations Master's Theses, and Doctoral Dissertations, and Graduate Capstone Projects 2009 Examining the accuracy of the

More information

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer

More information

Fundamental Probability and Statistics

Fundamental Probability and Statistics Fundamental Probability and Statistics "There are known knowns. These are things we know that we know. There are known unknowns. That is to say, there are things that we know we don't know. But there are

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science

UNIVERSITY OF TORONTO Faculty of Arts and Science UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator

More information

INFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS. Tao Jiang. A Thesis

INFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS. Tao Jiang. A Thesis INFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS Tao Jiang A Thesis Submitted to the Graduate College of Bowling Green State University in partial fulfillment of the

More information

Estimation of parametric functions in Downton s bivariate exponential distribution

Estimation of parametric functions in Downton s bivariate exponential distribution Estimation of parametric functions in Downton s bivariate exponential distribution George Iliopoulos Department of Mathematics University of the Aegean 83200 Karlovasi, Samos, Greece e-mail: geh@aegean.gr

More information

Introduction to Maximum Likelihood Estimation

Introduction to Maximum Likelihood Estimation Introduction to Maximum Likelihood Estimation Eric Zivot July 26, 2012 The Likelihood Function Let 1 be an iid sample with pdf ( ; ) where is a ( 1) vector of parameters that characterize ( ; ) Example:

More information

Dynamic System Identification using HDMR-Bayesian Technique

Dynamic System Identification using HDMR-Bayesian Technique Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

An Introduction to Multivariate Statistical Analysis

An Introduction to Multivariate Statistical Analysis An Introduction to Multivariate Statistical Analysis Third Edition T. W. ANDERSON Stanford University Department of Statistics Stanford, CA WILEY- INTERSCIENCE A JOHN WILEY & SONS, INC., PUBLICATION Contents

More information

Quick Tour of Basic Probability Theory and Linear Algebra

Quick Tour of Basic Probability Theory and Linear Algebra Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions

More information

Machine Learning, Fall 2012 Homework 2

Machine Learning, Fall 2012 Homework 2 0-60 Machine Learning, Fall 202 Homework 2 Instructors: Tom Mitchell, Ziv Bar-Joseph TA in charge: Selen Uguroglu email: sugurogl@cs.cmu.edu SOLUTIONS Naive Bayes, 20 points Problem. Basic concepts, 0

More information

Approximation of Posterior Means and Variances of the Digitised Normal Distribution using Continuous Normal Approximation

Approximation of Posterior Means and Variances of the Digitised Normal Distribution using Continuous Normal Approximation Approximation of Posterior Means and Variances of the Digitised Normal Distribution using Continuous Normal Approximation Robert Ware and Frank Lad Abstract All statistical measurements which represent

More information

Practice Exam 1. (A) (B) (C) (D) (E) You are given the following data on loss sizes:

Practice Exam 1. (A) (B) (C) (D) (E) You are given the following data on loss sizes: Practice Exam 1 1. Losses for an insurance coverage have the following cumulative distribution function: F(0) = 0 F(1,000) = 0.2 F(5,000) = 0.4 F(10,000) = 0.9 F(100,000) = 1 with linear interpolation

More information

Testing equality of two mean vectors with unequal sample sizes for populations with correlation

Testing equality of two mean vectors with unequal sample sizes for populations with correlation Testing equality of two mean vectors with unequal sample sizes for populations with correlation Aya Shinozaki Naoya Okamoto 2 and Takashi Seo Department of Mathematical Information Science Tokyo University

More information

Empirical Likelihood Inference for Two-Sample Problems

Empirical Likelihood Inference for Two-Sample Problems Empirical Likelihood Inference for Two-Sample Problems by Ying Yan A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Mathematics in Statistics

More information

BAYESIAN RELIABILITY ANALYSIS FOR THE GUMBEL FAILURE MODEL. 1. Introduction

BAYESIAN RELIABILITY ANALYSIS FOR THE GUMBEL FAILURE MODEL. 1. Introduction BAYESIAN RELIABILITY ANALYSIS FOR THE GUMBEL FAILURE MODEL CHRIS P. TSOKOS AND BRANKO MILADINOVIC Department of Mathematics Statistics, University of South Florida, Tampa, FL 3362 ABSTRACT. The Gumbel

More information

Working correlation selection in generalized estimating equations

Working correlation selection in generalized estimating equations University of Iowa Iowa Research Online Theses and Dissertations Fall 2011 Working correlation selection in generalized estimating equations Mi Jin Jang University of Iowa Copyright 2011 Mijin Jang This

More information

To Estimate or Not to Estimate?

To Estimate or Not to Estimate? To Estimate or Not to Estimate? Benjamin Kedem and Shihua Wen In linear regression there are examples where some of the coefficients are known but are estimated anyway for various reasons not least of

More information

TAMS39 Lecture 2 Multivariate normal distribution

TAMS39 Lecture 2 Multivariate normal distribution TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution

More information

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1 Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the

More information

Introduction to Machine Learning. Lecture 2

Introduction to Machine Learning. Lecture 2 Introduction to Machine Learning Lecturer: Eran Halperin Lecture 2 Fall Semester Scribe: Yishay Mansour Some of the material was not presented in class (and is marked with a side line) and is given for

More information

The Delta Method and Applications

The Delta Method and Applications Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability

More information

Stochastic Realization of Binary Exchangeable Processes

Stochastic Realization of Binary Exchangeable Processes Stochastic Realization of Binary Exchangeable Processes Lorenzo Finesso and Cecilia Prosdocimi Abstract A discrete time stochastic process is called exchangeable if its n-dimensional distributions are,

More information

1 Data Arrays and Decompositions

1 Data Arrays and Decompositions 1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is

More information

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Jae-Kwang Kim 1 Iowa State University June 26, 2013 1 Joint work with Shu Yang Introduction 1 Introduction

More information

Notes on the Multivariate Normal and Related Topics

Notes on the Multivariate Normal and Related Topics Version: July 10, 2013 Notes on the Multivariate Normal and Related Topics Let me refresh your memory about the distinctions between population and sample; parameters and statistics; population distributions

More information

MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models

MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models 1/13 MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models Dominique Guillot Departments of Mathematical Sciences University of Delaware May 4, 2016 Recall

More information

Chapter 4: Asymptotic Properties of the MLE (Part 2)

Chapter 4: Asymptotic Properties of the MLE (Part 2) Chapter 4: Asymptotic Properties of the MLE (Part 2) Daniel O. Scharfstein 09/24/13 1 / 1 Example Let {(R i, X i ) : i = 1,..., n} be an i.i.d. sample of n random vectors (R, X ). Here R is a response

More information

Confidence Intervals for the Ratio of Two Exponential Means with Applications to Quality Control

Confidence Intervals for the Ratio of Two Exponential Means with Applications to Quality Control Western Kentucky University TopSCHOLAR Student Research Conference Select Presentations Student Research Conference 6-009 Confidence Intervals for the Ratio of Two Exponential Means with Applications to

More information

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS Communications in Statistics - Simulation and Computation 33 (2004) 431-446 COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS K. Krishnamoorthy and Yong Lu Department

More information

Estimators as Random Variables

Estimators as Random Variables Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until

More information

Department of Statistical Science FIRST YEAR EXAM - SPRING 2017

Department of Statistical Science FIRST YEAR EXAM - SPRING 2017 Department of Statistical Science Duke University FIRST YEAR EXAM - SPRING 017 Monday May 8th 017, 9:00 AM 1:00 PM NOTES: PLEASE READ CAREFULLY BEFORE BEGINNING EXAM! 1. Do not write solutions on the exam;

More information

COMS 4721: Machine Learning for Data Science Lecture 1, 1/17/2017

COMS 4721: Machine Learning for Data Science Lecture 1, 1/17/2017 COMS 4721: Machine Learning for Data Science Lecture 1, 1/17/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University OVERVIEW This class will cover model-based

More information

Parameter Estimation

Parameter Estimation Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y

More information

A Distribution of the First Order Statistic When the Sample Size is Random

A Distribution of the First Order Statistic When the Sample Size is Random East Tennessee State University Digital Commons @ East Tennessee State University Electronic Theses and Dissertations 5-2017 A Distribution of the First Order Statistic When the Sample Size is Random Vincent

More information

FIRST YEAR EXAM Monday May 10, 2010; 9:00 12:00am

FIRST YEAR EXAM Monday May 10, 2010; 9:00 12:00am FIRST YEAR EXAM Monday May 10, 2010; 9:00 12:00am NOTES: PLEASE READ CAREFULLY BEFORE BEGINNING EXAM! 1. Do not write solutions on the exam; please write your solutions on the paper provided. 2. Put the

More information

Jackknife Empirical Likelihood for the Variance in the Linear Regression Model

Jackknife Empirical Likelihood for the Variance in the Linear Regression Model Georgia State University ScholarWorks @ Georgia State University Mathematics Theses Department of Mathematics and Statistics Summer 7-25-2013 Jackknife Empirical Likelihood for the Variance in the Linear

More information

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 2013-14 We know that X ~ B(n,p), but we do not know p. We get a random sample

More information

Bayesian Inference: Probit and Linear Probability Models

Bayesian Inference: Probit and Linear Probability Models Utah State University DigitalCommons@USU All Graduate Plan B and other Reports Graduate Studies 5-1-2014 Bayesian Inference: Probit and Linear Probability Models Nate Rex Reasch Utah State University Follow

More information

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables. Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just

More information

Next is material on matrix rank. Please see the handout

Next is material on matrix rank. Please see the handout B90.330 / C.005 NOTES for Wednesday 0.APR.7 Suppose that the model is β + ε, but ε does not have the desired variance matrix. Say that ε is normal, but Var(ε) σ W. The form of W is W w 0 0 0 0 0 0 w 0

More information

CS 195-5: Machine Learning Problem Set 1

CS 195-5: Machine Learning Problem Set 1 CS 95-5: Machine Learning Problem Set Douglas Lanman dlanman@brown.edu 7 September Regression Problem Show that the prediction errors y f(x; ŵ) are necessarily uncorrelated with any linear function of

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

COS513 LECTURE 8 STATISTICAL CONCEPTS

COS513 LECTURE 8 STATISTICAL CONCEPTS COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions

More information

Hypothesis Testing. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA

Hypothesis Testing. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA Hypothesis Testing Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA An Example Mardia et al. (979, p. ) reprint data from Frets (9) giving the length and breadth (in

More information

You can compute the maximum likelihood estimate for the correlation

You can compute the maximum likelihood estimate for the correlation Stat 50 Solutions Comments on Assignment Spring 005. (a) _ 37.6 X = 6.5 5.8 97.84 Σ = 9.70 4.9 9.70 75.05 7.80 4.9 7.80 4.96 (b) 08.7 0 S = Σ = 03 9 6.58 03 305.6 30.89 6.58 30.89 5.5 (c) You can compute

More information

Gaussian Models (9/9/13)

Gaussian Models (9/9/13) STA561: Probabilistic machine learning Gaussian Models (9/9/13) Lecturer: Barbara Engelhardt Scribes: Xi He, Jiangwei Pan, Ali Razeen, Animesh Srivastava 1 Multivariate Normal Distribution The multivariate

More information