DIRECT CALCULATION OF MAXIMUM LIKELIHOOD ESTIMATOR FOR THE BIVARIATE POISSON DISTRIBUTION

Similar documents
THE STRUCTURE OF MULTIVARIATE POISSON DISTRIBUTION

A New Demerit Control Chart for Monitoring the Quality of Multivariate Poisson Processes. By Jeh-Nan Pan Chung-I Li Min-Hung Huang

COMPLEX ALMOST CONTACT STRUCTURES IN A COMPLEX CONTACT MANIFOLD

Linear Regression and Its Applications

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Mathematical statistics

On a simple construction of bivariate probability functions with fixed marginals 1

Copula Regression RAHUL A. PARSA DRAKE UNIVERSITY & STUART A. KLUGMAN SOCIETY OF ACTUARIES CASUALTY ACTUARIAL SOCIETY MAY 18,2011

Non-separable AF-algebras

Collapsed Variational Inference for Sum-Product Networks

The Delta Method and Applications

EEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as

Introduction to Maximum Likelihood Estimation

STA 260: Statistics and Probability II

Formulas for probability theory and linear models SF2941

6.3 Forecasting ARMA processes

Statistics (1): Estimation

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics)

Practice Problems Section Problems

Multivariate Zero-Inflated Poisson Regression

Mathematical Statistics 1 Math A 6330

JOINT DISTRIBUTION OF THE FIRST HITTING TIME AND FIRST HITTING PLACE FOR A RANDOM WALK TADASHI NAKAJIMA

CHAPTER 1 INTRODUCTION

6348 Final, Fall 14. Closed book, closed notes, no electronic devices. Points (out of 200) in parentheses.

01 Probability Theory and Statistics Review

EC3062 ECONOMETRICS. THE MULTIPLE REGRESSION MODEL Consider T realisations of the regression equation. (1) y = β 0 + β 1 x β k x k + ε,

More on Distribution Function

Machine learning - HT Maximum Likelihood

Probability Distributions - Lecture 5

Solving the 100 Swiss Francs Problem

We introduce methods that are useful in:

EE5319R: Problem Set 3 Assigned: 24/08/16, Due: 31/08/16

MATHEMATICS 217 NOTES

1 Exercises for lecture 1

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Variational Mixture of Gaussians. Sargur Srihari

Example 1. Let be a random sample from. Please find a good point estimator for

ON THE AVERAGE NUMBER OF REAL ZEROS OF A CLASS OF RANDOM ALGEBRAIC CURVES

A skew Laplace distribution on integers

Bivariate distributions

1. Density and properties Brief outline 2. Sampling from multivariate normal and MLE 3. Sampling distribution and large sample behavior of X and S 4.

Lecture 23 Maximum Likelihood Estimation and Bayesian Inference

Part 6: Multivariate Normal and Linear Models

Weak convergence in Probability Theory A summer excursion! Day 3

Chapter 3. Matrices. 3.1 Matrices

Review: Probability. BM1: Advanced Natural Language Processing. University of Potsdam. Tatjana Scheffler

Supermodular ordering of Poisson arrays

Asymptotic properties of the likelihood ratio test statistics with the possible triangle constraint in Affected-Sib-Pair analysis

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours

Multivariate Normal-Laplace Distribution and Processes

3. Probability and Statistics

Maximum Likelihood Estimation

On prediction and density estimation Peter McCullagh University of Chicago December 2004

ON MIXED BOUNDARY VALUE PROBLEMS FOR PARABOLIC EQUATIONS IN SINGULAR DOMAINS

Mathematics for Business and Economics - I. Chapter 5. Functions (Lecture 9)

2. Matrix Algebra and Random Vectors

MacMahon s Partition Analysis VIII: Plane Partition Diamonds

Taylor and Laurent Series

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

Multivariate Statistics

SOLUTION FOR HOMEWORK 8, STAT 4372

Introduction to Statistics. By: Ewa Paszek

Statistics Ph.D. Qualifying Exam: Part II November 9, 2002

ACM 116: Lectures 3 4

Notes on Random Vectors and Multivariate Normal

Maximum Likelihood Estimation

Loglikelihood and Confidence Intervals

Probability Theory for Machine Learning. Chris Cremer September 2015

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA

5-DESIGNS IN AFFINE SPACES

Statistical Estimation

An Introduction to Expectation-Maximization

THE SUM OF THE ELEMENTS OF THE POWERS OF A MATRIX

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2011

Review of Probabilities and Basic Statistics

Conjugate Predictive Distributions and Generalized Entropies

Overall Objective Priors

Bayes Model Selection with Path Sampling: Factor Models

4. CONTINUOUS RANDOM VARIABLES

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

Probability reminders

Maximum variance formulation

Mean-field dual of cooperative reproduction

arxiv: v1 [math.ap] 17 May 2017

Integer-Valued Polynomials

Lecture 5: Moment Generating Functions

Chapter 5. Chapter 5 sections

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk

MATH 205C: STATIONARY PHASE LEMMA

On the bivariate Skellam distribution. hal , version 1-22 Oct Noname manuscript No. (will be inserted by the editor)

Near-exact approximations for the likelihood ratio test statistic for testing equality of several variance-covariance matrices

Lecture Note 1: Probability Theory and Statistics

Lecture 14: Multivariate mgf s and chf s

Eötvös Loránd University, Budapest. 13 May 2005

MLE for a logit model

ON THE EVOLUTION OF ISLANDS

THE APPROXIMATE SOLUTION OF / = F(x,y)

Stat 5101 Lecture Notes

Hidden Markov models: definition and properties

Geometric Skew-Normal Distribution

Transcription:

K KAWAMURA KODAI MATH J 7 (1984), 211 221 DIRECT CALCULATION OF MAXIMUM LIKELIHOOD ESTIMATOR FOR THE BIVARIATE POISSON DISTRIBUTION BY KAZUTOMO KAWAMURA Summary To estimate the parameter vector λ of bivariate Poisson distribution [1], [2] we would like to calculate maximum likelihood estimator (MLE) 1 This MLE 1 has not a simple expression as X y S 2, etc We only have information about MLE 1 by normal equations its variation forms [3], Holgate [4] shows the asymptotic property of MLE 1 In this paper we would like to show the calculating method of MLE 1 The method will be constructed by direct calculation of likelihood function by a searching routine of MLE 1 which maximizes the function value A sequence of rom numbers come from a bivariate Poisson distribution P(λ) will be shown The change of the value of likelihood function varying parameter λ in our rule will be calculated the work of the searching routine will be discussed in detail In the last part of this paper a numerical interpretation of our routine will be shown Section 1 Bivariate Poisson distribution P(λ) If (X, Y) has a bivariate distribution, Λ v n ' Λ ~ UJ βtt-kβ\γ\δr Γ r+δ=ι we shall call (X, Y) has a bivariate Poisson distribution P(λ), where k, I, β, γ δ are nonnegative integers nonnegative λ 10 are called as parameters λ means a vector of the three parameters (λ 10, λ n ) The moment generating function of the distribution is given by And the marginal distribution is given by Received July 7, 1983 211

212 KAZUTOMO KAWAMURA P(X=k) = p(k; λ lo +λ n ) {k=0, 1,2, ), P(Y=l)=p(l ; λ ol +λ 11 ) (/=0, 1, 2, - ), where (* Λ) is an univariate Poisson density about the distribution from the equalities: We will get more information E(X)=λ lo +λ n, V3r{X)=λ ί0 +λ n, E(Y)=λ Q1 +λ 11, Var(Y)=λ Oί +λ n, Cov(Z, Y)=λu If (Z, Y) has a bivariate Poisson distribution P(λ) then we will get the decomposition rule where X lo, X ol X n are mutually independent univariate Poisson distributions with parameter λ 10 respectively But getting a sample (x, y) of the distribution, we do not know the decomposed samples x 10, x Q1 x n satisfying the decomposition rule of the last two equalities This is the main reason why it is difficult to estimate the parameters from the samples of the distribution Section 2 MLE of the parameter λ Practical estimation of the parameter λ = (λ i0, λ 0L, λ n ) Denote n independent sample variables of a bivariate Poisson distribution P(λ) as (XuYJΛXtoY*), ~ ΛXn,Yn) denote the practical samples as (χi> yi), (xz, y*), - y (xn, y n ) The maximum likelihood estimator MLE of λ lo +λn will be given by Σ =i#ι/ft MLE of λ O i+λ n by Σ?=i3^/w To estimate the parameters λ 10, λ Q1 individually we need to discuss the method of the estimation in a more delicate way 2-1 Practical estimation of covariance value Λ n We shall consider the problem how to estimate λ n THEOREM 1 From n independent samples {x 1} y x ), (x 2, y 2 ),, (x n, y n ), MLE of vίio, ^oi λn is given by next three equalities, n ι=i p(x lf y τ )

DIRECT CALCULATION OF MAXIMUM LIKELIHOOD ESTIMATOR 213 n ι=ι p(χ t, y t ) where p{k, l) P{X k, Y l) with parameters λ 1Q, λ oι We usually call the relations the normal equalities Proof of the theorem To maximize the likelihood function Π?=i/>Uι, y%) with respect to the parameters λ 10, we have to differentiate the logarithm of likelihood function about the parameters to put its value zero, ~~πp(χr, yι)=σ-^r- ( r ΠίU, yθ=θ> σ/lo 1=1 1 = 1 P\X%y yi) 1 = 1 where the differential of the probability density is given by a convenient relation 3 _ Because, we have 3 (/3-1)! γ Ϊ a! ^ ^ 0! r! < r+δ=ί 2/3 }Γ ί<5 /lo/ol/ ^g-^io-^oi-^ii-,y? (fe Λ Therefore, we can verify, = p(k-l,l)-p(k, I) i, yύ ι=i t, 3 ;^ ι Π ρ{χι> y%) Γ «p(xi l, yι) 1", v The equivalent condition of the normal equation

214 KAZUTOMO KAWAMURA 1$ given by 4rliρ{χr, y τ )=0 oa 10 1=1 n p{χt-l, y τ ) *=i By using similar calculations about P(χι, yτ) d n gj- Tlρ(χι, y we will be given the equivalent conditions» ί(xi 1, ^i 1) respectively THEOREM 2 MLE o/ ^10, λ O ι denoted as λ 10 satisfy Xw x ^o]+ίn=j 7 - Proof of the theorem From the first normal equation a relation: we have ^ί(ft, l)=λ 1Q p(k-l, l)+λ n p(k-l, /-I), Λoί(^ι 1, yι)=xιp(χι, yι) Λnp(χi l, yt Then, the first normal equation is expressed as following: ^ χtp(χι, yiϊ λnpjxj l, y i l) λ 1<i ρ{x ι, y t )

DIRECT CALCULATION OF MAXIMUM LIKELIHOOD ESTIMATOR 215 by the third normal equation, we have a concluding equation Σxi-nW 10 +ω=0 1 = 1 n ^10 + ^11= Σ xjπ X 1 = 1 Similarly, we have another concluding equation from the second the third normal equations n Λoi+Λu= Σ yι/n = y t=l These two simple relations come only from the normal equations λ 10 λ u must be symbolized by MLE λ 10 Further calculation about λ ί0, λ O i is very difficult we have to calculate the individual estimator by numerical method In the next section we shall treat the practical calculating method showing how to get MLE λ lιh λ in In 2-2 Practical calculation of MLE l~(λ 10, λ Q1, l n ) THEOREM 3 MLE λ ί0, hi satisfy the relations ^10+^11 =, Λ01+Λ11 y 0^ί n^min(f, y) which maximize the logarithm of likelihood function logπ?=iί(#ι, y%) Proof of the theorem This theorem is a representation of the last theorem for our calculation of MLE Denote m=min(f, y) for simplicity of notation, λ 10, λ 01 are nonnegative parameters, so that we have 0^λ n^m To get λ n we have to move λ n in the interval [0, m] which maximize the logarithm of likelihood function /// To calculate MLE λ 10 numerically, we have to compare the value of /// on the scanning space of the three parameters This theorem states that we may find MLE which maximizes the value of /// in one dimensional parameter space, that is, our scanning space of the parameters reduce to one dimensional subspace ^io = Λ ^ii, ^01 =5 ^ii Λ n e[0, m] At the begining of computation, we compare the value /// in the rule ytn=00, 10, 20, ^m = (Jc, y) λ 1Q =x λ llf λ ol = y λ n, that is, λ n moves on nonnegative integers from 0 to the integer lower than m We will find one λ n which maximizes /// or two λ n which maximize the function Usually we get only one λ n which maximizes /// occasionally we look for two λ n which

216 KAZUTOMO KAWAMURA maximize the function In the first case we have to look for λ rι in the interval of both sides of λ n which maximizes the function we have adjusted the scanning step in the reduced parameter space to one quater of preceding scanning step 10-190 -190 u -200-200 llf -210-210 Fig 1 2-3 Automatic reduction rule of λ n scanning space Let us discuss in detail the reduction rule We put the scanning space of λ 1L as D o initially, D 0 ={0, 1, 2, -, m 0 }, where m 0 equals to the maximum integer lower than m=min(jc, y), (case 1) λ n =0 maximizes /// in the scanning space D o (case 2) One of λ n = k in 1, 2,, m o l maximizes /// in the scanning space Do (case 3) Two of λ lλ k, &-f-l in 0, 1, 2,, m 0 maximize ίίf in the scanning space IV (case 4) λ n =?n 0 maximizes /// in the scanning space Do- Scanning λ n <=D 0 as to maximize ///, we have four cases (case 1),, (case 4) In every case, to compute MLE λ n more detail, we have to reduce the scanning space the scanning step We have used the secondary scanning step as one quater of the preceding one Then we have a new scanning space D L as following: (case 1) D^{θ, \, \, f, l (case 2) ^ = {^-1, k- j, k-l+j, k-ί+j, k, k+j, k+j, k + j, (case 3) D^{k, k + j, k+j, k+j, k+l (case 4) Dί γn,, 1, m 0 1+γ, m 0 1+, m o l+, m 0, m o 4-j,

DIRECT CALCULATION OF MAXIMUM LIKELIHOOD ESTIMATOR 217 where ί is the maximum integer in 0, 1, 2, 3 such that m o +τ-^^r=min(i :, y) In the primary step, we compare the value of /// by λ n scanning only the integral values on [0, rrί], as to find λ n which maximizes /// Secondary step, if we find one λ n maximizing /// as (case 2), we can reduce the scanning interval [0, m] to [& 1, k+l~] where k is the value of λ n maximizing /// If we find one Λn=0 maximizing /// as (case 1), we can reduce the scanning interval to [0, 1] if we find one λ u m Q maximizing /// as (case 4), we can reduce the scanning interval to [m 0 1, m] If we find two λ n =k, k+1 (k=0 f 1,, m 0 1) maximizing /// we can reduce the scanning interval to [>, k+γ\ as (case 3) Our scanning of λ n is made in the reduced interval the scanning step is adjusted to a quater step of preceding scanning step, where the scanning step of λ n in the reduced interval may change under another reduction rule the total computing time will change Then we can reduce our scanning space D o to D x as denoted above Under this inductive routine, if we set 0001 as the exactness of the calculation of MLE λ n, then we will obtain MLE λ u involving the exactness after high resolution computing method Section 3 A sequence of rom numbers from P(λ) computation of MLE In by computer In this section a result of computer simulation will be demonstrated, one is a change of /// under our reduced linear space, the other is a computing process of finding MLE X n 3-1 Simulation of a sequence of bivariate Poisson distribution P(λ) Given parameters λ 10, λ O ι λ llf X lo, X ol X n are independent univariate Poisson distributions then (X, Y) from a bivariate Poisson distribution is expressed by X=X 10 -\-X n y = ^01+^11 To make a sequence of rom numbers of the distribution P{λ), we should make three series of independent univariate Poisson rom variables X 10, X ol X n Following figure is our flow-chart of making bivariate Poisson rom variables

218 KAZUTOMO KAWAMURA -XΊ 0 ; Poisson Poisson X ύι X n Poisson r V r V r V parameter parameter parameter ΛQX hi If the number of bivariate Poisson rvs equals to n then the system stops proceeds to the next calculations Poisson rom variables (X lf YJΛX*, Y\-ΛXn, Table 1 Histogram of Poisson rvs Computation of statistics X, 7, S 2 X, S 2 r, S xr Table 2 Computation of /// on our reduced parameter space Automatic computation of MLE j? u See 3-2 See 3-3 Fig 2 Poisson rom variables with parameters Λ 10 = 30, λ ol =3O λ λl =2Λ (6,5) (5,7) (4,3) (3,3) (9,5) (3,7) (6,2) (7,5) (2,4) (3,4) (5,6) (6,5) (7, 3) (3, 3) (5, 6) (6, 4) (6,4) (4,5) (6, 2) (2,3) (6,4) (4,4) (6,6) (7,5) (9, (5, 7) 7) (3, 5) (6, 3) (9, 7) (5, 5) (3, 2) (3, 3) (1,3) (1,6) (5,7) (6,2) (4,6) (6, 10) (1,4) (4, 6) (1,5) (5,2) (5,3) (9, 10) (7,9) (6,6) (5,6) (4,7) (3,6) (4, 5) (10, 2) (4, 4) (4,8) (8, Π) (7,9) (4,2) (7,4) (5,8) (7, 4) (5, 5) (4, 3) (5, 8) (5, 5) (5, 5) (4,4) (4,6) (2,7) (1, 1) (11, 13) (6,7) (3, 1) (5,2) (6,8) (4,9) (4,5) (5,8) (12, 10) (4, 7) (2,2) (9,7) (8,3) (5,3) (8, 9) (4, 1) (6, 2) (7, 7) (9,7) (3, 1) (3,4) (4,4) (1,5) (1,6) (3, 1) (4, 8) (7, 6) (3,4) (7, 4) (2, 2) (1 1) (4,3) Table 1

DIRECT CALCULATION OF MAXIMUM LIKELIHOOD ESTIMATOR 219 Histogram of the Poisson rom variables /=0, 1, 2, -, 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 - ----'-- 1 2 o o o Statistics based on the Poisson rom variables 1=493, F=503, S 2 *=53251 SI-=61891 where Table 2 3-2 The change of /// under our scanning rule Our reduced one dimensional parameter space is expressed as λ lo =X λn, λ ol =Y λ n i n [0,?n] where m=mm(x, y) From x=493, 5=503, we get m=493 so that our reduced parameter space is expressed; oi=493-λ n, 493] We tried to calculate /// on the space under a scanning step 01 of the third parameter λ n Next table graph explain a relation of variable λ n /// And we will find MLE of λ n close to Λ n =19

220 KAZUTOMO KAWAMURA 00 10 20 30 40 llf -1932464-1931536 -1930868-1930479 -1930389-1930625 -1931216-1932196 -1933607-1935494 -1937915 Fig 3 3-3 Automatic computing process of MLE λ n If we would like to know the function of our automatic computing process of MLE λ n, we can easily pull out the scanning spaces D Q, D u /// values on each spaces Following table graph explain the function Lower limit D o 0- A 1- D 2 1+3/4- D z 1+3/4+1/16- D 4 1+3/4+1/16+3/64- D 5 1 + 3/4+1/16+3/64+4/256 - D 6 1 + 3/4+1/16+3/64+4/256- + 3/1024 Upper limit -4 1 Scan -3 1/4-1+5/4 1/16-1+3/4+3/16 1/64-1+3/4+1/16+5/64 1/256-1+3/4+1/16+3/64+6/256 1/1024 1+3/4+1/16+3/64+4/256 +5/1024 Table 4

DIRECT CALCULATION OF MAXIMUM LIKELIHOOD ESTIMATOR 221 Graph of upper lower limits of reduced scanning spaces D o, D lf - Fig 4 D, A A A A A After our automatic computing process of MLE, we get MLE X n =1879 from the assertion of theorem 3, we get MLE ί lo =493 1879=3051 MLE 3oi=503 1879=3151 These are the maximum likelihood estimators of the bivariate Poisson rom variables simulated by computer We have another estimator for λ n, The parameter used to the simulation was Λ 10 =30, Λ 01 =30 Λ n =20 And our MLE is expressed as ί lo =3O51, ί 0 i=3151 ί n =1879 Another estimation for λ n is given by Sjrr=25021 Remark The aim of the last section is to answer the questions: how to make the bivariate Poisson rom variables (by simulation) how to calculate MLE of the parameter λ=(a 10, 2 01, 2 n ) But the next question to answer would be how to check the fitness of the bivariate Poisson distribution for given bivariate data (x lf yj, (x 2, y 2 ),, (x n, y n ) a s given in table 1 REFERENCES [ 1 ] KAWAMURA, K, The structure of bivariate Poisson distribution, Kδdai Math Sem Rep, 25, No 2, (1973) [2] POLAK, M, Poisson approximation for sums of independent bivariate Bernoulli vectors, Kodal Math J, 5, No 3, (1982) [3] JOHNSON, NL AND KOTZ S, Discrete distributions, Houghton Mifίiin Co (1969) [4] HOLGATE, P, Estimation for the bivaliate Poisson distribution, Biometrika, 51, (1964), 241-245 TOKYO INSTITUTE OF TECHNOLOGY