On Approximating a Generalized Binomial by Binomial and Poisson Distributions
|
|
- Bennett Lamb
- 6 years ago
- Views:
Transcription
1 International Journal of Statistics and Systems. ISSN Volume 3, Number (008), pp. 3 4 Research India Publications On Approximating a Generalized inomial by inomial and Poisson Distributions P. Wongkasem and K. Teerapabolarn * Department of Mathematics, Faculty of Science urapha University, Chonburi, 03 Thailand R. Gulasirima Applied Statistics Program, Faculty of Science and Technology Suan Dusit Rajabhat University, angkok, 0300 Thailand *Corresponding author kanint@buu.ac.th Abstract The aim of this paper is to approximate a generalized binomial distribution by binomial and Poisson distributions. Using the ω-function and the two Stein identities, the main result of each approximation in terms of the total variation distance between two distributions and its upper bound is obtained. Moreover, some numerical examples have been given to show applications of the results concerning the three special cases of generalized binomial distribution (binomial, hypergeometric and Pólya distributions). 000 Mathematics Subject Classification: Primary 60F05. Keywords: A generalized binomial distribution; binomial approximation; binomial distribution; hypergeometric distribution; Pólya distribution; Poisson approximation; Poisson distribution; Stein identity; ω-function.. Introduction In 979, Dwass [5] introduced a generalized discrete distribution that covered the binomial, hypergeometric and PÓlya distributions and he called the distribution that a generalized binomial distribution. The distribution depends on four parameters, A,, n and α, where A and are positive, n is a positive integer,α is an arbitrary real number satisfying (n )α and A (i) and (i) are never negative for i =,..., n, () i where x = x( x α) ( x ( i ) α).
2 4 P. Wongkasem et al. The distribution of the generalized binomial random variable is defined by p( k ) ( k) ( n k) n A =, k = 0,,..., n. ( n) k ( ) n [ AA ( α) ( A ( k ) α)][ ( α) ( ( n k ) α)] =, k ( )( α) ( ( n ) α) (.) k = 0,,..., n. (.) The form of (.) is similar to the distribution form with regard to the Pólya urn model in Johnson et al. [7] on pp. 58. Dwass [5] pointed out the three special cases of the distribution in (.) by α as follows: (i). If α = 0, then the binomial distribution with parameters n and result, that is, A is the k n k n A p( k ) =,k 0,,...,n. k = (.3) (ii). Ifα > 0, then the outcome is the hypergeometric distribution with parameters A,, n andα and, for A α and are integers, the distribution α A α α k n k p( k ) =,k = 0,,..., min{ n,a α} A α + α n (.4) is the well known classical hypergeometric distribution. (iii). If α < 0, then the result of (.) is the Pólya distribution with parameters A,, n and α given by p( k ) A α + k α + n k k n k =,k = 0,,...,n. A α α + n n (.5)
3 Generalized inomial by inomial and Poisson Distributions 5 Note that in the special case where α = the distribution is the beta-binomial distribution with parameters A, and n. If A,, n and α are all positive integers, then it is the same distribution as in rown and Phillips [3], and it is the negative hypergeometric distribution as α =. It is well known that the hypergeometric distribution can be approximated by the binomial distribution and both binomial and hypergeometric distributions can also be approximated by the Poisson distribution under certain conditions on their parameters. Similarly, if the conditions on the parameters of the generalized binomial, binomial and Poisson distributions are satisfied, then the generalized binomial distribution can also be approximated by the binomial distribution or the Poisson distribution. In this paper, the generalized binomial distribution has been approximated by binomial and Poisson distributions, and the accuracy of each approximation is measured in terms of the total variation distance and its upper bound. The total variation distance between generalized binomial and binomial distributions and the total variation distance between generalized binomial and Poisson distributions are defined by and d TV (G(A,,n,α),(n,p)) = sup G( A,,n, α)( E ) ( n, p )( E ) E (.6) d TV (G(A,,n,α),P(λ)) = sup G( A,,n, α)( E ) P( λ)( E ), (.7) E where E is a subset of {0,...,n} and E a subset of {0} and G(A,,n,α), (n, p) and P(λ) are generalized binomial, binomial and Poisson distributions, respectively. The main tools for determining an upper bound for each total variation distance are the ω-function associated with the generalized binomial random variable and the Stein identities for binomial and Poisson distributions. In 998, Majsnerowska [8] adapted the relation of ω-function associated with a non-negative integer-valued random variable X (Cacoullos and Papathanasiou [4]) to be pk ( ) μ ( k+ ) ω( k+ ) = ω( k) 0, k = 0,,..., (.8) pk ( + ) where ω(0) = μ, p(k) > 0 for all k 0 and μ and are mean and variance of X. The recurrence relation (.8) together with the two Stein identities are used to derive two main results in Section. In Section 3, some numerical examples have been given to show applications of the results concerning the binomial, hypergeometric and Pólya distributions which are the special cases of the generalized binomial distribution. The conclusions of this study are presented in the last section.
4 6 P. Wongkasem et al.. Main results This section uses the relation of ω-function in (.8) and the two Stein identities of binomial and Poisson distributions to derive the main results associated with (.6) and (.7). Let X be the generalized binomial random variable with probability distribution defined as in (.). Its mean and variance are μ = and ( nα ) =, respectively, and its associated ω-function is given as the ( ) ( α) following proposition. Proposition. Let ω(x) be the ω-function associated with the generalized binomial random variable X and p(k) > 0 for every 0 k n. Then ( n k)( A kα) ω( k) =, k = 0,,..., n ( ) (.) where ( nα ) =. ( ) ( α) Proof. Following Dwass [5] and (.8), the recurrence relation of ω-function associated with the random variable X can be expressed in the form k[ ( n k) α] k ω( k) = + ω( k ), k =,..., n, ( ) ( n k+ )[ A ( k ) α] where ω(0) =. ( ) Therefore, we have ( n )( A α) ( n )( A α) ( n n)( A nα ) ω() =, ω() =,, ω( n) =, ( ) ( ) ( ) which gives (.). Remark.. If α = 0, then (.) reduces to ( n k) A ω( k) =, k = 0,,..., n ( ) (.) with =, which is the ω-function associated with the binomial random ( ) variable and the distribution in (.3).
5 Generalized inomial by inomial and Poisson Distributions 7. If α > 0 and A α and α are integers, then the new form of ω-function in (.) can be written as ( n k)( Aα k) ω( k) =, k = 0,,...,min, ( Aα+ α) { n A α} (.3) with the same variance as in (.). This is the ω-function associated with the classical hypergeometric random variable and the distribution in (.4). 3. If α < 0, then (.) can be rewritten as ( n k)( Aα + k) ω( k) =, k = 0,,..., n, (.4) ( Aα α) ( nα ) where =. It is the ω-function associated with the Pólya ( ) ( α) random variable and the distribution in (.5). For giving a main result for each of the approximations, the relation of the ω- function is first mentioned, which was stated by Cacoullos and Papathanasiou [4], as follows: If a function g satisfies then where Δg(x) = g(x+) g(x). E ω(x)δg(x) < and = Var( X ) <, Cov(X,g(X)) = E ω(x)δg(x), (.5) y taking g(x) = x, E[ω(X)]= is obtained as a consequence of (.5).. A result of the binomial approximation The theorem below shows an estimate of the total variation distance between G(A,,n,α) and (n,p). To obtain the estimate the Stein identity for the binomial setting (see arbour et al. [] on pp or arbour and Chen [] on pp. 03) is applied for fixed parameters n and p = q (0,), every subset E of {0,...,n} and the bounded real valued function f = f E : {0} (defined as in []) G( A,,n, α)( E ) ( n, p)( E ) = E[( n X) pf( X + ) qxf( X)]. (.6) For any subset E of {0,...,n}, Ehm [6] showed that k, E k, E n+ n+ p q sup Δ f( k ) = sup f( k+ ) f( k ), (n + )pq (.7)
6 8 P. Wongkasem et al. where f(0) = f() and f(k) = f(n) for k n. Theorem. Let X be the generalized binomial random variable and p = A A + > 0. Then the following inequality holds: n A A ( G( A,, n, α), ( n, )) A + A + A + nn ( ) αδ( α), ( n+ )( α) where δ ( α) = if α 0 and δ( α) = if α < 0. Proof. It can be seen that + n+ (.8) E[( n X) pf( X + ) qxf( X)] = E[ npf ( X + ) px Δf ( X ) Xf ( X )] and, by (.6), (.6) and (.7), we have = E[ μ f( X + )] pe[ XΔf( X)] E[ Xf( X)] = E[ μ f( X + )] pe[ XΔ f( X)] Cov( X, f ( X )) E[ μ f ( X )] = E[ μδf( X)] pe[ XΔ f( X)] E[ ω( X) Δf( X)] (by (.5)) { μ ω( ) ( )} = E px X Δf X n+ n+ p q ( G( A,, n, α), ( n, p)) E μ px ω ( X). (.9) ( n+ ) pq y Proposition, ka ( n k)( A kα ) ( n k) kα μ pk ω( k) = =, A + this implies that For α 0, 0 if α 0, μ pk ω( k) 0 if α < 0.
7 Generalized inomial by inomial and Poisson Distributions 9 and E μ px ω( X) E μ px ω( X) Therefore, from both cases, we obtain = μ pe( X) E[ ω( X)] = μq pqn( n ) α =, A + α pqn( n ) α = for α < 0. α E μ px ω( X) = where δ ( α) = if α 0 and δ( α) = if α < 0. pqn( n ) α δ ( α ), α (.0) Substituting the right hand side of (.9) by (.0) and p and q by, respectively. Hence the inequality (.8) holds. A A + and The following two corollaries are consequences of Theorem with regard to approximating the hypergeometric and Pólya distributions by binomial distribution. Corollary. If α > 0 and A α and α are integers, then the following inequality holds: n+ n+ A A ( H( A,, n, α), ( n, )) A + A + A + nn ( ) α. (.) ( n+ )( α) n If additionally p = and A is an integer, then n n n ( H( A,, n, α), ( A, )) A + A + A + AA ( ) α, (.) ( )( α) where H(A,,n,α) is the hypergeometric distribution. Corollary. For α < 0, then A A ( PY( A,, n, α), ( n, )) A + A + A + nn ( )( α ), ( n+ )( α) n+ n+ (.3)
8 0 P. Wongkasem et al. where PY(A,,n,α) is the Pólya distribution. Remark. It should be noted that the result of Theorem or each result of corollaries and, yields a good binomial approximation provided that ( n ) α α is small, that is, α is small and A + α is large.. A result of the Poisson approximation Similar to the binomial approximation, the Stein identity for the Poisson case (see arbour et al. [] on pp. 6-7 or arbour and Chen [] on pp. 65) is applied for every positive constant λ, every subset E of {0} and the bounded real valued function f = f E : {0} (defined as in []). G( A,,n, α)( E) P( λ)( E ) = E[ λf( X + ) Xf( X)]. (.4) For any subset E of {0}, arbour et al. [] proved that Δ f k = f k+ f k λ e λ k, E k, E sup ( ) sup ( ) ( ) ( ). (.5) The following theorem presents a Poisson estimate of the total variation distance between G(A,,n,α) and P(λ) in terms of four parameters A,, n and α. Theorem. Let X be the generalized binomial random variable and 0, and let A (n )( α ) for α < 0. Then λ = A + > ( n ) α + A( A + α) ( G( A,, n, α), P( )) e. (.6) ( )( α) Proof. From (.4), we have d ( G( A,, n, α), P ( λ)) = λe[ f( X + )] E[ Xf( X)] TV = λe[ f( X + )] Cov( X, f( X)) μe[ f( X)] = λe[ Δf( X)] Cov( X, f( X)) = λe[ Δf( X)] E[ ω( X) Δ f( X)] (by (.5)) { λ ω( ) ( )} = E X Δf X ; which by (.5), λ = λ ( e ) E λ ω ( X) (.7)
9 Generalized inomial by inomial and Poisson Distributions thus As, by Proposition, λ ω( k) E λ ω( X) ( n k)( A kα) = [( n k) α + A] k = 0, = λ E[ ω( X)] = λ n ( ) α + AA ( + α) = λ. ( )( α) Substituting this result into (.7) and putting λ =, thus the theorem is proved. Immediately from (.6), additional results of the Poisson approximation are obtained for the three distributions as follows: Corollary 3. For α = 0, A ( G( n, ), P ( )) e. (.8) Corollary 4. For α > 0, if A α and α are integers, then ( n ) α + A( A + α) ( H( A,, n, α), P( )) e. (.9) ( )( α) Corollary 5. For α < 0, if A ( n )( α), then ( n ) α + A( A + α) ( PY( A,, n, α), P( )) e. (.0) ( )( α) A Remark 3.. If and α are small, then the result of (.6) is a good Poisson estimate.. It is noted that, for α > 0 and > 0, n ( ) α + AA ( + α) A n ( )( α) + AA ( + + α ) < <. ( )( α) ( )( + α) Hence, the bound in (.0) is always less than the bounds in (.8) and (.9).
10 P. Wongkasem et al. 3. Numerical examples The following numerical examples are given to illustrate how well binomial and Poisson distributions approximate generalized binomial distribution and to see how tight the upper bound for each total variation distance between two distributions by using the results in theorems and, or corollaries -5. Table : Sample values of each total variation distance ( d TV ) between generalized binomial and binomial distributions and its upper bound (U..) d TV U.. d TV U.. d TV U.. n A (.) (.) (.) (.) (.3) (.3) Table provides sample values of each total variation distance between generalized binomial and binomial distributions and its upper bound for fixed = 000 and α =, corresponding to corollaries and, respectively. Table : Sample values of each total variation distance between generalized binomial and Poisson distributions and its upper bound (U..) d TV U.. d TV U.. d TV U.. n A (.8) (.8) (.9) (.9) (.0) (.0) *******
11 Generalized inomial by inomial and Poisson Distributions 3 For the Poisson approximation, sample values of each total variation distance between generalized binomial and Poisson distributions and its upper bound have been presented for fixed A + = 000 and α = 0,,, corresponding to corollaries 3, 4 and 5, respectively, as seen in Table. The numerical results in the two tables suggest that binomial and Poisson approximation to the generalized binomial distribution are quite efficient when n and/or A are small, that is, the estimate of the total variation distance between two distributions of each result is close to the true value of the distance provided that n and/or A are small and A + α is large. y comparing the numerical results in Table, (.) and (.3), and the results in Table, (.9) and (.0), it indicates that the result in (.) is better than the result in (.9) and in the case where n and A are more different the result in (.3) is also better than the result in (.0). 4. Conclusions In this study the upper bounds in (.8) and (.6) are the estimates of the total variation distance between the generalized binomial and binomial distributions and the total variation distance between the generalized binomial and Poisson distributions, respectively. Furthermore, each upper bound in (.8) and (.6) is also a criterion for measuring the accuracy of the corresponding approximation, that is, if an obtained bound is small, then a good binomial or Poisson approximation to the generalized binomial distribution is obtained. Conversely, if such a bound is large, then the binomial or Poisson distribution is not appropriate to approximate the generalized binomial distribution. From the results of (.8) and (.6), it is found that the upper bound of the binomial approximation is small when α is small and A α is large, and the upper bound of the Poisson is small when both α and are small. Acknowledgements. The first two authors would like to thank Faculty of Science, urapha University, for financial support. References [] arbour, A. D., Holst, L., and Janson, S., 99, Poisson approximation, Oxford Studies in Probability, Clarendon Press, Oxford. [] arbour, A.D. and Chen, L.H.Y., 005, An introduction to Stein s method, Singapore University Press, Singapore. [3] rown, T. C., and Phillips, M. J., 999, Negative binomial approximation with Stein s method, Meth. Comp. Appl. Probab., (4), pp [4] Cacoullos, T., and Papathanasiou, V., 989, Characterization of distributions by variance bounds, Statist. Probab. Lett., 7(5), pp
12 4 P. Wongkasem et al. [5] Dwass, M., 979, A generalized binomial distribution, Amer. Statistician, 33(), pp [6] Ehm, W., 99, inomial approximation to the Poisson binomial distribution, Statist. Probab. Lett., (), pp [7] Johnson, N. L., Kotz, S., and Kemp, A. W., 005, Univariate Discrete Distributions, 3 rd edition, Wiley, New York. [8] Majsnerowska, M., 998, A note on Poisson approximation by ω- functions, Appl. Math., 5(3), pp
A Pointwise Approximation of Generalized Binomial by Poisson Distribution
Applied Mathematical Sciences, Vol. 6, 2012, no. 22, 1095-1104 A Pointwise Approximation of Generalized Binomial by Poisson Distribution K. Teerapabolarn Department of Mathematics, Faculty of Science,
More informationAPPROXIMATION OF GENERALIZED BINOMIAL BY POISSON DISTRIBUTION FUNCTION
International Journal of Pure and Applied Mathematics Volume 86 No. 2 2013, 403-410 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/10.12732/ijpam.v86i2.14
More informationNon Uniform Bounds on Geometric Approximation Via Stein s Method and w-functions
Communications in Statistics Theory and Methods, 40: 45 58, 20 Copyright Taylor & Francis Group, LLC ISSN: 036-0926 print/532-45x online DOI: 0.080/036092090337778 Non Uniform Bounds on Geometric Approximation
More informationA Note on Poisson Approximation for Independent Geometric Random Variables
International Mathematical Forum, 4, 2009, no, 53-535 A Note on Poisson Approximation for Independent Geometric Random Variables K Teerapabolarn Department of Mathematics, Faculty of Science Burapha University,
More informationA Non-uniform Bound on Poisson Approximation in Beta Negative Binomial Distribution
International Mathematical Forum, Vol. 14, 2019, no. 2, 57-67 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/imf.2019.915 A Non-uniform Bound on Poisson Approximation in Beta Negative Binomial Distribution
More informationPoisson Approximation for Independent Geometric Random Variables
International Mathematical Forum, 2, 2007, no. 65, 3211-3218 Poisson Approximation for Independent Geometric Random Variables K. Teerapabolarn 1 and P. Wongasem Department of Mathematics, Faculty of Science
More information1. Introduction. Let the distribution of a non-negative integer-valued random variable X be defined as follows:
International Journal of Pure and Applied Mathematics Volume 82 No. 5 2013, 737-745 ISSN: 1311-8080 (printed version); ISSN: 1314-335 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/.12732/ijpam.v82i5.7
More informationAN IMPROVED POISSON TO APPROXIMATE THE NEGATIVE BINOMIAL DISTRIBUTION
International Journal of Pure and Applied Mathematics Volume 9 No. 3 204, 369-373 ISSN: 3-8080 printed version); ISSN: 34-3395 on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/0.2732/ijpam.v9i3.9
More informationOn bounds in multivariate Poisson approximation
Workshop on New Directions in Stein s Method IMS (NUS, Singapore), May 18-29, 2015 Acknowledgments Thanks Institute for Mathematical Sciences (IMS, NUS), for the invitation, hospitality and accommodation
More informationOn the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method
On the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method Igal Sason Department of Electrical Engineering Technion - Israel Institute of Technology Haifa 32000, Israel ETH, Zurich,
More informationMathematical Statistics 1 Math A 6330
Mathematical Statistics 1 Math A 6330 Chapter 3 Common Families of Distributions Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 28, 2015 Outline 1 Subjects of Lecture 04
More informationON POINTWISE BINOMIAL APPROXIMATION
Iteratioal Joural of Pure ad Applied Mathematics Volume 71 No. 1 2011, 57-66 ON POINTWISE BINOMIAL APPROXIMATION BY w-functions K. Teerapabolar 1, P. Wogkasem 2 Departmet of Mathematics Faculty of Sciece
More informationPoisson approximations
Chapter 9 Poisson approximations 9.1 Overview The Binn, p) can be thought of as the distribution of a sum of independent indicator random variables X 1 + + X n, with {X i = 1} denoting a head on the ith
More informationarxiv: v1 [stat.me] 28 Mar 2011
arxiv:1103.5447v1 [stat.me] 28 Mar 2011 On matrix variance inequalities G. Afendras and N. Papadatos Department of Mathematics, Section of Statistics and O.R., University of Athens, Panepistemiopolis,
More informationStein s method and zero bias transformation: Application to CDO pricing
Stein s method and zero bias transformation: Application to CDO pricing ESILV and Ecole Polytechnique Joint work with N. El Karoui Introduction CDO a portfolio credit derivative containing 100 underlying
More informationChapter 1. Sets and probability. 1.3 Probability space
Random processes - Chapter 1. Sets and probability 1 Random processes Chapter 1. Sets and probability 1.3 Probability space 1.3 Probability space Random processes - Chapter 1. Sets and probability 2 Probability
More informationLIST OF FORMULAS FOR STK1100 AND STK1110
LIST OF FORMULAS FOR STK1100 AND STK1110 (Version of 11. November 2015) 1. Probability Let A, B, A 1, A 2,..., B 1, B 2,... be events, that is, subsets of a sample space Ω. a) Axioms: A probability function
More informationSeries of Error Terms for Rational Approximations of Irrational Numbers
2 3 47 6 23 Journal of Integer Sequences, Vol. 4 20, Article..4 Series of Error Terms for Rational Approximations of Irrational Numbers Carsten Elsner Fachhochschule für die Wirtschaft Hannover Freundallee
More informationOn discrete distributions with gaps having ALM property
ProbStat Forum, Volume 05, April 202, Pages 32 37 ISSN 0974-3235 ProbStat Forum is an e-journal. For details please visit www.probstat.org.in On discrete distributions with gaps having ALM property E.
More informationWeak convergence in Probability Theory A summer excursion! Day 3
BCAM June 2013 1 Weak convergence in Probability Theory A summer excursion! Day 3 Armand M. Makowski ECE & ISR/HyNet University of Maryland at College Park armand@isr.umd.edu BCAM June 2013 2 Day 1: Basic
More informationPower Series Solutions to the Legendre Equation
Power Series Solutions to the Legendre Equation Department of Mathematics IIT Guwahati The Legendre equation The equation (1 x 2 )y 2xy + α(α + 1)y = 0, (1) where α is any real constant, is called Legendre
More informationNotes on Poisson Approximation
Notes on Poisson Approximation A. D. Barbour* Universität Zürich Progress in Stein s Method, Singapore, January 2009 These notes are a supplement to the article Topics in Poisson Approximation, which appeared
More informationMath 324 Summer 2012 Elementary Number Theory Notes on Mathematical Induction
Math 4 Summer 01 Elementary Number Theory Notes on Mathematical Induction Principle of Mathematical Induction Recall the following axiom for the set of integers. Well-Ordering Axiom for the Integers If
More informationWeek 2. Review of Probability, Random Variables and Univariate Distributions
Week 2 Review of Probability, Random Variables and Univariate Distributions Probability Probability Probability Motivation What use is Probability Theory? Probability models Basis for statistical inference
More informationFisher Information, Compound Poisson Approximation, and the Poisson Channel
Fisher Information, Compound Poisson Approximation, and the Poisson Channel Mokshay Madiman Department of Statistics Yale University New Haven CT, USA Email: mokshaymadiman@yaleedu Oliver Johnson Department
More informationFisher information and Stam inequality on a finite group
Fisher information and Stam inequality on a finite group Paolo Gibilisco and Tommaso Isola February, 2008 Abstract We prove a discrete version of Stam inequality for random variables taking values on a
More informationA COMPOUND POISSON APPROXIMATION INEQUALITY
J. Appl. Prob. 43, 282 288 (2006) Printed in Israel Applied Probability Trust 2006 A COMPOUND POISSON APPROXIMATION INEQUALITY EROL A. PEKÖZ, Boston University Abstract We give conditions under which the
More informationarxiv: v1 [math.pr] 16 Jun 2009
A THREE-PARAMETER BINOMIAL APPROXIMATION arxiv:0906.2855v1 [math.pr] 16 Jun 2009 Vydas Čekanavičius, Erol A. Peköz, Adrian Röllin and Michael Shwartz Version from June 3, 2018 Abstract We approximate the
More informationMixtures and Random Sums
Mixtures and Random Sums C. CHATFIELD and C. M. THEOBALD, University of Bath A brief review is given of the terminology used to describe two types of probability distribution, which are often described
More informationCSE 312 Final Review: Section AA
CSE 312 TAs December 8, 2011 General Information General Information Comprehensive Midterm General Information Comprehensive Midterm Heavily weighted toward material after the midterm Pre-Midterm Material
More informationSTEIN S METHOD, SEMICIRCLE DISTRIBUTION, AND REDUCED DECOMPOSITIONS OF THE LONGEST ELEMENT IN THE SYMMETRIC GROUP
STIN S MTHOD, SMICIRCL DISTRIBUTION, AND RDUCD DCOMPOSITIONS OF TH LONGST LMNT IN TH SYMMTRIC GROUP JASON FULMAN AND LARRY GOLDSTIN Abstract. Consider a uniformly chosen random reduced decomposition of
More informationEntropy, Compound Poisson Approximation, Log-Sobolev Inequalities and Measure Concentration
Entropy, Compound Poisson Approximation, Log-Sobolev Inequalities and Measure Concentration Ioannis Kontoyiannis 1 and Mokshay Madiman Division of Applied Mathematics, Brown University 182 George St.,
More information3. DISCRETE RANDOM VARIABLES
IA Probability Lent Term 3 DISCRETE RANDOM VARIABLES 31 Introduction When an experiment is conducted there may be a number of quantities associated with the outcome ω Ω that may be of interest Suppose
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationContinuous Random Variables and Continuous Distributions
Continuous Random Variables and Continuous Distributions Continuous Random Variables and Continuous Distributions Expectation & Variance of Continuous Random Variables ( 5.2) The Uniform Random Variable
More informationIntroduction to Statistical Data Analysis Lecture 3: Probability Distributions
Introduction to Statistical Data Analysis Lecture 3: Probability Distributions James V. Lambers Department of Mathematics The University of Southern Mississippi James V. Lambers Statistical Data Analysis
More information1: PROBABILITY REVIEW
1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following
More informationOn the Poisson Approximation to the Negative Hypergeometric Distribution
BULLETIN of the Malaysian Mathematical Sciences Society http://mathusmmy/bulletin Bull Malays Math Sci Soc (2) 34(2) (2011), 331 336 On the Poisson Appoximation to the Negative Hypegeometic Distibution
More informationTopic 9 Examples of Mass Functions and Densities
Topic 9 Examples of Mass Functions and Densities Discrete Random Variables 1 / 12 Outline Bernoulli Binomial Negative Binomial Poisson Hypergeometric 2 / 12 Introduction Write f X (x θ) = P θ {X = x} for
More informationSTAT/MATH 395 PROBABILITY II
STAT/MATH 395 PROBABILITY II Chapter 6 : Moment Functions Néhémy Lim 1 1 Department of Statistics, University of Washington, USA Winter Quarter 2016 of Common Distributions Outline 1 2 3 of Common Distributions
More informationThings to remember when learning probability distributions:
SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationSuppose that you have three coins. Coin A is fair, coin B shows heads with probability 0.6 and coin C shows heads with probability 0.8.
Suppose that you have three coins. Coin A is fair, coin B shows heads with probability 0.6 and coin C shows heads with probability 0.8. Coin A is flipped until a head appears, then coin B is flipped until
More informationExam 3, Math Fall 2016 October 19, 2016
Exam 3, Math 500- Fall 06 October 9, 06 This is a 50-minute exam. You may use your textbook, as well as a calculator, but your work must be completely yours. The exam is made of 5 questions in 5 pages,
More informationProbability Distributions Columns (a) through (d)
Discrete Probability Distributions Columns (a) through (d) Probability Mass Distribution Description Notes Notation or Density Function --------------------(PMF or PDF)-------------------- (a) (b) (c)
More information5. Conditional Distributions
1 of 12 7/16/2009 5:36 AM Virtual Laboratories > 3. Distributions > 1 2 3 4 5 6 7 8 5. Conditional Distributions Basic Theory As usual, we start with a random experiment with probability measure P on an
More informationPower Series Solutions to the Legendre Equation
Department of Mathematics IIT Guwahati The Legendre equation The equation (1 x 2 )y 2xy + α(α + 1)y = 0, (1) where α is any real constant, is called Legendre s equation. When α Z +, the equation has polynomial
More informationMultivariate distributions
CHAPTER Multivariate distributions.. Introduction We want to discuss collections of random variables (X, X,..., X n ), which are known as random vectors. In the discrete case, we can define the density
More informationTwo finite forms of Watson s quintuple product identity and matrix inversion
Two finite forms of Watson s uintuple product identity and matrix inversion X. Ma Department of Mathematics SuZhou University, SuZhou 215006, P.R.China Submitted: Jan 24, 2006; Accepted: May 27, 2006;
More informationSTAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it
More informationClassical Probability
Chapter 1 Classical Probability Probability is the very guide of life. Marcus Thullius Cicero The original development of probability theory took place during the seventeenth through nineteenth centuries.
More informationp. 4-1 Random Variables
Random Variables A Motivating Example Experiment: Sample k students without replacement from the population of all n students (labeled as 1, 2,, n, respectively) in our class. = {all combinations} = {{i
More information2 Chance constrained programming
2 Chance constrained programming In this Chapter we give a brief introduction to chance constrained programming. The goals are to motivate the subject and to give the reader an idea of the related difficulties.
More informationTotal variation error bounds for geometric approximation
Bernoulli 192, 2013, 610 632 DOI: 10.3150/11-BEJ406 Total variation error bounds for geometric approximation EROL A. PEKÖZ 1, ADRIAN RÖLLIN 2 and NATHAN ROSS 3 1 School of Management, Boston University,
More informationClinical Trial Design Using A Stopped Negative Binomial Distribution. Michelle DeVeaux and Michael J. Kane and Daniel Zelterman
Statistics and Its Interface Volume 0 (0000) 1 10 Clinical Trial Design Using A Stopped Negative Binomial Distribution Michelle DeVeaux and Michael J. Kane and Daniel Zelterman arxiv:1508.01264v8 [math.st]
More informationTopic 3: The Expectation of a Random Variable
Topic 3: The Expectation of a Random Variable Course 003, 2017 Page 0 Expectation of a discrete random variable Definition (Expectation of a discrete r.v.): The expected value (also called the expectation
More informationMatchings in hypergraphs of large minimum degree
Matchings in hypergraphs of large minimum degree Daniela Kühn Deryk Osthus Abstract It is well known that every bipartite graph with vertex classes of size n whose minimum degree is at least n/2 contains
More informationMATH 151, FINAL EXAM Winter Quarter, 21 March, 2014
Time: 3 hours, 8:3-11:3 Instructions: MATH 151, FINAL EXAM Winter Quarter, 21 March, 214 (1) Write your name in blue-book provided and sign that you agree to abide by the honor code. (2) The exam consists
More informationComplete moment convergence of weighted sums for processes under asymptotically almost negatively associated assumptions
Proc. Indian Acad. Sci. Math. Sci.) Vol. 124, No. 2, May 214, pp. 267 279. c Indian Academy of Sciences Complete moment convergence of weighted sums for processes under asymptotically almost negatively
More informationProbability. Computer Science Tripos, Part IA. R.J. Gibbens. Computer Laboratory University of Cambridge. Easter Term 2008/9
Probability Computer Science Tripos, Part IA R.J. Gibbens Computer Laboratory University of Cambridge Easter Term 2008/9 Last revision: 2009-05-06/r-36 1 Outline Elementary probability theory (2 lectures)
More informationPCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities
PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets
More informationLecture 3. Discrete Random Variables
Math 408 - Mathematical Statistics Lecture 3. Discrete Random Variables January 23, 2013 Konstantin Zuev (USC) Math 408, Lecture 3 January 23, 2013 1 / 14 Agenda Random Variable: Motivation and Definition
More informationNorthwestern University Department of Electrical Engineering and Computer Science
Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability
More informationIndependence of some multiple Poisson stochastic integrals with variable-sign kernels
Independence of some multiple Poisson stochastic integrals with variable-sign kernels Nicolas Privault Division of Mathematical Sciences School of Physical and Mathematical Sciences Nanyang Technological
More informationConditional distributions (discrete case)
Conditional distributions (discrete case) The basic idea behind conditional distributions is simple: Suppose (XY) is a jointly-distributed random vector with a discrete joint distribution. Then we can
More informationProbability reminders
CS246 Winter 204 Mining Massive Data Sets Probability reminders Sammy El Ghazzal selghazz@stanfordedu Disclaimer These notes may contain typos, mistakes or confusing points Please contact the author so
More informationOn monotonicity of expected values of some run-related distributions
Ann Inst Stat Math (2016) 68:1055 1072 DOI 10.1007/s10463-015-0525-x On monotonicity of expected values of some run-related distributions Sigeo Aki 1 Katuomi Hirano 2 Received: 9 May 2014 / Revised: 11
More informationMATH 324 Summer 2011 Elementary Number Theory. Notes on Mathematical Induction. Recall the following axiom for the set of integers.
MATH 4 Summer 011 Elementary Number Theory Notes on Mathematical Induction Principle of Mathematical Induction Recall the following axiom for the set of integers. Well-Ordering Axiom for the Integers If
More informationPseudo-Boolean Functions, Lovász Extensions, and Beta Distributions
Pseudo-Boolean Functions, Lovász Extensions, and Beta Distributions Guoli Ding R. F. Lax Department of Mathematics, LSU, Baton Rouge, LA 783 Abstract Let f : {,1} n R be a pseudo-boolean function and let
More information1 Review of Probability and Distributions
Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote
More informationPACKING-DIMENSION PROFILES AND FRACTIONAL BROWNIAN MOTION
PACKING-DIMENSION PROFILES AND FRACTIONAL BROWNIAN MOTION DAVAR KHOSHNEVISAN AND YIMIN XIAO Abstract. In order to compute the packing dimension of orthogonal projections Falconer and Howroyd 997) introduced
More informationA Short Introduction to Stein s Method
A Short Introduction to Stein s Method Gesine Reinert Department of Statistics University of Oxford 1 Overview Lecture 1: focusses mainly on normal approximation Lecture 2: other approximations 2 1. The
More informationPAijpam.eu REPEATED RANDOM ALLOCATIONS WITH INCOMPLETE INFORMATION Vladimir G. Panov 1, Julia V. Nagrebetskaya 2
International Journal of Pure and Applied Mathematics Volume 118 No. 4 2018, 1021-1032 ISSN: 1311-8080 printed version; ISSN: 1314-3395 on-line version url: http://www.ipam.eu doi: 10.12732/ipam.v118i4.16
More informationIn N we can do addition, but in order to do subtraction we need to extend N to the integers
Chapter The Real Numbers.. Some Preliminaries Discussion: The Irrationality of 2. We begin with the natural numbers N = {, 2, 3, }. In N we can do addition, but in order to do subtraction we need to extend
More informationWittmann Type Strong Laws of Large Numbers for Blockwise m-negatively Associated Random Variables
Journal of Mathematical Research with Applications Mar., 206, Vol. 36, No. 2, pp. 239 246 DOI:0.3770/j.issn:2095-265.206.02.03 Http://jmre.dlut.edu.cn Wittmann Type Strong Laws of Large Numbers for Blockwise
More information18.175: Lecture 17 Poisson random variables
18.175: Lecture 17 Poisson random variables Scott Sheffield MIT 1 Outline More on random walks and local CLT Poisson random variable convergence Extend CLT idea to stable random variables 2 Outline More
More informationChap 2.1 : Random Variables
Chap 2.1 : Random Variables Let Ω be sample space of a probability model, and X a function that maps every ξ Ω, toa unique point x R, the set of real numbers. Since the outcome ξ is not certain, so is
More informationResearch Article On the Isolated Vertices and Connectivity in Random Intersection Graphs
International Cobinatorics Volue 2011, Article ID 872703, 9 pages doi:10.1155/2011/872703 Research Article On the Isolated Vertices and Connectivity in Rando Intersection Graphs Yilun Shang Institute for
More informationTHE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION
THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION JAINUL VAGHASIA Contents. Introduction. Notations 3. Background in Probability Theory 3.. Expectation and Variance 3.. Convergence
More informationRandom Variables. P(x) = P[X(e)] = P(e). (1)
Random Variables Random variable (discrete or continuous) is used to derive the output statistical properties of a system whose input is a random variable or random in nature. Definition Consider an experiment
More informationON THE MOMENTS OF ITERATED TAIL
ON THE MOMENTS OF ITERATED TAIL RADU PĂLTĂNEA and GHEORGHIŢĂ ZBĂGANU The classical distribution in ruins theory has the property that the sequence of the first moment of the iterated tails is convergent
More informationINDUCTION AND RECURSION. Lecture 7 - Ch. 4
INDUCTION AND RECURSION Lecture 7 - Ch. 4 4. Introduction Any mathematical statements assert that a property is true for all positive integers Examples: for every positive integer n: n!
More informationA Few Special Distributions and Their Properties
A Few Special Distributions and Their Properties Econ 690 Purdue University Justin L. Tobias (Purdue) Distributional Catalog 1 / 20 Special Distributions and Their Associated Properties 1 Uniform Distribution
More informationUnivariate Discrete Distributions
Univariate Discrete Distributions Second Edition NORMAN L. JOHNSON University of North Carolina Chapel Hill, North Carolina SAMUEL KOTZ University of Maryland College Park, Maryland ADRIENNE W. KEMP University
More informationChapter 4 Multiple Random Variables
Review for the previous lecture Theorems and Examples: How to obtain the pmf (pdf) of U = g ( X Y 1 ) and V = g ( X Y) Chapter 4 Multiple Random Variables Chapter 43 Bivariate Transformations Continuous
More informationThe first divisible sum
The first divisible sum A.D. Barbour and R. Grübel Universität Zürich and Technische Universiteit Delft now: "Universität-Gesamthochschule-Paderborn" Abstract. We consider the distribution of the first
More informationYi Wang Department of Applied Mathematics, Dalian University of Technology, Dalian , China (Submitted June 2002)
SELF-INVERSE SEQUENCES RELATED TO A BINOMIAL INVERSE PAIR Yi Wang Department of Applied Mathematics, Dalian University of Technology, Dalian 116024, China (Submitted June 2002) 1 INTRODUCTION Pairs of
More informationChiang Mai J. Sci. 2016; 43(3) : Contributed Paper
Chiang Mai J Sci 06; 43(3) : 67-68 http://epgsciencecmuacth/ejournal/ Contributed Paper Upper Bounds of Generalized p-values for Testing the Coefficients of Variation of Lognormal Distributions Rada Somkhuean,
More informationDiscrete Distributions
A simplest example of random experiment is a coin-tossing, formally called Bernoulli trial. It happens to be the case that many useful distributions are built upon this simplest form of experiment, whose
More informationarxiv: v1 [math.nt] 8 Jan 2014
A NOTE ON p-adic VALUATIONS OF THE SCHENKER SUMS PIOTR MISKA arxiv:1401.1717v1 [math.nt] 8 Jan 2014 Abstract. A prime number p is called a Schenker prime if there exists such n N + that p nandp a n, wherea
More information1 Random Variable: Topics
Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?
More informationSharp threshold functions for random intersection graphs via a coupling method.
Sharp threshold functions for random intersection graphs via a coupling method. Katarzyna Rybarczyk Faculty of Mathematics and Computer Science, Adam Mickiewicz University, 60 769 Poznań, Poland kryba@amu.edu.pl
More informationA Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.
A Probability Primer A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes. Are you holding all the cards?? Random Events A random event, E,
More informationFactorization of integer-valued polynomials with square-free denominator
accepted by Comm. Algebra (2013) Factorization of integer-valued polynomials with square-free denominator Giulio Peruginelli September 9, 2013 Dedicated to Marco Fontana on the occasion of his 65th birthday
More informationThe Binomial distribution. Probability theory 2. Example. The Binomial distribution
Probability theory Tron Anders Moger September th 7 The Binomial distribution Bernoulli distribution: One experiment X i with two possible outcomes, probability of success P. If the experiment is repeated
More informationFundamental Tools - Probability Theory II
Fundamental Tools - Probability Theory II MSc Financial Mathematics The University of Warwick September 29, 2015 MSc Financial Mathematics Fundamental Tools - Probability Theory II 1 / 22 Measurable random
More informationSusceptible-Infective-Removed Epidemics and Erdős-Rényi random
Susceptible-Infective-Removed Epidemics and Erdős-Rényi random graphs MSR-Inria Joint Centre October 13, 2015 SIR epidemics: the Reed-Frost model Individuals i [n] when infected, attempt to infect all
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationM378K In-Class Assignment #1
The following problems are a review of M6K. M7K In-Class Assignment # Problem.. Complete the definition of mutual exclusivity of events below: Events A, B Ω are said to be mutually exclusive if A B =.
More informationWeak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria
Weak Leiden University Amsterdam, 13 November 2013 Outline 1 2 3 4 5 6 7 Definition Definition Let µ, µ 1, µ 2,... be probability measures on (R, B). It is said that µ n converges weakly to µ, and we then
More information