Department of Mathematics

Similar documents
Department of Mathematics

Department of Mathematics

Things to remember when learning probability distributions:

Introducing the Normal Distribution

Introducing the Normal Distribution

The Poisson Arrival Process

1 Review of Probability and Distributions

Introduction to Probability and Statistics Winter 2015

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

Probability and Distributions

Stat 5101 Notes: Brand Name Distributions

Chapter 5. Chapter 5 sections

Probability Distributions Columns (a) through (d)

Department of Mathematics

Mathematical Statistics 1 Math A 6330

Department of Mathematics

Continuous Probability Spaces

Sampling Distributions of Statistics Corresponds to Chapter 5 of Tamhane and Dunlop

Random variables and expectation

Random Variables and Their Distributions

Random variables and expectation

Northwestern University Department of Electrical Engineering and Computer Science

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Week 1 Quantitative Analysis of Financial Markets Distributions A

7.3 The Chi-square, F and t-distributions

15 Discrete Distributions

Common probability distributionsi Math 217 Probability and Statistics Prof. D. Joyce, Fall 2014

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

MATH Notebook 5 Fall 2018/2019

Chapter 1. Sets and probability. 1.3 Probability space

S n = x + X 1 + X X n.

Exercises and Answers to Chapter 1

Lecture 22: A Review of Linear Algebra and an Introduction to The Multivariate Normal Distribution

Guidelines for Solving Probability Problems

Special distributions

Expectation is a positive linear operator

Continuous Distributions

Lecture 2. Distributions and Random Variables

BMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Discrete Distributions

Common ontinuous random variables

Order Statistics; Conditional Expectation

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

Discrete Distributions Chapter 6

3 Modeling Process Quality

Statistical Methods in Particle Physics

3 Continuous Random Variables

Probability Distributions - Lecture 5

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

PAS04 - Important discrete and continuous distributions

Chapter 3 Common Families of Distributions

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Review for the previous lecture

Example 1. Assume that X follows the normal distribution N(2, 2 2 ). Estimate the probabilities: (a) P (X 3); (b) P (X 1); (c) P (1 X 3).

Stat 5101 Notes: Brand Name Distributions

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment:

Statistical distributions: Synopsis

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x!

ECE 313 Probability with Engineering Applications Fall 2000

Chapter 2. Discrete Distributions

1 Presessional Probability

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Statistics 1B. Statistics 1B 1 (1 1)

1 Probability and Random Variables

Statistics 3657 : Moment Generating Functions

Mathematical Statistics

A Few Special Distributions and Their Properties

Brief Review of Probability

Continuous-time Markov Chains

Probability distributions. Probability Distribution Functions. Probability distributions (contd.) Binomial distribution

Chapter 5 continued. Chapter 5 sections

18.440: Lecture 28 Lectures Review

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Statistics and data analyses

1 Review of Probability

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

MFM Practitioner Module: Quantitative Risk Management. John Dodson. September 23, 2015

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Ching-Han Hsu, BMES, National Tsing Hua University c 2015 by Ching-Han Hsu, Ph.D., BMIR Lab. = a + b 2. b a. x a b a = 12

ACM 116: Lectures 3 4

MAS113 Introduction to Probability and Statistics

Name: Firas Rassoul-Agha

2 Random Variable Generation

MATH Solutions to Probability Exercises

Continuous Random Variables

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

APM 504: Probability Notes. Jay Taylor Spring Jay Taylor (ASU) APM 504 Fall / 65

MATH : EXAM 2 INFO/LOGISTICS/ADVICE

Math Stochastic Processes & Simulation. Davar Khoshnevisan University of Utah

Continuous Random Variables and Continuous Distributions

Chapter 3.3 Continuous distributions

Math/Stats 425, Sec. 1, Fall 04: Introduction to Probability. Final Exam: Solutions

Monotonicity and Aging Properties of Random Sums

Lecture 17: The Exponential and Some Related Distributions

Tom Salisbury

THE QUEEN S UNIVERSITY OF BELFAST

Data, Estimation and Inference

Chapter 4 Multiple Random Variables

Transcription:

Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Supplement 2: Review Your Distributions Relevant textbook passages: Pitman [10]: pages 476 487. Larsen Marx [9]: Chapter 4 A terrific reference for the familiar distributions is the compact monograph by Forbes, Evans, Hastings, and Peacock [4]. There are also the classics by Johnson and Kotz [6, 7, 8]. S2.1 Bernoulli The Bernoulli distribution is a discrete distribution that generalizes coin tossing. A Bernoulli(p) random variable X takes on two values: 1 ( success ), with probability p, and 0 ( failure ), with probability 1 p. The probability mass function is { p x = 1, p (X = x) = 1 p x = 0. The Bernoulli(p) has mean Moreover so the variance is E X = x=0,1 E X 2 = x=0,1 Note that the variance is maximized for xp(x = x) = 0(1 p) + 1p = p. x 2 p(x = x) = 0 2 (1 p) + 1 2 p = p, Var X = E(X 2 ) (E X) 2 = p p 2 = p(1 p). Also note that every moment is the same: p = 1/2. E X α = 0 α (1 p) + 1 α p = p. S2 1

KC Border Review Your Distributions S2 2 ( / ) 0.0 ( / ) E E E E E E - -0.5 0.5 1.5 2.0 S2.2 Binomial The Binomial(n, p) is the distribution of the number X of successes in n independent Bernoulli(p) trials. It is the sum of n independent Bernoulli(p) random variables. v. 2017.01.25::17.49 KC Border

KC Border Review Your Distributions S2 3 The probability mass function is P (X = k) = ( ) n p k (1 p) n k. k Since expectation is a linear operator, the expectation of a Binomial is the sum of the expectations of the underlying Bernoullis, so if X has a Binomial(n, p) distribution, E X = np, and since the variance of the sum of independent random variables is the sum of the variances, Var X = np(1 p). 0.15 ( = ) p= p=0.5 p= 0.10 0.05 10 20 30 40 KC Border v. 2017.01.25::17.49

KC Border Review Your Distributions S2 4 p= p=0.5 p= ( = ) 10 20 30 40 Larsen Marx [9]: Section 1, pp. 494 499 Pitman [10]: p. 155 S2.3 The Multinomial Distribution In one sense, the Multinomial distribution generalizes the binomial distribution to independent random experiments with more than two outcomes. It is the distribution of a vector that counts how many times each outcome occurs. A Multinomial random vector is an m-vector X of counts of outcomes in a sequence of n independent repetitions of a random experiment with m distinct outcomes. If the experiment has m possible outcomes, and the i th outcome has probability p i, then the Multinomial(n, p) probability mass function is given by where k 1 + + k m = n. P (X = (k 1,..., k m )) = n! k 1! k 2! k m! pk 1 1 pk 2 2 pk m m. If we count outcome k as a success, then it is obvious that each X k is simply a Binomial(n, p k ) random variable. But the components are not independent, since they sum to n. v. 2017.01.25::17.49 KC Border

KC Border Review Your Distributions S2 5 Pitman [10]: p. 213 Larsen Marx [9]: 4.5 S2.4 The Negative Binomial Distribution Replicate a Bernoulli(p) experiment independently until the r th success occurs (r 1). Let X be the number of the trial on which the r th success occurs. Then X is said to have a Negative Binomial(r, p) Distribution. Warning: There is another definition of the Negative Binomial Distribution, namely the distribution of the number of failures before the r th success. (This is the definition employed by Jacod and Protter [5, p. 31], R, and Mathematica.) The relationship between the two is quite simple. If X is a negative binomial in the Pitman Larsen Marx sense, and F is negative binomial in the R Mathematica sense, then F = X r. A simple way to tell which definition your software is using is to ask it the probability that the rv equals zero. For PLM, it is always 0, but for RM, it is p r. What is the probability that the r th success occurs on trial t, for t r? For this to happen, there must be t r failures and r 1 successes in the first t 1 trials, with a success on trial t. By independence, this happens with the binomial probability for r 1 successes on t 1 trials times the probability p of success on trial t: P (X = t) = ( ) t 1 p r (1 p) t r r 1 (t r). Of course, the probability is 0 for t < r. The special case r = 1 (number of trials to the first success) is called the Geometric Distribution(p). The mean of the Geometric Distribution is reasonably straightforward: E X = tp(1 p) t 1 = 1 p. t=1 (See my notes on sums, paying attention to the fact that I deviously switched the roles of p and 1 p.) Moreover, in Section S1.4 we showed Var X = E(X 2 ) (E X) 2 = 1 p p 2. Now observe that a Negative Binomial(r, p) is really the sum of r independent Geometric(p) random variables. (Wait for the first success, start over, wait for the next success,, stop after r successes.) Now we use the fact that expectation is a positive linear operator to conclude that the mean of Negative Binomial(r, p) is r times the mean the Geometric(p), and the variance of an independent sum is the sum of the variances, so If X Negative Binomial(r, p), then E X = r p and Var X = r(1 p) p 2. KC Border v. 2017.01.25::17.49

KC Border Review Your Distributions S2 6 S2.5 Rademacher The Rademacher(p) distribution is a recoding of the Bernoulli distribution, 1 still indicates success, but failure is coded as 1. If Y is a Bernoulli(p) random variable, then X = 2Y 1 is a Rademacher(p) random variable. The probability mass function is P (X = x) = A Rademacher(p) random variable X has mean E X = x=0,1 { p x = 1, 1 p x = 1. xp(x = x) = 1(1 p) + 1p = 2p 1. Moreover X 2 = 1 so so the variance is E(X 2 ) = 1, Var X = E(X 2 ) (E X) 2 = 1 (2p 1) 2 = 4p(1 p). A sequence of successive sums of independent Rademacher(p) random variables is called a random walk. That is, if X i are iid Rademacher(1/2) random variables, the sequence S 1, S 2,... is a random walk, where S n = X 1 + + X n. Since expectation is a linear operator, E S n = 0 for all n, and since the variance of a sum of independent random variables is the sum of the variances, Var S n = n for all n. S2.6 Poisson(µ) A Poisson(µ) random variable N models the count of successes when the probability of success is small, but the number of independent trials is large, so that the average success rate is µ. I will explain the above comment soon. distribution Here is the formal definition of the Poisson(µ) v. 2017.01.25::17.49 KC Border

KC Border Review Your Distributions S2 7 The Poisson(µ) probability mass function is So P (N = k) = e µ µk, k = 0, 1,... k! E N = µ, and Var N = µ. 0.15 μ=5 μ=10 μ=20 μ=40 0.10 0.05 10 20 30 40 50 60 λ=5 λ=10 λ=20 λ=40 10 20 30 40 50 60 KC Border v. 2017.01.25::17.49

KC Border Review Your Distributions S2 8 Ladislaus von Bortkiewicz [11] referred to the Poisson distribution as The Law of Small Numbers. It can be thought of as a peculiar limit of Binomial distributions. Consider a sequence of Binomial(n, p) random variables, where the probability p of success is going to zero, but the number n of trials is growing in such a way that np = µ remains fixed. Then we have the following. S2.6.1 Poisson s Limit Theorem For every k, lim Binomial(n, µ/n)(k) = Poisson(µ)(k). n Proof : Fix n and p, and let µ = np to rewrite the Binomial probability of k successes in n trials as n! k!(n k)! pk (1 p) n k = = µk k! = µk k! = µk k! n(n 1) (n k + 1) k! (pn) k n k (1 p)n (1 p) k n n 1 n n n k + 1 (1 p) n (1 p) k n ( 1 µ ) n n 1 n n n k + 1 ( 1 µ ) k n n ( 1 µ ) n n 1 n n µ n k + 1 n µ. Now consider what happens when ( we allow n to grow but keep µ fixed by setting p n = µ/n. It is well known 5 that lim n 1 µ n n) = e µ, and each of the last k terms has limit 1 as n. Since k is held fixed, the product of the last k terms converges to 1. So in the limit as n, the binomial probability of k successes, when the probability of success is µ/n, is just µ k k! e µ. 5 In fact, for any x, we have ( 1 + n) x n ( )) = (e ln 1+ x n ( ) n = e n ln 1+ x n ( so we first find lim n n ln 1 + n) x. Note that even if x < 0, for large enough n we will have 1 + x > 0, so the n logarithm is defined. Using Taylor s Theorem we get ln ( 1 + x n) = ln(1) + ln (1) x n + R(n) = x n + R(n), where the remainder R(n) satisfies R(n)/( x ) = nr(n)/x 0. Thus n ( n ) ( lim n ln 1 + x n n = lim x n n n + R(n)) = x + lim nr(n) = x. n Since the exponential function is continuous, lim ln(1 + x n n )n = (1+ lim e n ln x n n ) ( ) = e lim n n ln 1+ x n = e x. v. 2017.01.25::17.49 KC Border

KC Border Review Your Distributions S2 9 0.08 ( / ) ( ) Binomial Poisson 0.06 0.04 0.02 10 20 30 40 50 60 0.08 ( / ) ( ) Binomial Poisson 0.06 0.04 0.02 10 20 30 40 50 60 S2.7 The Normal family According to the Central Limit Theorem, the limiting distribution of the standardized sum a large number of independent random variables, each with negligible variance relative to the total, is a Normal distribution. KC Border v. 2017.01.25::17.49

KC Border Review Your Distributions S2 10 The N(µ, σ 2 ) density is f (µ,σ2 )(x) = 1 2πσ e (x µ)2 2σ 2. The mean of an N(µ, σ 2 ) random variable X E X = µ and variance Var X = σ 2. The cdf has no closed form and is simply denoted Φ (µ,σ 2 ). S2.7.1 The Standard Normal The standard normal has µ = 0 and σ 2 = 1, so the density is It has mean and variance f(x) = 1 2π e x2 2. E X = 0 Var X = 1. The cdf has no closed form and is simply denoted Φ. Φ(t) = 1 2π t e x2 2 dx. S2.7.2 The Normal Family If Z is a standard normal, then σz + µ N(µ, σ 2 ) and if X N(µ, σ 2 x µ ), then N(0, 1). σ If X and Z are independent normals, then ax + bz is also normal. (μ = σ = ) 0.5 0.3 0.1-3 -2-1 0 1 2 3 5 N.B. Mathematica and R parameterize the Normal by its mean and standard deviation, not its variance. v. 2017.01.25::17.49 KC Border

KC Border Review Your Distributions S2 11 (μ = σ = ) -3-2 -1 1 2 3 (μ = ) σ=0.5 σ= σ=1.5-3 -2-1 1 2 3 (μ = ) σ=0.5 σ= σ=1.5-3 -2-1 1 2 3 S2.8 Binomial and Normal S2.8.1 DeMoivre Laplace Limit Theorem Let X be Binomial(n, p) random variable. Its standardization is (X np)/ np(1 p). For any real numbers a, b, ( ) lim P a X np b = 1 b z z2 /2 dx. n np(1 p) 2π In practice, this approximation requires that n max{p/(1 p)p, (1 p)/p}. a KC Border v. 2017.01.25::17.49

KC Border Review Your Distributions S2 12 ( ) ( ) Binomial Normal 0.3 0.1-4 -2 2 4 S2.9 Exponential The Exponential family is used to model waiting times, where what we are waiting for is typically called an arrival, or a failure, or death. Exponential(λ) has density and cdf and survival function f(t) = λe λt F (t) = 1 e λt G(t) = 1 F (t) = e λt. It has a constant hazard rate f(t)/g(t) equal to λ. The mean of the Exponential(λ) is E T = 1 λ and the variance is Var T = 1 λ 2. The Exponential is memoryless, that is, P ( T < t + s T > t ) = P (T > s). v. 2017.01.25::17.49 KC Border

KC Border Review Your Distributions S2 13 2.0 λ=0.5 λ=1 λ=2 1.5 0.5 0.5 1.5 2.0 2.5 3.0 λ=0.5 λ=1 λ=2 0.5 1.5 2.0 2.5 3.0 S2.10 Gamma The Gamma(r, λ) family of distributions is a versatile one. The distribution of the sum of r independent Exponential(λ) random variables, the distribution of the r th arrival time in a Poisson process with arrival rate λ is the Gamma(r, λ) distribution. The distribution of the sum of squares of n independent standard normals the χ 2 (n) distribution is the Gamma(n/2, 1/2) distribution. KC Border v. 2017.01.25::17.49

KC Border Review Your Distributions S2 14 The general Gamma(r, λ) distribution (r > 0, λ > 0) has density given by f(t) = λr Γ(r) tr 1 e λt (t > 0). The parameter r is the shape parameter, and λ is the scale parameter. S2.10.1 The Γ function The Gamma function is defined by Γ(r) = 0 t r 1 e t dt. It satisfies Γ(r + 1) = rγ(r) and Γ(n) = (n 1)!. The mean and variance of a Gamma(r, λ) random variable are given by E X = r λ, Var X = r λ 2. Unfortunately there are two definitions of the Gamma distribution in use! The one I just gave is the one used by Pitman [10], Feller [3], Cramér [2], and Larsen and Marx [9]. The other definition replaces λ by 1/λ in the density, and is used by Casella and Berger [1], and also by R and Mathematica. 0.5 Λ Λ 0.5 Λ 1 Λ 2 0.3 0.1 2 4 6 8 v. 2017.01.25::17.49 KC Border

KC Border Review Your Distributions S2 15 Λ Λ 0.5 Λ 1 Λ 2 2 4 6 8 0.5 ( ) r = 0.5 r = 1 r = 3 r = 5 0.3 0.1 2 4 6 8 KC Border v. 2017.01.25::17.49

KC Border Review Your Distributions S2 16 ( ) r = 0.5 r = 1 r = 3 r = 5 2 4 6 8 S2.11 Cauchy If X and Y are independent standard normals, then Y/X has a Cauchy distribution. The Cauchy distribution is also the distribution of the tangent of an angle randomly selected from [ π, π]. If C is a Cauchy random variable, then so is 1/C. The density is and the cdf is f(x) = 1 π(1 + x 2 ) F (t) = 1 π arctan(t) + 1 2. S2.11.1 so The expectation of the Cauchy does not exist! 0 t t 0 x π(1 + x 2 ) dx = ln(1 + t2 ) 2π t x π(1 + x 2 ) dx = ln(1 + t2 ) 2π t x π(1 + x 2 dx is divergent and meaningless. ) v. 2017.01.25::17.49 KC Border

KC Border Review Your Distributions S2 17 0.5 0.3 0.1-3 -2-1 0 1 2 3-3 -2-1 0 1 2 3 ( / ) 0.3 Normal Cauchy 0.1-4 -2 0 2 4 KC Border v. 2017.01.25::17.49

Ma 3/103 KC Border Winter 2017 S2 18 Review Your Distributions ( / ) Normal Cauchy -4 0-2 2 4 The Law of Large Numbers versus the Cauchy Distribution 4 2 2 109 4 109 6 109 8 109 1 1010-2 -4 This depicts a simulation of Sn /n, where Sn is the sum of n simulated Cauchy random variables, for n = 1,..., 10, 000, 000, 000. Of course, no computer can accurately simulate a Cauchy random variable. S2.12 Uniform { Density: f (x) 0 { CDF: F (x) EX = v. 2017.01.25::17.49 a+b, 2 1 b a x a b a 0 x [a, b] otherwise. x [a, b] otherwise. Var X = (b a)2 12 KC Border

KC Border Review Your Distributions S2 19 [ ] -0.5 0.5 1.5 [ ] -0.5 0.5 1.5 S2.13 Beta The k th order statistic of a sample of n independent U[0, 1] random variables has a Beta(k, n k + 1) distribution. where The density is f(x) = { 1 β(r,s) xr 1 (1 x) s 1 x [0, 1] 0 otherwise β(r, s) = Γ(r)Γ(s) Γ(r + s), where the Gamma function satisfies Γ(n) = (n 1)! and Γ(s + 1) = sγ(s) for s > 0. It has mean and variance given by E X = r r + s Var X = rs (r + s) 2 (1 + r + s). KC Border v. 2017.01.25::17.49

KC Border Review Your Distributions S2 20 3.5 ( ) 3.0 2.5 2.0 s=1/4 s=2 s=4 1.5 0.5 0.0 ( ) r=1/4 s=2 s=4 3.5 ( ) 3.0 2.5 2.0 r=1/4 r=2 r=4 1.5 0.5 0.0 v. 2017.01.25::17.49 KC Border

KC Border Review Your Distributions S2 21 ( ) r=1/4 r=2 r=4 S2.14 The χ 2 (n) distribution Let Z 1,..., Z n be independent standard normals. The chi-square distribution with n degrees of freedom, χ 2 (n), is the distribution of R 2 n = Z 2 1 + + Z 2 n. One can show (Pitman [10, pp. 358, 365]) that the density of R 2 is given by and we write f R 2 n (t) = 1 2 n/2 Γ(n/2) t(n/2) 1 e t/2, (t > 0), R 2 n χ 2 (n). This is also Gamma(n/2, 1/2) distribution. 6 The mean and variance are easily characterized: E R 2 n = n, Var R 2 n = 2n. It also follows from the definition that if X and Y are independent and X χ 2 (n) and Y χ 2 (m), then (X + Y ) χ 2 (n + m). - ( ) t=1 t=2 t=3 t=4 t=5 0 1 2 3 4 5 6 6 Recall that the Gamma distribution has two naming conventions. In the other convention this would be the Gamma(n/2, 2) distribution. KC Border v. 2017.01.25::17.49

KC Border Review Your Distributions S2 22 - ( ) t=1 t=2 t=3 t=4 t=5 0 1 2 3 4 5 6 - ( ) ( ) 0.06 0.05 Chi-square Normal 0.04 0.03 0.02 0.01 10 20 30 40 S2.14.1 Preview: chi-square test This is an important distribution in inferential statistics and is the basis of the chi-square goodness of fit test and the method of minimum chi-square estimation. If there are m possible outcomes of an experiment, and let p i be the probability that outcome i occurs. If the experiment is repeated independently N times, let N i be the number of times that outcome i is observed, so N = N 1 + + N m. Then the chi-square statistic n (N i Np i ) 2 i=1 Np i v. 2017.01.25::17.49 KC Border

KC Border Review Your Distributions S2 23 converges in distribution as N to the χ 2 (m 1) chi-square distribution with m 1 degrees of freedom. Bibliography [1] G. Casella and R. L. Berger. 2002. Statistical inference, 2d. ed. Pacific Grove, California: Wadsworth. [2] H. Cramér. 1946. Mathematical methods of statistics. Number 34 in Princeton Mathematical Series. Princeton, New Jersey: Princeton University Press. Reprinted 1974. [3] W. Feller. 1968. An introduction to probability theory and its applications, 3d. ed., volume 1. New York: Wiley. [4] C. Forbes, M. Evans, N. Hastings, and B. Peacock. 2011. Statistical distributions, 4th. ed. Hoboken, New Jersey: John Wiley & Sons. [5] J. Jacod and P. Protter. 2004. Probability essentials, 2d. ed. Berlin and Heidelberg: Springer. [6] N. L. Johnson and S. Kotz. 1969. Discrete distributions. Boston: Houghton Mifflin. [7]. 1970. Continuous univariate distributions I. New York: Houghton Mifflin. [8]. 1970. Continuous univariate distributions II. Boston: Houghton Mifflin. [9] R. J. Larsen and M. L. Marx. 2012. An introduction to mathematical statistics and its applications, fifth ed. Boston: Prentice Hall. [10] J. Pitman. 1993. Probability. Springer Texts in Statistics. New York, Berlin, and Heidelberg: Springer. [11] L. von Bortkiewicz. 1898. Das Gesetz der kleinen Zahlen [The law of small numbers]. Leipzig: B.G. Teubner. KC Border v. 2017.01.25::17.49