May 15, 2023
In my studies I have encountered commonly only 5 discrete and 7 continuous distributions.
There is little to be said of the uniform distributions, except that they are the most basic of the distributions. In fact, the uniform distribution forms the basis of scenarios which help derive the following distributions.
The Normal distribution is forced by two requirements.
Of note is the integral
which via transformation into polar coordinates shall equal , thus giving the normal distribution its constant.
Consider this scenario:
A bakery sees on average customers per minute, whose arrival times are distributed uniformly.
Questions on this scenario lead to the four distributions.
Geometric: One way this scenario could be fulfilled is by the following underlying mechanic: with probability , one customer arrives in each minute. Otherwise, no customer arrives in that minute. In which minute will the next customer arrive?
We take PDF as the probability that the next customer arrives in the -th minute. It must be then that for minutes prior, no customers arrive:
It is self-evident that the mean should be . The variance is not easy to compute, but is , and is had by separating .
Exponential: Dropping the geometric assumption earlier, how many minutes will pass before the next customer arrives?
The Exponential distribution is simply the continuous analogue of the geometric. Should we divide a minute into moments, we would expect customers per moment. Alternatively, with large enough , we expect a customer with probability .
For minutes to pass before the next customer arrives, then, we would see moments with no customer and a final moment with a customer:
recalling that .
Indeed, must sum to , and we have
so must equal .
As the continuous analogue of the geometric, the mean should remain the same, at . The variance is computed by a double integration by parts, and is .
Gamma: How many minutes will pass before the next customers arrive?
Immediately, this is simply the sum of exponential distributions, and by its name, must invoke the Gamma function. To intuit the Gamma function, consider how it must extend the factorial function with an integral. Certainly, this integral must be evaluated by cascading integration by parts:
Perhaps the in the exponent is unintuitive. But consider this: all the trash terms in the integration by parts must vanish, leaving only the factorial term remaining. vanishes the limit, so the term must vanish the limit. Hence, the exponent must be negative.
Alternatively, consider the weak Sterling’s approximation derived earlier:
Given that must be an integral, the inside of the integral must then be similar to the derivative of . Though weakly linked, there is indeed some similarity in form between and .
To intuit the PDF of the Gamma distribution, we take first the sum of two exponential RVs with PDF :
of which the final line may be inferred from the general form on the line prior. It remains to scale this PDF to sum to unity. To do this, of course, we invoke the function, with a clever substitution :
And thus we scale the proportional PDF earlier by :
In other literature, you may see the variables replaced with and replaced with .
As the Gamma distribution is the sum of i.i.d. exponential distributions, its mean is necessarily and its variance .
Poisson: Suppose now that , for ease of notation. How many customers will arrive in a minute?
Again, the approach of taking number of moments is fruitful. We have for large moments per minute that with probability a customer is slated to arrive. So for customers to arrive in the minute, of those moments must have a customer, and the rest must not.
By definition, the mean is . The variance, perhaps surprisingly, is also . The intuition to this lies in the binomial distribution: that when intervals are used, the mean is (i)P(x−i)di=∫i=0xλ2e−λie−λ(x−i)di=λ2[e−λxi]i=0x=λ2e−λxx=λnxn−1e−λx
of which the final line may be inferred from the general form on the line prior. It remains to scale this PDF to sum to unity. To do this, of course, we invoke the function, with a clever substitution :
And thus we scale the proportional PDF earlier by :
In other literature, you may see the variables replaced with and replaced with .
As the Gamma distribution is the sum of i.i.d. exponential distributions, its mean is necessarily and its variance .
Poisson: Suppose now that , for ease of notation. How many customers will arrive in a minute?
Again, the approach of taking number of moments is fruitful. We have for large moments per minute that with probability a customer is slated to arrive. So for customers to arrive in the minute, of those moments must have a customer, and the rest must not.
By definition, the mean is . The variance, perhaps surprisingly, is also . The intuition to this lies in the binomial distribution: that when intervals are used, the mean is , but the variance is , but vanishes to unity as goes to , and so the variance is the same as the mean, at .
Notably, the Poisson distribution is also the limit of the Binomial distribution taken .
There is little to be said of the Bernoulli and Binomial: it is too common, and it is easily derived. The mean is , and the variance , which is useful in gaining intuition for the Poisson distribution.
To derive the Beta distribution, consider the following scenario.
A possibly unfair coin is flipped times, and it comes up heads times and tails times. Suppose if you guess the bias of