Chapter 7

The binomial distribution

If we are interested in the probability for a specific outcome to occur \(x\) times in a series of \(n\) independent and identical random experiments, the answer is provided by the binomial distribution. This could be the total number of occurrences of the number \(6\) in \(n\) throws of a die or the total number of people with a specific characteristic in a random sample of size \(n\).

Educational objectives

After having worked through this chapter, you will be familiar with the binomial distribution and its parameters, its probability distribution and its expected value. You will be able to name various concrete examples of binomially distributed variables. You will also know the conditions under which a binomial distribution can be approximated by a normal distribution.

Moreover, you can describe the shape of the histogram of probabilities of a binomial distribution based on its parameters, calculate the expected value of a binomially distributed variable and determine the normal distribution which best approximates a given binomial distribution.

You know how the Poisson distribution can be seen as an extreme form of a binomial distribution. You can give examples of variables whose variation can be described by a Poisson distribution under certain conditions, and you can name these conditions.

Key words: binomial distribution, expected value, normal approximation, Poisson distribution

Previous knowledge: variable (chap. 1), relative frequency, bar chart (chap. 2), sample (chap. 4), probabilities (chap. 5), normal distribution (chap. 6)

Central questions: You draw several random population samples of fixed size and then count the number of smokers in each sample. How does this count vary from one sample to another? What does the frequency distribution of these counts look like? Which count can be expected on average? In which range will 95% of these counts lie?

Figure 7.1. Scheme of chapter 7

7.1. The binomial distribution

If a fair coin is tossed \(n\) times and \(n\) is sufficiently large, we will generally observe about the same number of "heads" and "tails". We thus expect to get about \( n/2 \) times head and \( n/2 \) times tail. Although the count (number) \( X \) of heads can vary in a relatively wide range, large deviations from the value \( n/2 \) are rare.

The frequency distribution of the count of heads among \( n = 20 \) tosses in a very large (i.e., ideally infinite) number of repetitions of \( n = 20 \) tosses is illustrated in the following diagram. It shows the probabilities of the so-called "binomial distribution" with the parameters \( n = 20 \) (number of repetitions of the random experiment - in our case the tossing of a coin) and \( \pi = 0.5 \) (probability of the outcome of interest in each random experiment). This probability distribution is symmetrical and the value occurring most often is \(10\).

Figure 7.2. Frequency distribution of the number of heads in 20 tosses of a fair coin

Another classical random experiment is the repeated throwing of a die, where we may count the number of sixes. The probability to get a six in one throw is \( 1/6 \). How often do we e.g. get exactly seven sixes, only sixes or no sixes at all in a series of 20 throws of the die? We thus ask for the probability distribution of the number of sixes in a series of \(20\) throws of a die.

Figure 7.3 below illustrates the probability distribution of the number of sixes in a series of \(20\) throws. We notice that the probability of only getting sixes is very small. (The bar at \(20\) is too small to be visible.) The same is already true for values bigger than nine. The most frequent values are \(3\) and \(4\). They are the integer neighbours of \( 20/6 = 3.333... \). Accordingly, the probability distribution is asymmetrical.

Figure 7.3. Frequency distribution of the number of sixes in 20 throws of a die

Epidemiological studies often involve the estimation of the frequency of a specific characteristic in a population. In such cases, a random sample of a certain size \(n\) is drawn from this population, and one counts how often the respective characteristic occurs in the sample (absolute frequency or count). Due to chance, this number - we denote it by \( X \) - can take various values. Most likely, we would get a different count in a second sample.

If the sample only represents a tiny fraction of the population, the probability pattern of the various possible values of \(X\) can be described by the "binomial distribution" with the parameters \(n\) (sample size) and \( \pi \) (relative frequency of the respective characteristic in the population).

In the preceding examples of tossing "heads" and throwing a six, \( \pi \) had the values \( 1/2 \) and \( 1/6 \), respectively, and \( n \) was equal to \(20\) in both examples. Any random drawing of a sample element can be regarded as an individual random experiment. Note: Here, the symbol \( X \) is not used to denote a variable which is observed in randomly chosen individuals, but it denotes a sample statistic which varies between different random samples.

Definition 7.1.1

If the relative frequency of a characteristic in a population is \( \pi \) and the number of observational units with the respective characteristic in a random sample of size \( n \) is denoted by \( X \) , then \( X \) is said to have a binomial distribution with the parameters \( n \) and \( \pi \), or simply to be binomially distributed with these two parameters.

The same applies to the number \( X \) of outcomes of interest in a series of \( n \) independent random experiments, if the probability of the outcome of interest equals \( \pi \) in each of the random experiments.

Work through the following three questions with the help of the applet "Binomial distribution".

Figure 7.4. Applet "Binomial distribution"

Question Nr. 7.1.1 You toss a fair coin fifteen times. What is the probability of getting exactly ten times tail?
approximately 0.69
approximately 0.07
approximately 0.09
approximately 0.6
almost 0
Question Nr. 7.1.2 You toss a fair coin eight times. What is the approximate probability that you either get three or five times tail?
approximately 0.54
approximately 0.25
approximately 0.43
approximately 0.35
approximately 0.21
Question Nr. 7.1.3 The relative frequency of atopic people in the adult Swiss population is estimated at approx. 30%. Hence the probability that there are more atopic than non-atopic people in a random sample of size 10 from the adult Swiss population is:
approximately 20%
approximately 5%
approximately 10%
approximately 1%
approximately 0.1%

Now we are interested in the average number \( \mu \) of outcomes of interest observed in a series of \( n \) independent random experiments, where the probability of the outcome of interest equals \( \pi \) in each of the experiments. Intuition tells us that \( \mu \) must be proportional to both \( n \) and \( \pi \). Hence, \( \mu = c \times n \times \pi \), for some constant number \( c \). Since \( \mu = n \), if \( \pi = 1 \), we have \( c \times n \times 1 = n \), implying that \( c = 1 \). Therefore \( \mu = n \times \pi \). The same holds true for the average number \( \mu \) of carriers of a certain characteristic in a random sample of size \( n \) from a population in which the characteristic has a relative frequency of \( \pi \).

Definition 7.1.2

Suppose that we

a) know the relative frequency \( \pi \) of a certain characteristic in the population and

b) draw a very large number of random samples of a given size \( n \) from this population.

Then the number \( X \) of the carriers of the characteristic will vary from one sample to another, but the mean value of \( X \) across all samples can be predicted with very high accuracy. It is called "expected value" of \( X \) and is denoted by \( E(X) \). In fact, \( E(X) \) is a synonym of \( \mu_X \).

Synopsis 7.1.1

If \( X \) has a binomial distribution with parameters \( n \) and \( \pi \), then its expected value \( E(X) \) is given by \[ E(X) = n \times \pi \, . \] Typically, such a variable \( X \) is either

(A) the absolute frequency (count) of a certain characteristic in a random sample of size \( n \)

or

(B) the number of outcomes of interest in a series of \( n \) independent and identical random experiments.

(A) If the characteristic has a relative frequency of \( \pi \) in the population, then we expect, on average, \( n \times \pi \) carriers of the respective characteristic in repeated random samples of size \( n \).

(B) If the outcome of interest occurs with a probability of \( \pi \) in each random experiment, then we expect, on average, \( n \times \pi \) outcomes of interest in repeated series of \( n \) independent random experiments.

Answer the following question without the help of the applet.

Question Nr. 7.1.4 The relative frequency of diabetics in the elderly Swiss population (people over 60 years of age) is approx. 8%. In a random sample of 150 people from this population we therefore expect
about 8 diabetics.
about 16 diabetics.
about 12 diabetics.
about 90 diabetics.
about 13 diabetics.

In the example of the coin tosses you saw that the binomial distribution with the parameters \( n = 20 \) and \( \pi = 0.5 \) is symmetrical. The example of die throws illustrates that this is not always the case. Here, the parameters were \( n = 20 \) and \( \pi = 1/6 \). Take a look at figure 7.5 to see this.

Figure 7.5. Comparison of three binomial distributions with \( n = 40 \) and different values of \( \pi \)

We can see that the two probability distributions with the values \( \pi = 0.05 \) and \( \pi = 0.95 \) are highly asymmetrical. For \( \pi = 0.5 \) we get a symmetrical probability distribution. In figure 7.5, \( n \) was fixed and \( \pi \) varied. But what will happen if we fix \( \pi \) and vary \( n \)? Take a look at figure 7.6 to see what happens.

Figure 7.6. Comparison of three binomial distributions with \( \pi = 0.1 \) and varying sample size \( n \)

For \( n = 25 \) we get a highly asymmetrical probability distribution. With increasing \( n \), the probability distribution becomes more and more symmetrical. These observations are summarised in the following synopsis:

Synopsis 7.1.2

The following applies to the bar chart of probabilities of the binomial distribution:

(i) It is exactly symmetrical if \( \pi = 0.5 \), i.e., \( P(X = x) = P(X = n - x) \) for all values \(x\).

(ii) If the sample size is fixed, the asymmetry of the distribution increases the closer \( \pi \) gets to the limits \( 0 \) or \( 1 \).

(iii) If \( \pi \) is fixed, the asymmetry decreases with increasing sample size \( n \) .

7.2. Approximation of the binomial distribution by the normal distribution

Take another look at the lower graph of figure 7.6. It is striking that the bar chart representing the probabilities of the respective binomial distribution has the same shape as a Gaussian bell curve. This suggests to draw the Gaussian bell curve which best fits the bar chart. We shall now see how this can be done (thereby, we ignore that binomially distributed variables are discrete and that only continuous variables can be normally distributed).

The solution is quite simple if we remember that the normal distribution is characterised by its mean value \( \mu \) and its standard deviation \( \sigma \). If a binomial distribution with the parameters \( n \) and \( \pi \) is given, then we simply approximate it by the normal distribution, which has the same mean and the same standard deviation as the respective binomial distribution.

We have seen that the mean value of the binomial distribution equals \( \mu = n \times \pi \). Moreover, the standard deviation of the binomial distribution is \( \sqrt{n \times \pi \times (1 - \pi)} \) (you just have to believe this). We thus choose \( \mu \) and \( \sigma \) as follows: \[ \mu = n \times \pi \] and \[ \sigma = \sqrt{n \times \pi \times (1 - \pi)} \]

Now take a look at figure 7.7.

Figure 7.7. Approximation of the binomial distribution by the normal distribution (\(y\)-axis: probability in \(\%\))

In these graphs the approximating Gaussian bell curves were laid over the bar charts of probabilities of the binomial distributions.

In the upper two graphs, the approximation is not convincing, because the respective probability distributions are asymmetrical. Notice that the approximating bell curve of the upper left graph has negative \(x\)-values.

In the left graph of the middle row, the approximation is quite good even though \( n = 40 \) is relatively small. Here, the binomial distribution is symmetrical.

If we increase the sample size to \(200\), the approximations become pretty good also for the binomial distributions with \( \pi = 0.05 \) or \( \pi = 0.95 \).

When can we consider the approximation of a binomial distribution by the corresponding normal distribution as satisfactory? A general rule says that this is the case if \[ n \times \pi \times (1 - \pi) \ge 10 \, .\] This condition is met by the middle left ( \( n = 40, \pi = 0.5 \)) and the lower right (\( n = 200, \pi = 0.5 \)) distribution. The condition is clearly not met by the two upper distributions with \( n \times \pi \times (1 - \pi) = 40 \times \ 0.05 \times 0.95 = 1.9 \). In the remaining cases, where we found the approximations to be pretty good, \( n \times \pi \times (1 - \pi) = 200 \times \ 0.05 \times 0.95 = 9.5 \). Thus, the condition is almost met in these two cases.

Synopsis 7.2.1

If the parameters \( n \) and \( \pi \) of a binomial distribution satisfy \[n \times \pi \times (1 - \pi) \ge 10 \] then the binomial distribution can be approximated by a normal distribution with \[ \mu = n \times \pi \,\,\, \text{(mean)} \] and \[ \sigma = \sqrt{ n \times \pi \times (1 - \pi)} \,\,\, \text{(standard deviation)} \]

If the variable \( X \) has a normal distribution with mean \( \mu \) and standard deviation \( \sigma \) , then the probability that \( X \) lies in the interval \( [ \mu - 2 \times \sigma, \, \mu + 2 \times \sigma ] \) is approximately \(0.95\) (chapter 6).

If \( X \) has a binomial distribution with parameters \( n \) and \( \pi \) satisfying \( n \times \pi \times (1 - \pi) \ge 10 \) , then we can consider \( X \) to be approximately normally distributed with a mean value of \( n \times \pi \) and a standard deviation of \( \sqrt{ n \times \pi \times (1 - \pi) } \).

From this we can infer that \( X \) takes a value in the interval \[ [ n \times \pi - 2 \times \sqrt{ n \times \pi \times (1 - \pi) }, \, n \times \pi + 2 \times \sqrt{ n \times \pi \times (1 - \pi) } ] \] with a probability of about \(95\%\).

Question Nr. 7.2.1 Assume that there is an equal number of men and women in a given population.
Let X denote the number of women in a random sample of one hundred persons from this population. In which interval does the value of X lie with a probability of approx. 95%?
(45, 55)
(30, 70)
(33, 67)
(40, 60)

Synopsis 7.2.2

If \( X \) is binomially distributed with parameters \( n \) and \( \pi \) satisfying \[n \times \pi \times (1 - \pi) \geq 10 \] then, with a probability of about \(0.95\), the value of \( X \) lies in the interval \[ [n \times \pi - 2 \times \sqrt{ n \times \pi \times (1 - \pi) }, \, n \times \pi + 2 \times \sqrt{ n \times \pi \times (1 - \pi) }] \]

7.3. The Poisson distribution

If we let \( n \) increase and \( \pi \) decrease simultaneously in a binomial distribution, while keeping the product \( n \times \pi \) constant, then a so-called Poisson distribution results in the limit. The expected value \( n \times \pi \) of the number of events of interest is usually denoted by \( \lambda \) in a Poisson distribution.

If we consider \( \lambda = n \times \pi \) as constant, then \( \pi = \lambda/n \) must hold. Hence \( \pi \) approaches \(0\) with increasing \( n \). Based on the formula of the normal distribution approximation of a binomially distributed variable \( X \), we can see that the standard deviation \(\sigma\) of \(X\) satisfies \[ \sigma = \sqrt{ n \times \pi \times (1 - \pi) } \approx \sqrt{ n \times \pi } = \sqrt{\lambda} \] if \( \pi \) is very small. Thus, the standard deviation \(\sigma\) of a variable \( X \) with a Poisson-distribution and an expected value \( E(X) = \lambda \) equals \[ \sigma = \sqrt{\lambda} \, .\]

Therefore, the normal distribution approximating the Poisson-distribution with parameter \( \lambda \) has a mean value of \( \lambda \) and a standard deviation of \( \sqrt{\lambda} \). In analogy to the binomial distribution, this approximation is satisfactory if the expected number \( \lambda \) of events of interest is at least \(10\) (cf. figure 7.8).

Among other, the Poisson distribution is suitable for describing the number of occurrences of a rare event (e.g., of a rare disease) in a specific population during a certain period of time (e.g. during one year). However, for such counts to follow a Poisson distribution across several identical time intervals, the population itself and the conditions surrounding it must remain stable. Otherwise, \(\lambda\) may change over time.

One can often assume that the expected value \( \lambda \) is proportional to the length of the observation period. Figure 7.8. illustrates two such situations.

With the applet "Poisson distribution" you can generate such probability distributions for arbitrary parameter values \( \lambda \).

Figure 7.8. Probability distribution of the number (count) of new cases of disease

7.4. Summary

If we repeatedly draw random samples of size \(n\) from a population and count the number of observational units with a certain characteristic each time, then these counts are binomially distributed with parameters \(n\) and \( \pi \) (where \( \pi \) is the probability or relative frequency of the characteristic in the population).

The histogram of probabilities of the binomial distribution is symmetrical if \( \pi = 0.5 \), independent of the sample size \( n \), while it is highly asymmetrical for small to moderate sample sizes, if \( \pi \) is close to \(1\) or \(0\). The mean value of the observed counts in a series of such samples approaches the value \( n \times \pi \), the so-called expected value, if the number of samples increases.

If \( n \times \pi \times (1 - \pi) \geq 10 \), then the distribution of these counts can be considered as approximately normal, with a mean value of \( n \times \pi \) and a standard deviation of \( \sqrt{ n \times \pi \times (1 - \pi)} \). In this case, we can expect that the absolute frequency (count) of the characteristic of interest in a random sample of size \(n\) will lie in the interval \( n \times \pi \pm 2 \times \sqrt{ n \times \pi \times (1 - \pi)} \) with a probability of approx. \(95\%\).

The Poisson distribution can be considered as an "extreme case" of a binomial distribution where \( \pi \) is very small and \( n \) very large. The expectation \(E(X) = n \times \pi\) is denoted by \(\lambda\). The Poisson distribution is completely defined by this parameter \(\lambda\).

The mean value and standard deviation of a variable \(X\) with a Poisson distribution are \( \lambda\) and \(\sqrt{\lambda}\), respectively.

If \(\lambda \geq 10\), then the distribution of \(X\) can be approximated by a normal distribution with \(\mu = \lambda\) and \(\sigma = \sqrt{\lambda}\).

Among other, the number of occurrences of a rare event (e.g., a rare disease) in a fixed population during separate time intervals of equal length can generally be well described by a Poisson distribution, provided that \(\lambda\) does not change over time.

Figure 7.9. Scheme of chapter 7 with gaps