5.4: Binomial Distribution
- Page ID
- 56126
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\( \newcommand{\dsum}{\displaystyle\sum\limits} \)
\( \newcommand{\dint}{\displaystyle\int\limits} \)
\( \newcommand{\dlim}{\displaystyle\lim\limits} \)
\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)
( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\id}{\mathrm{id}}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\kernel}{\mathrm{null}\,}\)
\( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\)
\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\)
\( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)
\( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)
\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)
\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vectorC}[1]{\textbf{#1}} \)
\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)
\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)
\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\(\newcommand{\longvect}{\overrightarrow}\)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)The binomial distribution is a discrete probability distribution used to find the probability of success when there are two outcomes to each trial, and there are a set number of independent trials with the same probability of occurrence. To develop the process for calculating the probabilities in a binomial experiment, consider the following example.
Suppose you are given a three-question multiple-choice test. Each question has four responses and only one is correct. Suppose you want to find the probability that you can just guess at the answers and get two questions correct. To help with the idea that you are going to guess, suppose the test is on string theory.
Solution
a) The random variable is x = number of correct answers. 1. There are three questions, and each question is a trial, so there are a fixed number of trials. In this case, n = 3. 2. Getting the first question right has no effect on getting the second or third question correct, thus the trials are independent. 3. Either you get the question right or you get it wrong, so there are only two outcomes. In this case, the success is getting the question right. 4. The probability of getting a question right is one out of four. This is the same for every trial since each question has 4 responses. In this case, p = ¼ and q = 1 – ¼ = ¾.
b) To answer this question, start with the sample space. SS = {RRR, RRW, RWR, WRR, WWR, WRW, RWW, WWW}, where RRW means you get the first question right, the second question right, and the third question wrong.
Now the event space for getting 2 right is {RRW, RWR, WRR}. What you did in chapter four was just to find three divided by eight to compute the probability outcome. However, this would not be correct in this case, because the probability of getting a question right is different from getting a question wrong. What else can you do?
Look at just P(RRW) for the moment. Again, that means P(RRW) = P(R on 1st, R on 2nd, and W on 3rd). Since the trials are independent, then P(RRW) = P(R on 1st, R on 2nd, and W on 3rd) = P(R on 1st) * P(R on 2nd) * P(W on 3rd). Just multiply p · p · q = P(RRW) = ¼ ∙ ¼ ∙ ¾ = (¼)2 ∙ (¾)1.
Similarly, you can compute the P(RWR) and P(WRR). To find the probability of 2 correct answers, just add these three probabilities together.
P(2 correct answers) = P(RRW) + P(RWR) + P(WRR) = (¼)2 ∙ (¾)1 + (¼)2 ∙ (¾)1 + (¼)2 ∙ (¾)1 = 3∙(¼)2 ∙ (¾)1 .
c) You could go through the same argument that you did above and come up with the following:

Do you see the resulting pattern? You can now write the general formula for the probabilities for a binomial experiment.
- Is this a binomial experiment? Since all of the properties are met, this is a binomial experiment.
- What is the probability of getting two questions right?
- What is the probability of getting zero right, one right, two right and all three right?
First, the random variable in a binomial experiment is x = number of successes.
A binomial probability distribution results from a random experiment that meets all of the following requirements.
- The procedure has a fixed number of trials (or steps), which is denoted by n.
- The trials must be independent.
- Each trial must have exactly two categories that can be labeled “success” and “failure.”
- The probability of a “success,” denoted by p, remains the same in all trials. The probability of “failure” is often denoted by q, thus q = 1 – p.
- Random Variable, X, counts the number of “successes.”
If a random experiment satisfies all of the above, the distribution of the random variable X, where X counts the number of successes, is called a binomial distribution. A binomial distribution is described by the population proportion p and the sample size n. If a discrete random variable X has a binomial distribution with population proportion p and sample size n, we write X ~ B(n, p).
The general formula for the binomial distribution is given by
\[
P(X = k) = \frac{n!}{k!(n - k)!} \cdot p^k \cdot (1 - p)^{n - k}
\]
Where:
- n = number of independent trials
- k = number of successes
- p = probability of success on each trial
- (1−p)(1 - p)(1−p) = probability of failure
- n! = factorial of n
Be careful, a “success” is not always a “good” thing. Sometimes a success is something that is “bad,” like finding a defect or getting in a car crash. The success will be the event from the probability question.
The geometric and binomial distributions are easy to mix up. Keep in mind that the binomial distribution has a given sample size, whereas the geometric is sampling until you get a success.
As you read through a problem look for some of the following key phrases in Figure 5-6. Once you find the phrase then match up to what sign you would use and then use the table to walk you through the computer or calculator formula. The same idea about signs applies to all the discrete probabilities that follow.

Figure 5-6
As of 2018, the Centers for Disease Control and Prevention (CDC) reported that about 1 in 88 children in the United States have been diagnosed with autism spectrum disorder (ASD). A researcher randomly selects 10 children. Compute the probability that 2 children have been diagnosed with ASD.
Solution
The random variable x = number of children with ASD. There are 10 children, and each child is a trial, so there are a fixed number of trials. In this case, n = 10. If you assume that each child in the group is chosen at random, then whether a child has ASD does not affect the chance that the next child has ASD. Thus, the trials are independent. Either a child has been diagnosed with ASD or they have not been diagnosed ASD, so there are two outcomes. In this case, the success is a child has ASD. The probability of a child having ASD is 1/88. This is the same for every trial since each child has the same chance of having ASD, \(p=\frac{1}{88} \text { and } q=1-\frac{1}{88}=\frac{87}{88}\)
Using the formula: \(\mathrm{P}(X=2)={ }_{10} C_{2} \cdot\left(\frac{1}{88}\right)^{2} \cdot\left(\frac{87}{88}\right)^{(10-2)}=0.0053\)
Flip a fair coin exactly 10 times.
a) What is the probability of getting exactly 8 tails?
b) What is the probability of getting 8 or more tails?
Solution
a) There are only two outcomes to each trial, heads or tails. The coin flips are independent and the probability of a success, flipping a tail, p = ½ = 0.5 is the same for each trial. This is a binomial experiment with a sample size n = 10. Using the formula: P(X = 8) = 10C8·0.58 ·0.5(10-8) = 0.0439
b) We still have p = 0.5 and n = 10, however we need to look at the probability of 8 or more. The P(X ≥ 8) = P(X = 8) + P(X =9) + P(X = 10). We can stop at x = 10 since the coin was only flipped 10 times. P(X ≥ 8) = 10C8· 0.58 · 0.5(10-8) + 10C9· 0.59 · 0.5(10-9) + 10C10· 0.510 · 0.5(10-10) = 0.0439 + 0.0098 + 0.0010 = 0.0547.
So far, most of the examples for the binomial distribution were for exactly x successes. If we want to find the probability of accumulation of x values then we would use the cumulative distribution function (cdf) instead of the pdf. Phrases such as “at least,” “more than,” or “below” can drastically change the probability answers.
Approximately 10.3% of American high school students drop out of school before graduation. Choose 10 students entering high school at random. Find the following probabilities.
a) No more than two drop out.
b) At least 6 students graduate.
c) Exactly 10 stay in school and graduate.
Solution
a) There is a set sample size of independent trials. The person either has or has not dropped out of school before graduation. A “success” is what we are finding the probability for, so in this case a success is to drop out so p = 0.103 and q = 1 – 0.103 = 0.897.
P(X ≤ 2) = P(X = 0) + P(X = 1) + P(X = 2) = 10C0 · 0.1030 · 0.89710 + 10C1 · 0.1031 ·0.8979 + 10C2 · 0.1032 · 0.8978 = 0.3372 + 0.3872 + 0.2001 = 0.9245
b) A success is to graduate so p = 0.897 and q = 0.103.
P(X ≥ 6) = P(X = 6) + P(X = 7) + P(X = 8) + P(X = 9) + P(X = 10) = 10C6 · 0.8976 · 0.1034 + 10C7 · 0.8977 · 0.1033 + 10C8 · 0.8978 · 0.1032 + 10C9 · 0.8979 · 0.1031 + 10C10 · 0.89710 · 0.1030 = 0.0123 + 0.0613 + 0.2001 + 0.3872 + 0.3372 = 0.9981.
This is a lot of work to do each one so we will often use technology to find the answer.
c) A success is to graduate so p = 0.897 and q = 0.103. Find P(X = 10) = 10C10 · 0.89710 · 0.1030 = 0.3372.
When looking at a person’s eye color, it turns out that only 2% of people in the world have green eyes (not to be confused with hazel colored eyes). Consider a randomly selected group of 20 people.
a) Compute the probability that at most 3 have green eyes.
b) Compute the probability that less than 3 have green eyes.
c) Compute the probability that more than 3 have green eyes.
d) Compute the probability that 3 or more have green eyes.
Solution
a) This fits a binomial experiment. P(X ≤ 3) = P(X = 0) + P(X = 1) + P(X = 2) + P(X = 3) = 20C0 · 0.020 · 0.8020 + 20C1 · 0.021 · 0.8019 + 20C2 · 0.022 · 0.8018 + 20C3 · 0.023 · 0.8017 = 0.667608 + 0.272493 + 0.05283 + 0.006469 = 0.9994.
b) P(X < 3) = P(X = 0) + P(X = 1) + P(X = 2) = 0.667608 + 0.272493 + 0.05283 = 0.9929.
c) P(X > 3) = 1 – P(X ≤ 3) = 1 – 0.9994 = 0.0006.
d) P(X ≥ 3) = 1 – P(X ≤ 2) = 1 – 0.9929 = 0.0071
As of 2018, the Centers for Disease Control and Prevention (CDC) reported that about 1 in 88 children in the United States have been diagnosed with autism spectrum disorder (ASD). A researcher randomly selects 10 children. Compute the probability that at least 2 children have been diagnosed with ASD.
Solution
\(\mathrm{P}(X \geq 2)=1-\mathrm{P}(X \leq 1)=1-\left({ }_{10} C_{0} \cdot\left(\frac{1}{88}\right)^{0} \cdot\left(\frac{87}{88}\right)^{10-0)}+{ }_{10} C_{1} \cdot\left(\frac{1}{88}\right)^{1} \cdot\left(\frac{87}{88}\right)^{(10-1)}\right)=1-(0.89202+0.102528)=0.0055\)
Mean, Variance & Standard Deviation of a Binomial Distribution
If you list all possible values of x in a Binomial distribution, you get the Binomial Probability Distribution (pdf). You can then find the mean, the variance, and the standard deviation using the general formulas μ = Σ(xi ∙ P(xi)) and σ2 = ∑(xi2 ∙P(xi)) – μ2 . This, however, would take a lot of work if you had a large value for n. If you know the type of distribution, like binomial, then you can find the mean, variance and standard deviation using easier formulas. They are derived from the general formulas. For a Binomial distribution, μ, the expected number of successes, σ2 , the variance, and σ, the standard deviation for the number of success are given by the formulas, where p is the probability of success and q = 1 – p.
The expected value is calculated by \[
\mu = E[X] = n p
\]
The variance is calculated by \[
\sigma^2 = \text{Var}(X) = n p (1 - p)
\]
The standard deviation is calculated by \[
\sigma = \sqrt{n p (1 - p)}
\]
A random experiment consists of flipping a coin three times. Let X = the number of heads that show up. Compute the mean and standard deviation of X, that is, the mean and standard deviation for the number of heads that show up when a coin is flipped three times.
Solution
This experiment follows a binomial distribution; hence, we can use the mean and standard deviation formulas for a binomial. The mean of number of heads is µ = 3·0.5 = 1.5. The standard deviation of X is \(\sigma=\sqrt{n \cdot p \cdot q}=\sqrt{(3 \cdot 0.5 \cdot(1-0.5))}=0.8660\)
When looking at a person’s eye color, it turns out that only 2% of people in the world have green eyes (not to be confused with hazel colored eyes). Consider a randomly selected group of 20 people. Compute the mean, variance and standard deviation.
Solution
Since this is a binomial experiment, then you can use the formula μ = n ∙ p. So μ = 20 · 0.02 = 0.4 people. You would expect on average that out of 20 people, less than 1 would have green eyes. The variance would be σ2 = n ∙ p ∙ q = 20(0.02)(0.98) = 0.392 people2 . Once you have the variance, you just take the square root of the variance to find the standard deviation σ = \(\sqrt0.392\) = 0.6261 people. We would expect on average spread of the distribution to have 0.4 \(\pm\) 0.6261 or 0 to 1 person out of 20 people to have green eyes.


