Skip to main content
Statistics LibreTexts

5.7: Binomial Distribution

  • Page ID
    2360
  • \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

    \( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

    ( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\id}{\mathrm{id}}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\kernel}{\mathrm{null}\,}\)

    \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\)

    \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\)

    \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

    \( \newcommand{\vectorA}[1]{\vec{#1}}      % arrow\)

    \( \newcommand{\vectorAt}[1]{\vec{\text{#1}}}      % arrow\)

    \( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vectorC}[1]{\textbf{#1}} \)

    \( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

    \( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

    \( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

    \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

    \(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)

    Learning Objectives

    • Define binomial outcomes
    • Compute the probability of getting \(X\) successes in \(N\) trials
    • Compute cumulative binomial probabilities
    • Find the mean and standard deviation of a binomial distribution

    When you flip a coin, there are two possible outcomes: heads and tails. Each outcome has a fixed probability, the same from trial to trial. In the case of coins, heads and tails each have the same probability of \(1/2\). More generally, there are situations in which the coin is biased, so that heads and tails have different probabilities. In the present section, we consider probability distributions for which there are just two possible outcomes with fixed probabilities summing to one. These distributions are called binomial distributions.

    A Simple Example

    The four possible outcomes that could occur if you flipped a coin twice are listed below in Table \(\PageIndex{1}\). Note that the four outcomes are equally likely: each has probability \(1/4\). To see this, note that the tosses of the coin are independent (neither affects the other). Hence, the probability of a head on \(\text{Flip 1}\) and a head on \(\text{Flip 2}\) is the product of \(P(H)\) and \(P(H)\), which is \(1/2 \times 1/2 = 1/4\). The same calculation applies to the probability of a head on \(\text{Flip 1}\) and a tail on \(\text{Flip 2}\). Each is \(1/2 \times 1/2 = 1/4\).

    Table \(\PageIndex{1}\): Four Possible Outcomes
    Outcome First Flip Second Flip
    1 Heads Heads
    2 Heads Tails
    3 Tails Heads
    4 Tails Tails

    The four possible outcomes can be classified in terms of the number of heads that come up. The number could be two (Outcome \(1\), one (Outcomes \(2\) and \(3\)) or \(0\) (Outcome \(4\)). The probabilities of these possibilities are shown in Table \(\PageIndex{2}\) and in Figure \(\PageIndex{1}\).

    binomial1.jpg
    Figure \(\PageIndex{1}\): Probabilities of \(0\), \(1\), and \(2\) heads

    Since two of the outcomes represent the case in which just one head appears in the two tosses, the probability of this event is equal to \(1/4 + 1/4 = 1/2\). Table \(\PageIndex{2}\) summarizes the situation.

    Table \(\PageIndex{2}\): Probabilities of Getting \(0\), \(1\), or \(2\) Heads
    Number of Heads Probability
    0 1/4
    1 1/2
    2 1/4

    Figure \(\PageIndex{1}\) is a discrete probability distribution: It shows the probability for each of the values on the \(X\)-axis. Defining a head as a "success," Figure \(\PageIndex{1}\) shows the probability of \(0\), \(1\), and \(2\) successes for two trials (flips) for an event that has a probability of \(0.5\) of being a success on each trial. This makes Figure \(\PageIndex{1}\) an example of a binomial distribution.

    The Formula for Binomial Probabilities

    The binomial distribution consists of the probabilities of each of the possible numbers of successes on \(N\) trials for independent events that each have a probability of \(\pi\) (the Greek letter pi) of occurring. For the coin flip example, \(N = 2\) and \(\pi =0.5\). The formula for the binomial distribution is shown below:

    \[P(x)=\dfrac{N!}{x!(N-x)!}\pi ^x(1-\pi )^{N-x}\]

    where \(P(x)\) is the probability of x successes out of \(N\) trials, \(N\) is the number of trials, and \(\pi\) is the probability of success on a given trial. Applying this to the coin flip example,

    \[\begin{align} P(0)&=\dfrac{2!}{0!(2-0)!}(0.5)^0(1-0.5)^{2-0} \\[5pt] &=\dfrac{2}{2}(1)(0.25) \\[5pt] &=0.25 \end{align}\]

    \[\begin{align}P(0)&=\dfrac{2!}{1!(2-1)!}(0.5)^1(1-0.5)^{2-1} \\[5pt] &=\dfrac{2}{1}(0.5)(0.5) \\[5pt] &=0.50\end{align}\]

    \[\begin{align}P(0)&=\dfrac{2!}{2!(2-2)!}(0.5)^2(1-0.5)^{2-2} \\[5pt] &=\dfrac{2}{2}(0.25)(1) \\[5pt] &=0.25\end{align}\]

    If you flip a coin twice, what is the probability of getting one or more heads? Since the probability of getting exactly one head is \(0.50\) and the probability of getting exactly two heads is \(0.25\), the probability of getting one or more heads is \(0.50 + 0.25 = 0.75\).

    Now suppose that the coin is biased. The probability of heads is only \(0.4\). What is the probability of getting heads at least once in two tosses? Substituting into the general formula above, you should obtain the answer \(0.64\).

    Cumulative Probabilities

    We toss a coin \(12\) times. What is the probability that we get from \(0\) to \(3\) heads? The answer is found by computing the probability of exactly \(0\) heads, exactly \(1\) head, exactly \(2\) heads, and exactly \(3\) heads. The probability of getting from \(0\) to \(3\) heads is then the sum of these probabilities. The probabilities are: \(0.0002\), \(0.0029\), \(0.0161\), and \(0.0537\). The sum of the probabilities is \(0.073\). The calculation of cumulative binomial probabilities can be quite tedious. Therefore we have provided a binomial calculator to make it easy to calculate these probabilities.

    Mean and Standard Deviation of Binomial Distributions

    Consider a coin-tossing experiment in which you tossed a coin \(12\) times and recorded the number of heads. If you performed this experiment over and over again, what would the mean number of heads be? On average, you would expect half the coin tosses to come up heads. Therefore the mean number of heads would be \(6\). In general, the mean of a binomial distribution with parameters \(N\) (the number of trials) and \(\pi\) (the probability of success on each trial) is:

    \[\mu =N\pi\]

    where \(\mu\) is the mean of the binomial distribution. The variance of the binomial distribution is:

    \[\sigma ^2=N\pi (1-\pi )\]

    where \(\sigma ^2\) is the variance of the binomial distribution.

    Let's return to the coin-tossing experiment. The coin was tossed \(12\) times, so \(N = 12\). A coin has a probability of \(0.5\) of coming up heads. Therefore, \(\pi =0.5\). The mean and variance can therefore be computed as follows:

    \[\mu =N\pi =(12)(0.5)=6\]

    \[\sigma ^2=N\pi (1-\pi )=(12)(0.5)(1.0-0.5)=3.0\]

    Naturally, the standard deviation (\(\sigma\)) is the square root of the variance (\(\sigma ^2\)).

    \[\sigma =\sqrt{N\pi (1-\pi )}\]


    This page titled 5.7: Binomial Distribution is shared under a Public Domain license and was authored, remixed, and/or curated by David Lane via source content that was edited to the style and standards of the LibreTexts platform.