Skip to main content
Statistics LibreTexts

8.1: Confidence Intervals Using the z-Distribution

  • Page ID
    51821
  • \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

    \( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

    ( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\id}{\mathrm{id}}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\kernel}{\mathrm{null}\,}\)

    \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\)

    \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\)

    \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

    \( \newcommand{\vectorA}[1]{\vec{#1}}      % arrow\)

    \( \newcommand{\vectorAt}[1]{\vec{\text{#1}}}      % arrow\)

    \( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vectorC}[1]{\textbf{#1}} \)

    \( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

    \( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

    \( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

    \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

    \(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)

    With confidence intervals we will make our first statistical inference. Confidence intervals give us a direct inference about the population from a sample. The probability statement is one about hypotheses about the mean \(\mu\) of the population based on the mean \(\overline{x}\) and standard deviation \(s\) of the sample. This is a fine point. The frequentist definition of probability gives no way to assign a probability to a hypothesis. How do you count hypotheses? The central limit theorem makes a statement about the sample means \(\overline{x}\) on the basis of a hypothesis about a population, about its mean \(\mu\) and standard deviation \(\sigma\). If the population is fixed then the central limit theorem gives the results of counting sample means, frequentist probabilities. If we let \(H\) represent a hypothesis about a population (i.e. that it is described by \(\mu\) and \(\sigma\)) and let \(D\) represent data (with mean \(\overline{x}\)) then the central limit theorem gives the probability \(P(D \mid H) = P(\overline{x} \mid \mu, \sigma)\). The confidence intervals that we’ll look at first give \(P(H \mid D) = P(\mu \mid \overline{x}, \sigma)\). We’ll look at the recipe for computing confidence intervals for means first, then return to this discussion about probabilities for hypotheses.

    Our goal is to define a symmetric interval about the population mean \(\mu\) that will contain all potentially measured values of \(\bar{x}\) with a probability[1] of \(\cal{C}\).

    Typically \(\cal{C}\) will be

    \[ {\cal{C}} = 0.90 \hspace{.5in} \mbox{(90\% confidence)}\] \[ {\cal{C}} = 0.95 \hspace{.5in} \mbox{(95\% confidence)}\] \[ {\cal{C}} = 0.99 \hspace{.5in} \mbox{(99\% confidence)}\]

    The assumptions that we need in order to use the \(z\)-distribution to compute confidence intervals for means are :

    1. The population standard deviation, \(\sigma\), is known (a somewhat artificial assumption since it is usually not known in an experimental situation) or
    2. The sample size is greater than (or equal to) 30, \(n \geq 30\) and we use \(\sigma = s\), the sample standard deviation in our confidence interval formula.

    Definition : Let quicklatex.com-bd26baae39159fc531985204b9b489d1_l3.png where \({\cal{C}} = 1 - \alpha\) be the \(z\)-value, from the Standard Normal Distribution Table that corresponds to an area, between 0 and \(z_{\cal{C}}\) of quicklatex.com-a71afe5a00ea529b4ad965b53b128ad3_l3.png as shown in Figure 8.1.

    fig54apng-300x188.png
    Figure 8.1 : The \(z\)-distribution areas of interest associated with quicklatex.com-bd26baae39159fc531985204b9b489d1_l3.png.

    To get our confidence interval we simply inverse \(z\)-transform the picture of Figure 8.1, taking the mean of 0 to the sample mean \(\bar{x}\) and the standard deviation of 1 to the standard error quicklatex.com-697bcc7588a01e3546d608e57792873e_l3.png as shown in Figure 8.2.

    fig54bpng-300x177.jpg
    Figure 8.2 : The inverse \(z\)-transformation of Figure 8.1 gives the confidence interval for \(\mu\).

    So here is our recipe from Figure 8.2. The \(\cal{C}\)-confidence interval for the mean, under one of the two assumptions given above, is :

    or using notation that we will use as a standard way of denoting symmetric confidence intervals

    \[\begin{equation*} \bar{x} - E < \mu < \bar{x} + E \end{equation*}\]

    where

    \[ E = z_{\cal{C}} \left( \frac{\sigma}{\sqrt{n}}\right). \]

    The notation \(z_{\cal{C}}\) is more convenient for us than quicklatex.com-0104532c676aa4524826b08363df33f9_l3.png because we will use the t Distribution Table in the Appendix to find \(z_{\cal{C}}\) very quickly. We could equally well write

    \[ \mu = \bar{x} \pm E \]

    but we will use Equation (8.1) because it explicitly gives the bounds for the confidence interval.

    Notice how the confidence interval is backwards from the picture that the central limit theorem gives, the picture shown in Figure 8.3. We actually had no business using the inverse \(z\)-transformation quicklatex.com-b94567fea35875120a66d0f9a513b722_l3.png to arrive at Figure 8.2. It reverses the roles of \(\mu\) and \(\bar{x}\). We’ll return to this point after we work through the mechanics of an example.

    fig54png-300x172.jpg
    Figure 8.3 : The central limit theorem is about distributions of sample means.

    Example 8.2 : What is the 95\(\%\) confidence interval for student age if the population \(\sigma\) is 2 years, sample \(n = 50\), \(\bar{x} = 23.2\)?

    Solution : So \({\cal{C}} = 0.95\). First write down the formula prescription so you can see with numbers you need:

    \[ \bar{x} - E < \mu < \bar{x} + E \mbox{\hspace{2em}where\hspace{2em}} E = z_{95\%} \frac{\sigma}{\sqrt{n}}. \]

    First determine quicklatex.com-bd26baae39159fc531985204b9b489d1_l3.png. With the tables in the Appendices, there are two ways to do this. The first way is to use the Standard Normal Distribution Table noting that we need the \(z\) associated with a table area of . Using the table backwards we find \(z_{\cal{C}} = 1.96\). The second way, the recommended way especially during exams, is to use the t Distribution Table. Simply find the column for the 95\(\%\) confidence level and read the \(z\) from the last line of the table. We quickly find \(z_{95\%} = 1.960\).

    Either way we now find

    \[ E = 1.96( \frac{2}{\sqrt{50}}) = 0.6\]

    so

    \[\begin{eqnarray*} \bar{x} - E &< \mu <& \bar{x} + E\\ 23.2 - 0.6 &< \mu <& 23.2 + 0.6 \\ 22.6 &< \mu <& 23.8 \end{eqnarray*}\]

    with 95\(\%\) confidence.


    1. Because of this issue about probabilities of hypotheses, many prefer to say "confidence" and not probability. But we will learn enough about Bayesian probability to say "probability".

    This page titled 8.1: Confidence Intervals Using the z-Distribution is shared under a CC BY-NC-SA 4.0 license and was authored, remixed, and/or curated by Gordon E. Sarty via source content that was edited to the style and standards of the LibreTexts platform.