8.1: Confidence Intervals Using the z-Distribution
- Page ID
- 51821
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)
( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\id}{\mathrm{id}}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\kernel}{\mathrm{null}\,}\)
\( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\)
\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\)
\( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)
\( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)
\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)
\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vectorC}[1]{\textbf{#1}} \)
\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)
\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)
\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)With confidence intervals we will make our first statistical inference. Confidence intervals give us a direct inference about the population from a sample. The probability statement is one about hypotheses about the mean \(\mu\) of the population based on the mean \(\overline{x}\) and standard deviation \(s\) of the sample. This is a fine point. The frequentist definition of probability gives no way to assign a probability to a hypothesis. How do you count hypotheses? The central limit theorem makes a statement about the sample means \(\overline{x}\) on the basis of a hypothesis about a population, about its mean \(\mu\) and standard deviation \(\sigma\). If the population is fixed then the central limit theorem gives the results of counting sample means, frequentist probabilities. If we let \(H\) represent a hypothesis about a population (i.e. that it is described by \(\mu\) and \(\sigma\)) and let \(D\) represent data (with mean \(\overline{x}\)) then the central limit theorem gives the probability \(P(D \mid H) = P(\overline{x} \mid \mu, \sigma)\). The confidence intervals that we’ll look at first give \(P(H \mid D) = P(\mu \mid \overline{x}, \sigma)\). We’ll look at the recipe for computing confidence intervals for means first, then return to this discussion about probabilities for hypotheses.
Our goal is to define a symmetric interval about the population mean \(\mu\) that will contain all potentially measured values of \(\bar{x}\) with a probability[1] of \(\cal{C}\).
Typically \(\cal{C}\) will be
\[ {\cal{C}} = 0.90 \hspace{.5in} \mbox{(90\% confidence)}\] \[ {\cal{C}} = 0.95 \hspace{.5in} \mbox{(95\% confidence)}\] \[ {\cal{C}} = 0.99 \hspace{.5in} \mbox{(99\% confidence)}\]The assumptions that we need in order to use the \(z\)-distribution to compute confidence intervals for means are :
- The population standard deviation, \(\sigma\), is known (a somewhat artificial assumption since it is usually not known in an experimental situation) or
- The sample size is greater than (or equal to) 30, \(n \geq 30\) and we use \(\sigma = s\), the sample standard deviation in our confidence interval formula.
Definition : Let where \({\cal{C}} = 1 - \alpha\) be the \(z\)-value, from the Standard Normal Distribution Table that corresponds to an area, between 0 and \(z_{\cal{C}}\) of
as shown in Figure 8.1.


To get our confidence interval we simply inverse \(z\)-transform the picture of Figure 8.1, taking the mean of 0 to the sample mean \(\bar{x}\) and the standard deviation of 1 to the standard error as shown in Figure 8.2.

So here is our recipe from Figure 8.2. The \(\cal{C}\)-confidence interval for the mean, under one of the two assumptions given above, is :
or using notation that we will use as a standard way of denoting symmetric confidence intervals
\[\begin{equation*} \bar{x} - E < \mu < \bar{x} + E \end{equation*}\]where
\[ E = z_{\cal{C}} \left( \frac{\sigma}{\sqrt{n}}\right). \]The notation \(z_{\cal{C}}\) is more convenient for us than because we will use the t Distribution Table in the Appendix to find \(z_{\cal{C}}\) very quickly. We could equally well write
but we will use Equation (8.1) because it explicitly gives the bounds for the confidence interval.
Notice how the confidence interval is backwards from the picture that the central limit theorem gives, the picture shown in Figure 8.3. We actually had no business using the inverse \(z\)-transformation to arrive at Figure 8.2. It reverses the roles of \(\mu\) and \(\bar{x}\). We’ll return to this point after we work through the mechanics of an example.

Example 8.2 : What is the 95\(\%\) confidence interval for student age if the population \(\sigma\) is 2 years, sample \(n = 50\), \(\bar{x} = 23.2\)?
Solution : So \({\cal{C}} = 0.95\). First write down the formula prescription so you can see with numbers you need:
\[ \bar{x} - E < \mu < \bar{x} + E \mbox{\hspace{2em}where\hspace{2em}} E = z_{95\%} \frac{\sigma}{\sqrt{n}}. \]First determine . With the tables in the Appendices, there are two ways to do this. The first way is to use the Standard Normal Distribution Table noting that we need the \(z\) associated with a table area of
. Using the table backwards we find \(z_{\cal{C}} = 1.96\). The second way, the recommended way especially during exams, is to use the t Distribution Table. Simply find the column for the 95\(\%\) confidence level and read the \(z\) from the last line of the table. We quickly find \(z_{95\%} = 1.960\).
Either way we now find
\[ E = 1.96( \frac{2}{\sqrt{50}}) = 0.6\]so
\[\begin{eqnarray*} \bar{x} - E &< \mu <& \bar{x} + E\\ 23.2 - 0.6 &< \mu <& 23.2 + 0.6 \\ 22.6 &< \mu <& 23.8 \end{eqnarray*}\]with 95\(\%\) confidence.
▢
- Because of this issue about probabilities of hypotheses, many prefer to say "confidence" and not probability. But we will learn enough about Bayesian probability to say "probability". ↵