6.2: Mean and Standard Deviation of Sampling Distributions
- Page ID
- 58911
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\( \newcommand{\dsum}{\displaystyle\sum\limits} \)
\( \newcommand{\dint}{\displaystyle\int\limits} \)
\( \newcommand{\dlim}{\displaystyle\lim\limits} \)
\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)
( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\id}{\mathrm{id}}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\kernel}{\mathrm{null}\,}\)
\( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\)
\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\)
\( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)
\( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)
\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)
\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vectorC}[1]{\textbf{#1}} \)
\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)
\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)
\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\(\newcommand{\longvect}{\overrightarrow}\)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)In the last section, we saw how we can build a sampling distribution: take many samples from the same population, calculate a statistic (like the mean) for each sample, and plot those values.
Now we focus more carefully on two key questions:
- What is the center of that sampling distribution?
- How much do the sample statistics typically vary? That is, how wide is the distribution?
To answer these, we revisit two ideas from earlier chapters: mean and standard deviation. But this time, we’re not looking at raw data, we’re looking at a distribution of statistics from many samples. And that leads us to two new terms: unbiased estimators and standard error.
Review: Mean and Standard Deviation
- Mean: The average of a set of values; center of a distribution
- Standard Deviation (SD): A measure of how spread out values are around the mean on average
In a sampling distribution, the values we’re working with are not individual data points, they’re sample statistics, like sample means from different samples. So when we talk about “the mean of all the sample means” or their spread, we’re talking about the sampling distribution of the statistic.
Definition: Mean of the Sampling Distribution
The mean of the sampling distribution is the mean of all of the sample statistics from all possible samples.
For sample means, the mean of the sampling distribution is equal to the population mean:
\[ \mu_{\bar{x}} = \mu \]
For sample proportions, the mean of the sampling distribution is equal to the population proportion:
\[ \mu_{\hat{p}} = p \]
The sample mean and sample proportion are both called unbiased estimators since the mean of all the sample means matches the population mean and the mean of all the sample proportions matches the population proportion. The statistics "target" the population parameter. This means that on average the sample mean (and proportion) is correct, even though individual samples may vary. This is NOT true of other sample statistics like the range or median. The mean of all sample ranges may NOT match the population range.
This concept, known as unbiasedness, is important: it means that even though sample statistics fluctuate, the process doesn’t consistently overestimate or underestimate the population value. Remember, not all statistics are unbiased!
Definition: Standard Error (SE)
The standard error is the standard deviation of a sampling distribution. It measures how much the sample statistic varies from sample to sample. We often abbreviate standard error as SE. However, it may also be written like \( \sigma_{\bar{x}} \) for sample means, \( \sigma_{\hat{p}} \) for sample proportions, or \( \sigma_{s} \) for sample standard deviations.
As an example, for sample means the standard error is:
\[ \sigma_{\bar{x}} = \frac{\sigma}{\sqrt{n}} \]
where \( n\) is the sample size and \(\sigma\) is the population standard deviation.
Note the \(n\) in the denominator of the fraction. As sample size increases, the standard error shrinks. Larger samples give more stable sample means. This is generally true for all sampling distributions, not just sample means, but this particular formula \(\frac{\sigma}{\sqrt{n}}\) is specific to sample means. You can see an example of this plotted below.
Interpreting the Horizontal Axis
When we draw or simulate a sampling distribution (like we did in Section 6.1), the x-axis now shows sample statistics, not individual data points.
The center of the distribution is \( \mu_{\bar{x}} \) or \(\mu_{\hat{p}}\) or whichever statistic you're using.
The spread of the distribution is \( \sigma_{\bar{x}} \), the standard error.
Nightly Sleep Amounts
Suppose that the average college student gets 6.8 hours of sleep per night with a population standard deviation is 1.1 hours. If we were studying student sleep habits, we would not know these actual population values! Instead, we would take a sample to determine an estimate.
We take a random sample of 25 students and compute this sample mean. What would the center and spread of the sampling distribution of sample means look like?
Mean: \( \mu_{\bar{x}} = 6.8 \) hours. The average sample mean is just the pre-existing population mean.
Standard Error: \( \sigma_{\bar{x}} = \frac{1.1}{\sqrt{25}} = \frac{1.1}{5} = 0.22 \text{ hours} \). We have to calculate the standard error by dividing by the sample size. This is because as the sample size goes up, the likelihood of having a sample mean further from the population mean decreases.
Conclusion: Most of the sample means (not individual sleep values!) would fall within 0.22 hours of 6.8 on average. Using our rule of thumb, we can say that with perfectly random sampling, there is a 68% of getting a sample within one standard deviation, or between 6.58 and 7.02 hours.
A Larger Sample
Let's suppose we change our minds and double our sample size for our sleep study. We now have \( n = 50 \) and can recompute the standard error.
\( \sigma_{\bar{x}} = \frac{1.1}{\sqrt{50}} = \frac{1.1}{7.07} = 0.156\)
Note that with a doubling of sample size, the standard error does not get twice as small. This is still helpful however, as our 68% area would now span between 6.64 and 6.96, which is closer to the true mean.
Connecting to Chapter 5
Back in Chapter 5, we plotted raw data and observed shapes and spread. That was the distribution of the data.
Here in Chapter 6, we are plotting statistic values from many samples. It’s a shift in thinking:
- Before: Each point = one person’s data value
- Now: Each point = one sample's mean (or another statistic)
There are two major points to consider however:
- In reality, we don't know the population values and therefore cannot calculate our sample distribution. Instead, we have a sample that came from that distribution.
- We don't take multiple samples to build the distribution. It's a tool that is used to assess a single sample that comes from the distribution.


