Skip to main content
Statistics LibreTexts

11.5.1: Pairwise Comparison Post Hoc Tests for Critical Values of Mean Differences

  • Page ID
    22115
  • \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

    \( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

    ( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\id}{\mathrm{id}}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\kernel}{\mathrm{null}\,}\)

    \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\)

    \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\)

    \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

    \( \newcommand{\vectorA}[1]{\vec{#1}}      % arrow\)

    \( \newcommand{\vectorAt}[1]{\vec{\text{#1}}}      % arrow\)

    \( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vectorC}[1]{\textbf{#1}} \)

    \( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

    \( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

    \( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

    \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

    \(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)

    As just discussed , a post hoc test is used only after we find a statistically significant (reject the null hypothesis) result and need to determine where our differences truly came from. This implies that when the null hypothesis is retained, you do not need to conduct pairwise comparisons; if there's no differences between the means, why would you look for where a mean difference is? The term “post hoc” comes from the Latin for “after the event”.

    Mean Differences

    The next set of post-hoc analyses compare the difference between each pair of means, then compares that to a critical value. Let's start by determining the mean differences. Table \(\PageIndex{1}\) shows the mean test scores for the three IV levels in our job applicant scenario.

    Table \(\PageIndex{1}\)- Average Test Scores by Job Applicant Degree
    IV Levels Average Test Score (\(\overline{X}\))
    No Degree 43.7
    Related Degree 84.3
    Unrelated Degree 63.2

    Using the average test scores for each level of the IV (Degree), we could find the difference between each pair of means:

    \[ \overline{X}_{N} - \overline{X}_{R} = 43.7 – 84.3 = -40.6 \nonumber\]

    \[ \overline{X}_{N} - \overline{X}_{U} = 43.7 – 63.2 = -19.5 \nonumber\]

    \[ \overline{X}_{R} - \overline{X}_{U} = 84.3 – 63.2 = 21.1 \nonumber\]

    We could avoid negative answers if we always put the bigger number first, but Dr. MO likes to follow the order from the research hypotheses. As with all critical values, we use the absolute value of the calculated statistic anyway, so the big issue is to remember is to make sure that you subtract each mean from each other mean.

    That’s it on mean differences, let’s now learn about some ways to calculate a critical value to compare these mean differences to; as always:

    \(Critical < |Calculated| =\) Reject null \(=\) means are different \(= p<.05\)

    \(Critical > |Calculated| =\) Retain null \(=\) means are similar \(= p>.05\)

    Pairwise Comparisons

    For this type of post-hoc analysis, you compare each of these mean differences (that you just calculated by subtracting one mean from another mean) to a critical value. What should you do if the calculated mean difference is further from zero (bigger) than the critical value? Yep, you reject the null hypothesis and say that those two means are different from each other! Sound familiar? It’s just like when using a critical z, critical t, or critical F.

    As with converting the raw p-values, the critical mean difference can be computed in different ways. Tukey’s Honestly Significant Difference will be discussed here, but just know that there are other types of pairwise comparison tests that statistical software can complete with ease.

    Pairwise Comparison Steps:

    1. Compute a mean difference for each pair of variables.
    2. Find the critical mean difference.
    3. Compare each calculated mean difference to the critical mean.
    4. Decide whether to retain or reject the null hypothesis for that pair of means.

    Tukey's HSD (Honestly Significant Difference)

    Tukey’s HSD corrects for the alpha inflation caused by doing a bunch of statistical tests by correcting for the probability of a Type I error so that all the pairwise comparisons combined is p<.05, sorta like the Bonferroni that we just discussed. Tukey’s makes it so that each individual pairwise comparison has a much smaller p-value.

    Formula:

    \[ HSD = q * \sqrt{\dfrac{MSw}{n_{group}}} \nonumber \]

    There are a couple things to know about this formula. First, the q is found in yet another table. You probably won’t need this too much because statistical software tells you which means are different from which other means, but you can find a table of q-values at this Real-Statistics.com webpage: https://www.real-statistics.com/statistics-tables/studentized-range-q-table/; make sure that you use the Alpha = 0.05 set of tables!

    To find the q-value in the table, you find the column with the total number of groups in the analysis (k) on the top, then find the Degrees of Freedom for the denominator (\(df_{W}\)) on the side. Remember, there are different tables for the different alpha levels so don’t just use the first table.

    Example \(\PageIndex{1}\)

    What would the critical value (Tukey’s HSD) be for our job applicant scenario?

    Solution

    Let’s first find the q-value. We know that we have 3 groups (k = 3), and 30 participants. The \(df_{W}\) is calculated as N – k, so that’s 30 – 3 = 27. According to the Alpha = 0.05 table at the table of q-values, q = 3.506.
    Now we can use the q-value we just found and information from the ANOVA Summary Table to complete this formula:
    \[ HSD = q * \left( \sqrt{\dfrac{MSw}{n_{group}}}\right) \nonumber \]
    \[ HSD = 3.506 * \left( \sqrt{\dfrac{111.85}{10}}\right) \nonumber \]
    \[ HSD = 3.506 * \left( \sqrt{11.185}\right) \nonumber \]
    \[ HSD = 3.506 * \left( 3.34\right) \nonumber \]
    \[ HSD = 11.73 \nonumber \]

    Okay, but what do we do with this number? Tukey’s HSD is a critical value, so if any of your mean differences are bigger than this critical value of 3.82, then the null hypothesis is rejected and that set of means is statistically significantly different from each other. If the critical value (Tukey’s HSD that you just calculated) is bigger than any of the differences between the means, then you retain the null hypothesis and say that they are not different enough to think that they are from different populations.

    \(Critical < |Calculated| =\) Reject null \(=\) means are different \(= p<.05\)

    \(Critical > |Calculated| =\) Retain null \(=\) means are similar \(= p>.05\)

    Let’s look at that in Table \(\PageIndex{2}\).

    Table \(\PageIndex{2}\)- Interpreting Pairwise Comparison Results
    Means Compared Mean Difference Is Tukey’s HSD of 11.73 Smaller than the Absolute Value of the Mean Difference? Reject or Retain the Null Hypothesis? Are the Means Similar or Different?
    \[ \overline{X}_{N} - \overline{X}_{R} = 43.7 – 84.3 \nonumber\] -40.6 Yes Reject the null hypothesis. These two means are different from each other.
    \[ \overline{X}_{N} - \overline{X}_{U} = 43.7 – 63.2 \nonumber\] -19.5 Yes Reject the null hypothesis. These two means are different from each other.
    \[ \overline{X}_{R} - \overline{X}_{U} = 84.3 – 63.2 \nonumber\] 21.1 Yes Reject the null hypothesis. These two means are different from each other.

    But what does it mean?

    To complete this analysis of the job applicant scenario, we need to go back to the research hypothesis. We had said that the applicants with No Degree will have a lower average test score than those with a Related Degree, but those with No Degree will have a similar average test score to those with an Unrelated degree. The average test score for those with a Related Degree will also have a higher average test score compared to those with an Unrelated Degree. In symbols, this looks like:

    • \( \overline{X}_{N} < \overline{X}_{R} \)
    • \( \overline{X}_{N} = \overline{X}_{U} \)
    • \( \overline{X}_{R} > \overline{X}_{U} \)

    Now, let’s look at our mean differences. All of them were statistically significantly different, which means that our research hypothesis was partially supported.

    • RH1: \( \overline{X}_{N} < \overline{X}_{R} \) = Supported; those without a degree scored statistically significantly lower than those with a related degree.
    • RH2: \( \overline{X}_{N} = \overline{X}_{U} \) = Not supported; the research hypothesis was that the group without a degree would have average test scores that were similar to the group with an unrelated degree, but those with an unrelated degree actually scored higher than those with no degree.
    • RH3: \( \overline{X}_{R} > \overline{X}_{U} \) = Supported; those with a related degree did score higher than those with an unrelated degree.

    And that leads us to our final write-up!

    Write-Up

    Exercise \(\PageIndex{1}\)

    Write up a conclusion to the job applicant scenario.

    Answer

    Look at your conclusion. Did you include the four components that should be in all write-ups?

    • The statistical test is preceded by the descriptive statistics (means).
    • The description tells you what the research hypothesis being tested is.
    • A "statistical sentence" showing the results is included.
    • The results are interpreted in relation to the research hypothesis.

    Look at Dr. MO’s conclusion.

    The research hypothesis that the applicants with No Degree will have a lower average test score than those with a Related Degree, but will have a similar average test score to those with an unrelated degree, and that the average test score for those with a Related Degree will also have a higher average test score compared to those with an Unrelated Degree was partially supported (F(2,27) = 36.86, p < 0.05. Those with a Related Degree (M = 84.3) did score higher than both those with an Unrelated Degree (M = 63.2) and those with No Degree (M = 43.7). However, it was hypothesized that those with an Unrelated Degree would have similar average test scores than those with No Degree, when those with an Unrelated Degree actually scored significantly higher than those with No Degree.

    Note

    Did Dr. MO include the four components that should be in all write-ups?

    Make sure to include the statistical sentence for the ANOVA with both Degrees of Freedom (with the smaller one from the numerator first), but you don’t have to include all of that information for each pairwise comparison of the mean differences.

    Summary

    There are many more post hoc tests than just the ones discussed here, and they all approach the task in different ways, with some being more conservative and others being more powerful. In general, though, they will give highly similar answers. What is important here is to be able to interpret a post hoc analysis.

    So, that’s it! You’ve learned a Between Groups ANOVA and pairwise comparisons to test the null hypothesis! Let’s try one full example next!


    This page titled 11.5.1: Pairwise Comparison Post Hoc Tests for Critical Values of Mean Differences is shared under a CC BY-NC-SA 4.0 license and was authored, remixed, and/or curated by Michelle Oja.