11.3: The Foundations of Repeated-Measures ANOVA Formula
- Page ID
- 50155
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)
( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\id}{\mathrm{id}}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\kernel}{\mathrm{null}\,}\)
\( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\)
\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\)
\( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)
\( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)
\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)
\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vectorC}[1]{\textbf{#1}} \)
\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)
\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)
\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)The omnibus test is performed using the ANOVA F-formula. When we test a hypothesis using repeated measures ANOVA, it is because we expect that a substantial proportion of variability will be attributed to the conditions relative to how much is attributed to random error. Therefore, the focus of the computations is how much variation is attributed to the conditions relative to how much is attributed to random error. The repeated measures ANOVA formula uses sums of squares to measure variability. Though the focus is on two sources, the variability can come from three important sources: different conditions, differences among participants, and random error. Therefore, the sums of squares must be parsed out and attributed to these three different sources so that the proportions attributed to conditions vs. random error can be compared.
The formula for repeated measures ANOVA can be written as follows:
\[F=\dfrac{M S S_{\text {between }}}{M S S_{\text {error }}} \nonumber \]
The formula can be understood as follows:
\[F=\dfrac{\text { variation between conditions }}{\text { error variance }} \nonumber \]
This is similar to the independent groups ANOVA (see Chapter 10) but with one very important difference: the denominator for repeated measures ANOVA is a computation of random error instead of the sum of squares within (\(S S_w\)). The numerator for repeated-measures ANOVA focuses on differences between conditions and the denominator focuses on random, unsystematic differences observed, also known as the error variance.
The obtained value, \(F\), tells us the ratio of difference between conditions relative to random error. Another way to say this is that it tells us how different the means for conditions are from one another after removing individual variation that is consistent and taking into account random variation that is not attributed to the conditions nor to the individuals. The goal is to see how much of the difference observed is attributed to the conditions relative to how much is attributed to random, unsystematic variation alone. The variation of interest between conditions can be thought of as systematic differences caused by the independent variable (i.e. the conditions) rather than individual differences (i.e. individual variation) or random error. The portion of variation between and attributed to the conditions is known as the treatment effects. However, when we observe variation between conditions, it will include not only the treatment effects but also some random, unsystematic variation that can occur known as random error. Because both treatment effects and random error exist between groups, random error must be accounted for in the denominator of the ANOVA formula. To summarize, we can understand the formula’s main construction as follows:
\[F=\dfrac{\text { variation between conditions }}{\text { error variance }}=\dfrac{\text { treatment effects }+ \text { random error }}{\text { random error }} \nonumber \]
The final \(F\)-value for repeated-measures ANOVA, thus, indicates the ratio of variability that is between conditions relative to that which is attributed to random error. The lowest an \(F\)-value can be for an ANOVA is 1.00. When \(F\) = 1.00, all of the observed variability is attributed to random error and not to the different conditions. This is in keeping with the null hypothesis. However, the greater the \(F\)-value, the greater the ratio of variability that is uniquely attributed to the conditions (i.e. the IV) rather than to random differences.
When there is no treatment effect, the numerator and denominator will be equal and the result will be \(F\) = 1.00. Thus, the lowest \(F\)-value possible is 1.00 which occurs when no treatment effect is observed. However, the greater the treatment effect, the greater the \(F\)-value will be.
When this formula is written out to summarize the computational elements, it is written as follows:
\[F=\dfrac{M S S_b}{M S S_e}=\dfrac{S S_{b \div} d f_b}{S S_{e \div} d f_e} \nonumber \]
Keep in mind that between calculations for repeated measures ANOVA refer to differences between conditions (i.e. differences between groups of data for the same participants) not differences between different groups of participants. Therefore, in repeated measures ANOVA there can be differences that occur among the participants within each group but there are no differences that occur from having different people in each group.
Sources of Variation in Repeated Measures ANOVA
Let’s take a moment to understand the different sources of variation for repeated-measures ANOVA before looking at the way things are calculated so we can understand the rationale for those calculations. There are three things to which variation can be attributed and estimated in ANOVA:
- Treatment Effects (\(S S_{\text {treatment }}\)): variation attributed to the independent variable (i.e. variation attributed to exposure to different conditions);
- Individual Variation (\(S S_{\text {participants }}\) or \(S S_p\)): variation attributed to pre-existing, systematic differences among individuals that they bring with them to each condition;
- Error Variance (\(S S_{\text {error }}\) or \(S S_e\); also known as Residual Error): unexplained variation not accounted for or attributed to either of the aforementioned forms.
It is important to note that the proportions of variance attributed to each of these sources cannot all be directly computed. Instead, some sources can be more directly estimated while others must be deduced by partitioning them out during computations.
Parsing out Sources of Variation
For repeated-measures ANOVA, we must parse out the variation attributed to several sources before the F-formula can be used. This can be done using four important computations of variability:
- \(S_{\text {betweenconditions }}\left(S S_b\right)\): This includes variation between groups that is attributed to the conditions/treatment (\(S S_t\)) and any random error (\(S S_e\)) together. This can also be referred to as \(SS\) between treatments or simply \(SS\) between.
- \(S S_{\text {within }}\left(S S_w\right)\): This includes variation that exists within group caused by both participant differences that are systematic (\(S S_p\)) and random differences (error) which is non-systematic (\(S S_e\)).
- If you compute and sum \(S S_p\) and \(S S_e\), you will get \(S S_w\)
- \(S_{\text {participants }}\left(S S_p\right)\): This includes variation due to differences between participants that they bring with them to each condition.
- This form of error is part of \(S S_w\) but is not part of \(S S_b\).
- \(S S_{\text {error }}\left(S S_e\right)\): This includes variation that is random and is not attributed to the treatment (i.e. conditions) nor to differences between participants.
- \(S S_e\) exists within both \(S S_b\) and \(S S_w\).
\(S S_b\) is short for the “sum of squares between conditions.” \(S S_b\) is the portion of variance attributed to the independent variable (i.e. the different conditions) plus random error (\(S S_e\)). \(S S_w\) is short for “sum of squares within conditions.” There are two sources of variability within conditions that together comprise the \(S S_w: 1\). The sum of squares for participants (\(S S_p\)) which is the portion of variability attributed to pre-existing differences among people and 2. The sum of squares error (\(S S_e\)) which is the variability that is attributed to randomness. Notice that both \(S S_b\) and \(S S_w\) include random error but only \(S S_w\) includes error attributed to differences between people. Therefore, we want to remove the error attributed to pre-existing differences among participants (\(S S_p\)) from the \(S S_w\) when creating our denominator; this is so that both the numerator and denominator will account for the same sources of error (namely, just the \(S S_e\)).
The sum of squares total (\(S S_T\)) can also be computed but is not a necessity so we will only briefly review it here. \(S S_T\) includes variability from all three sources (i.e. treatment effects, individual variation, and error variance) without partitioning them. \(S S_T\) can be computed directly or found by summing \(S S_b\) and \(S S_w\) because these together contain the variability for all three sources.
Reading Review 11.2
- What two sources of variance together make up the \(SS\) between?
- What two sources of variance together make up the \(SS\) within?
- What source of variance is included in both the \(SS\) between and the \(SS\) within?
- Which \(SS\) includes any variance attributed to the independent variable?