20.5: Consequences
- Page ID
- 57813
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\( \newcommand{\dsum}{\displaystyle\sum\limits} \)
\( \newcommand{\dint}{\displaystyle\int\limits} \)
\( \newcommand{\dlim}{\displaystyle\lim\limits} \)
\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)
( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\id}{\mathrm{id}}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\kernel}{\mathrm{null}\,}\)
\( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\)
\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\)
\( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)
\( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)
\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)
\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vectorC}[1]{\textbf{#1}} \)
\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)
\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)
\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\(\newcommand{\longvect}{\overrightarrow}\)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)The formal definitions of sample statistics, their distributions, and their relationship to population parameters are not merely academic distinctions — they are the very engine of statistical inference. From these foundational ideas flow consequences that govern how we learn from data. These consequences directly underpin the theory, diagnostics, and interpretation of linear models.
This section explores the most important of these implications, including the behavior of estimators, the decomposition of variance, the derivation of standard errors, and the theoretical justification for key assumptions. Understanding these logical outcomes is what transforms simple calculation into reliable and defensible statistical analysis.
✦•················• ✦ •··················•✦
Basic Matrix Results
First, when taking the transpose of a product, you switch the order of the multiplication:
\begin{equation}
\left( \mathbf{A}\mathbf{B}\right)^\prime = \mathbf{B}^\prime \mathbf{A}^\prime
\end{equation}
A similar result holds with taking the transpose of inverses. The only difference is that all three inverses must exist:
\begin{equation}
\left( \mathbf{A}\mathbf{B}\right)^{-1} = \mathbf{B}^{-1} \mathbf{A}^{-1}
\end{equation}
For any matrix \(\mathbf{X}\), the matrix \(\mathbf{X}^\prime \mathbf{X}\) is symmetric.
Proof.
I leave the proof as an exercise for you.
If you know the determinant of a matrix, you can easily calculate the determinant of a scalar multiple of that matrix.
If \( c \in \) ℝ and \(\mathbf{A} \in \mathcal{M}_n\), then \(\det c\mathbf{A} = c^n \det \mathbf{A}\).
There is a similar result with the determinant of a trace of a matrix.
If \( c \in \) ℝ and \(\mathbf{A} \in \mathcal{M}_n\), then \(\text{tr}\ c\mathbf{A} = c\ \text{tr}\ \mathbf{A}\).
Positive Definite Matrices
There are a lot of interesting properties of positive definite (pd) matrices. So, let us break these into a separate subsection.
The diagonal elements of a pd matrix are all positive. That is, let \(\mathbf{A} \in \mathscr{M}_{n}\) be positive definite, then \(a_{ii} > 0,\ \forall i \in \set{1, 2, \ldots, n}\).
Proof.
This can easily be shown by letting the \(\mathbf{q}^\prime\) row vector be \(\mathbf{e_i}\). The quadratic form \(\mathbf{q}^\prime \mathbf{A} \mathbf{q}\) would therefore equal the diagonal element at position \(i\). Since \(\mathbf{A}\) is positive definite, that element must be greater than 0.
\(\blacksquare\)
Note that the converse is not true, in general. Just because the diagonal elements are all positive does not mean that the matrix is positive definite. For an example, note
\begin{equation}
\mathbf{A} = \left[ \begin{matrix} 1&1\\ 1&1\\ \end{matrix} \right]
\end{equation}
is not positive definite. To see this, its determinant is 0.
A few additional results:
- The inverse of a pd matrix is also pd. That is, if \(\mathbf{A}\) is positive definite, then so is \(\mathbf{A}^{-1}\).
- Since the determinant of a pd matrix is positive, all of the eigenvalues are positive. And, since all of the diagonal elements of a pd matrix are positive, then the trace is positive. Since the trace is used to calculate the degrees of freedom, the matrix must be positive definite.
- If \(\mathbf{X}\) is a full rank matrix, even if not square, \(\mathbf{X}^\prime\mathbf{X}\) is positive definite.
- And, most importantly, the covariance matrix is positive definite if the design matrix, \(\mathbf{X}\), is full rank. Otherwise, it is positive semi-definite and has a determinant of zero.


