Skip to main content
Statistics LibreTexts

1.6.2: Calculating z-scores

  • Page ID
  • \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

    \( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

    ( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\id}{\mathrm{id}}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\kernel}{\mathrm{null}\,}\)

    \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\)

    \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\)

    \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

    \( \newcommand{\vectorA}[1]{\vec{#1}}      % arrow\)

    \( \newcommand{\vectorAt}[1]{\vec{\text{#1}}}      % arrow\)

    \( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vectorC}[1]{\textbf{#1}} \)

    \( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

    \( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

    \( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

    \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

    \(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)

    A \(z\)-score is a standardized version of a raw score (\(x\)) that gives information about the relative location of that score within its distribution. The formula for converting a raw score from a sample into a \(z\)-score is:

    \[z=\dfrac{x-\overline{X}}{s} \nonumber \]

    As you can see, \(z\)-scores combine information about where the distribution is located (the mean/center) with how wide the distribution is (the standard deviation/spread) to interpret a raw score (\(x\)). Specifically, \(z\)-scores will tell us how far the score is away from the mean in units of standard deviations and in what direction.

    The value of a \(z\)-score has two parts: the sign (positive or negative) and the magnitude (the actual number). The sign of the \(z\)-score tells you in which half of the distribution the z-score falls: a positive sign (or no sign) indicates that the score is above the mean and on the right hand-side or upper end of the distribution, and a negative sign tells you the score is below the mean and on the left-hand side or lower end of the distribution. The magnitude of the number tells you, in units of standard deviations, how far away the score is from the center or mean. The magnitude can take on any value between negative and positive infinity, but for reasons we will see soon, they generally fall between -3 and 3.

    A \(z\)-score value of -1.0 tells us that this z-score is 1 standard deviation (because of the magnitude 1.0) below (because of the negative sign) the mean. Similarly, a \(z\)-score value of 1.0 tells us that this \(z\)-score is 1 standard deviation above the mean. Thus, these two scores are the same distance away from the mean but in opposite directions. A \(z\)-score of -2.5 is two-and-a-half standard deviations below the mean and is therefore farther from the center than both of the previous scores, and a \(z\)-score of 0.25 is closer than all of the ones before. For now, we will use a rough cut-off of 1.5 standard deviations in either direction as the difference between close scores (those within 1.5 standard deviations or between \(z\) = -1.5 and \(z\) = 1.5) and extreme scores (those farther than 1.5 standard deviations – below \(z\) = -1.5 or above \(z\) = 1.5).


    Raw Score to z-score

    We can convert raw scores into \(z\)-scores to get a better idea of where in the distribution those scores fall. Let’s say we get a score of 68 on an exam. We may be disappointed to have scored so low, but perhaps it was just a very hard exam. Having information about the distribution of all scores in the class would be helpful to put some perspective on ours. We find out that the class got an average score of 54 with a standard deviation of 8. To find out our relative location within this distribution, we simply convert our test score into a \(z\)-score.

    \[z=\dfrac{x-\overline{X}}{s} =\dfrac{68-54}{8} = 1.75 \nonumber \]

    We find that we are 1.75 standard deviations above the average, above our rough cut off for close and far. Suddenly our 68 is looking pretty good!

    Two symmetrical line graphs with a dashed line showing where 1.75 standard deviatons are in scores (upper line graph) and z-scores (lower line graph).
    Figure \(\PageIndex{1}\): Raw and standardized versions of a single score (CC-BY-NC-SA Foster et al. from An Introduction to Psychological Statistics)

    Figure \(\PageIndex{1}\) shows both the raw score and the \(z\)-score on their respective distributions. Notice that the red line indicating where each score lies is in the same relative spot for both. This is because transforming a raw score into a \(z\)-score does not change its relative location, it only makes it easier to know precisely where it is.

    \(Z\)-scores are also useful for comparing scores from different distributions. Let’s say we take the SAT and score 501 on both the math and critical reading sections. Does that mean we did equally well on both? Scores on the math portion are distributed normally with a mean of 511 and standard deviation of 120, so our \(z\)-score on the math section is

    \[z_{math}=\dfrac{501-511}{120}=-0.08 \nonumber \]

    which is just slightly below average (note that use of “math” as a subscript; subscripts are used when presenting multiple versions of the same statistic in order to know which one is which and have no bearing on the actual calculation). The critical reading section has a mean of 495 and standard deviation of 116, so

    \[z_{C R}=\frac{501-495}{116}=0.05 \nonumber \]

    So even though we were almost exactly average on both tests, we did a little bit better on the critical reading portion relative to other people.

    Finally, \(z\)-scores are incredibly useful if we need to combine information from different measures that are on different scales. Let’s say we give a set of employees a series of tests on things like job knowledge, personality, and leadership. We may want to combine these into a single score we can use to rate employees for development or promotion, but look what happens when we take the average of raw scores from different scales, as shown in Table \(\PageIndex{1}\):

    Table \(\PageIndex{1}\): Raw test scores on different scales (ranges in parentheses).
    Raw Scores Job Knowledge (0 – 100) Personality (1 –5) Leadership (1 – 5) Average
    Employee 1 98 4.2 1.1 34.43
    Employee 2 96 3.1 4.5 34.53
    Employee 3 97 2.9 3.6 34.50

    Because the job knowledge scores were so big and the scores were so similar, they overpowered the other scores and removed almost all variability in the average. However, if we standardize these scores into \(z\)-scores, our averages retain more variability and it is easier to assess differences between employees, as shown in Table \(\PageIndex{2}\).

    Table \(\PageIndex{2}\): Standardized scores.
    \(z\)-Scores Job Knowledge (0 – 100) Personality (1 –5) Leadership (1 – 5) Average
    Employee 1 1.00 1.14 -1.12 0.34
    Employee 2 -1.00 -0.43 0.81 -0.20
    Employee 3 0.00 -0.71 0.30 -0.14

    z-score to Raw Score

    Another convenient characteristic of \(z\)-scores is that they can be converted into any “scale” that we would like. Here, the term scale means how far apart the scores are (their spread) and where they are located (their central tendency). This can be very useful if we don’t want to work with negative numbers or if we have a specific range we would like to present. The formula for transforming \(z\) to \(x\) for a sample is:

    \[x=z s+\overline{X} \nonumber \]

    Notice that this is just a rearrangement of the original formulas for calculating \(z\) from raw scores. Dr. MO prefers to just use the original z-score formula and do algebra to figure out the raw score when she has a z-score because she doesn't want to have to remember more formulas, but others prefer to do less algebra and have more formulas; it's your choice!

    Let’s say we create a new measure of intelligence, and initial calibration finds that our scores have a mean of 40 and standard deviation of 7. Three people who have scores of 52, 43, and 34 want to know how well they did on the measure. We can convert their raw scores into \(z\)-scores:

    \[\begin{aligned} z &=\dfrac{52-40}{7}=1.71 \\ z &=\dfrac{43-40}{7}=0.43 \\ z &=\dfrac{34-40}{7}=-0.80 \end{aligned} \nonumber \]

    A problem is that these new \(z\)-scores aren’t exactly intuitive for many people. We can give people information about their relative location in the distribution (for instance, the first person scored well above average), or we can translate these \(z\) scores into the more familiar metric of IQ scores, which have a mean of 100 and standard deviation of 16:

    \[\begin{array}{l}{\mathrm{IQ}=1.71 * 16+100=127.36} \\ {\mathrm{IQ}=0.43 * 16+100=106.88} \\ {\mathrm{IQ}=-0.80 * 16+100=87.20}\end{array} \nonumber \]

    We would also likely round these values to 127, 107, and 87, respectively, for convenience.

    Contributors and Attributions

    This page titled 1.6.2: Calculating z-scores is shared under a CC BY-NC-SA 4.0 license and was authored, remixed, and/or curated by Michelle Oja.