# 15.5: Alternating Renewal Processes

- Page ID
- 10281

\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

\( \newcommand{\Span}{\mathrm{span}}\)

\( \newcommand{\id}{\mathrm{id}}\)

\( \newcommand{\Span}{\mathrm{span}}\)

\( \newcommand{\kernel}{\mathrm{null}\,}\)

\( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\)

\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\)

\( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

\( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)

\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)

\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vectorC}[1]{\textbf{#1}} \)

\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

\(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)## Basic Theory

### Preliminaries

An alternating renewal process models a system that, over time, alternates between two states, which we denote by 1 and 0 (so the system starts in state 1). Generically, we can imagine a device that, over time, alternates between *on* and *off* states. Specializing further, suppose that a device operates until it fails, and then is replaced with an identical device, which in turn operates until failure and is replaced, and so forth. In this setting, the times that the device is functioning correspond to the *on* state, while the replacement times correspond to the *off* state. (The device might actually be *repaired* rather than *replaced*, as long as the repair returns the device to pristine, new condition.) The basic assumption is that the pairs of random times successively spent in the two states form an independent, identically distributed sequence. Clearly the model of a system alternating between two states is basic and important, but moreover, such alternating processes are often found embedded in other stochastic processes.

Let's set up the mathematical notation. Let \( \bs{U} = (U_1, U_2, \ldots) \) denote the successive lengths of time that the system is in state 1, and let \( \bs{V} = (V_1, V_2, \ldots)\) the successive lengths of time that the system is in state 0. So to be clear, the system starts in state 1 and remains in that state for a period of time \( U_1 \), then goes to state 0 and stays in this state for a period of time \( V_1 \), then back to state 1 for a period of time \( U_2 \), and so forth. Our basic assumption is that \( \bs{W} = \left((U_1, V_1), (U_2, V_2), \ldots\right) \) is an independent, identically distributed sequence. It follows that \( \bs{U} \) and \( \bs{V} \) each are independent, identically distributed sequences, but \( \bs{U} \) and \( \bs{V} \) might well be dependent. In fact, \( V_n \) might be a function of \( U_n \) for \( n \in \N_+ \). Let \( \mu = \E(U) \) denote the mean of a generic time period \( U \) in state 1 and let \( \nu = \E(V) \) denote the mean of a generic time period \( V \) in state 0. Let \( G \) denote the distribution function of a time period \( U \) in state 1, and as usual, let \( G^c = 1 - G \) denote the right distribution function (or reliability function) of \( U \).

Clearly it's natural to consider returns to state 1 as the arrivals in a renewal process. Thus, let \( X_n = U_n + V_n \) for \( n \in \N_+ \) and consider the renewal process with interarrival times \( \bs{X} = (X_1, X_2, \ldots) \). Clearly this makes sense, since \( \bs{X} \) is an independent, identically distributed sequence of nonnegative variables. For the most part, we will use our usual notation for a renewal process, so the common distribution function of \( X_n = U_n + V_n \) is denoted by \( F \), the arrival time process is \( \bs{T} = (T_0, T_1, \ldots) \), the counting process is \( \{N_t: t \in [0, \infty)\} \), and the renewal function is \( M \). But note that the mean interarrival time is now \( \mu + \nu \).

The renwal process associated with \( \bs{W} = ((U_1, V_1), (U_2, V_2), \dots) \) as constructed above is known as an alternating renewal process.

### The State Process

Our interest is the state \( I_t \) of the system at time \( t \in [0, \infty) \), so \( \bs{I} = \left\{I_t: t \in [0, \infty)\right\} \) is a stochastic process with state space \( \{0, 1\} \). Clearly the stochastic processes \( \bs{W} \) and \( \bs{I} \) are equivalent in the sense that we can recover one from the other. Let \( p(t) = \P(I_t = 1) \), the probability that the device is on at time \( t \in [0, \infty) \). Our first main result is a renewal equation for the function \( p \).

The function \( p \) satisfies the renewal equation \( p = G^c + p * F \) and hence \( p = G^c + G^c * M \).

## Proof

By now, the approach should be clear—we're going to condition on the first arrival \( X_1 \): \[ \P(I_t = 1) = \P(I_t = 1, X_1 \gt t) + \P(I_t = 1, X_1 \le t) = \P(I_t = 1, X_1 \gt t) + \int_0^t \P(I_t = 1 \mid X_1 = s) \, dF(s) \] But \( \{I_t = 1, X_1 \gt t\} = \{U_1 \gt t) \) so \( \P(I_t = 1, X_1 \gt t) = \P(U_1 \gt t) = G^c(t) \). By the fundamental renewal property (the process restarts, independently of the past, at each arrival) \( \P(I_t = 1 \mid X_1 = s) = p(t - s) \) for \( s \le t \). Hence we have \[ p(t) = G^c(t) + \int_0^t p(t - s) \, dF(s), \quad t \in [0, \infty) \] or equivalently, \( p = G^c + p * F \). By the fundamental theorem on renewal equations, the solution is \( p = G^c + G^c * M \), so \( p(t) = G^c(t) + \int_0^t G^c(t - s) \, dM(s), \quad t \in [0, \infty) \)

We can now apply the key renewal theorem to get the asymptotic behavior of \( p \).

If the renewal process is non-arithmetic, then \[ p(t) \to \frac{\mu}{\mu + \nu} \text{ as } t \to \infty\]

## Proof

From the result above, \( p = G^c + G^c * m \). First, \( G^c(t) \to 0 \) as \( t \to \infty \) as a basic property of the right distribution function. Next, by the key renewal theorem, \[ (G^c * M)(t) \to \frac{1}{\mu + \nu} \int_0^\infty G^c(s) \, ds \text{ as } t \to \infty \] But by another basic property of the right distribution function, \( \int_0^\infty G^c(s) \, ds = \mu \).

Thus, the limiting probability that the system is on is simply the ratio of the mean of an *on* period to the mean of an *on-off* period. It follows, of course, that \[ \P(I_t = 0) = 1 - p(t) \to \frac{\nu}{\mu + \nu} \text{ as } t \to \infty\] so in particular, the fact that the system starts in the on state makes no difference in the limit. We will return to the asymptotic behavior of the alternating renewal process in the next section on renewal reward processes.

## Applications and Special Cases

With a clever definition of *on* and *off*, many stochastic processes can be turned into alternating renewal processes, leading in turn to interesting limits, via the basic limit theorem above.

### Age Processes

The last remark applies in particular to the age processes of a standard renewal process. So, suppose that we have a renewal process with interarrival sequence \( \bs{X} \), arrival sequence \( \bs{T} \), and counting process \( \bs{N} \). As usual, let \( \mu \) denote the mean and \( F\) the probability distribution function of an interarrival time, and let \( F^c = 1 - F \) denote the right distribution function (or reliability function).

For \( t \in [0, \infty) \), recall that the current life, remaining life and total life at time \( t \) are \[ C_t = t - T_{N_t}, \quad R_t = T_{N_t + 1} - t, \quad L_t = C_t + R_t = T_{N_t+1} - T_{N_t} = X_{N_t + 1} \] respectively. In the usual terminology of reliability, \( C_t \) is the age of the device in service at time \( t \), \( R_t \) is the time remaining until this device fails, and \( L_t \) is total life of the device. We will use limit theorem above to derive the limiting distributions these age processes. The limiting distributions were obtained earlier, in the section on renewal limit theorems, by a direct application of the key renewal theorem. So the results are not new, but the method of proof is interesting.

If the renewal process is non-arithmetic then \[ \lim_{t \to \infty} \P(C_t \le x) = \lim_{t \to \infty} \P(R_t \le x) = \frac{1}{\mu} \int_0^x F^c(y) \, dy, \quad x \in [0, \infty) \]

## Proof

Fix \( x \in [0, \infty) \). For the current life limit, define the *on* period corresponding to the interarrival time \( X_n \) to be \( U_n = \min\{X_n, x\} \) for \( n \in \N_+ \), so that the *off* period is \(V_n = X_n - \min\{X_n, x\} \). Note that the system is on at time \( t \in [0, \infty) \) if and only if \( C_t \le x \), and hence \( p(t) = \P(C_t \le x) \). It follows from the limit theorem above that \[ \P(C_t \le x) \to \frac{\E[\min\{X, x\}]}{\mu} \text{ as } t \to \infty \] where \( X \) is a generic interarrival time. But \[G^c(y) = \P[\min\{X, x\} \gt y] = \P(X \gt y) \bs{1}(x \gt y) = F^c(y) \bs{1}(x \gt y), \quad y \in [0, \infty)\] Hence \[ \E[\min(X, x)] = \int_0^\infty G^c(y) \, dy = \int_0^x F^c(y) \, dy \] For the remaining life limit we reverse the on-off periods. Thus, define the *on* period corresponding to the interarrival time \( X_n \) to be \( U_n = X_n - \min\{X_n, x\} \) for \( n \in \N_+ \), so that the *off* period is \( V_n = \min\{X_n, x\} \). Note that the system is off at time \( t \) if and only if \( R_t \le x \), and hence \( 1 - p(t) = \P(R_t \le x) \). From the limit theorem above, \[ \P(R_t \le x) \to \frac{\E[\min\{X, x\}]}{\mu} \text{ as } t \to \infty \]

As we have noted before, the fact that the limiting distributions are the same is not surprising after a little thought. After a long time, the renewal process looks the same forward and backward in time, and reversing the arrow of time reverses the roles of current and remaining time.

If the renewal process is non-arithmetic then \[ \lim_{t \to \infty} \P(L_t \le x) = \frac{1}{\mu} \int_0^x y \, dF(y), \quad x \in [0, \infty) \]

## Proof

Fix \( x \in [0, \infty) \). For \( n \in \N_+ \), define the *on* period associated with interarrival time \( X_n \) by \( U_n = X_n \bs{1}(X_n \gt x) \). Of course, the *off* period corresponding to \( X_n \) is \(V_n = X_n - U_n \). Thus, each renewal period is either totally on or totally off, depending on whether or not the interarrival time is greater than \( x \). Note that the system is on at time \( t \in [0, \infty) \) if and only if \( L_t \gt x \), so from the basic limit theorem above, \[ \P(L_t \gt x) \to \frac{1}{\mu} \E[X \bs{1}(X \gt x)] \] where \( X \) is a generic interarrival time. But \[ \E[X \bs{1}(X \gt x)] = \int_x^\infty y \, dF(y)\]