Processing math: 45%
  • AR(1) Visualization
  • Mathematical Representation
    • Expectation
    • Autocovariance
    • Autocorrelation

AR(1) Visualization

The first-order autoregressive process, denoted AR(1), is εt=ρεt1+wt where wt is a strictly stationary and ergodic white noise process with 0 mean and variance σ2w.

White noise process with $\sigma=20$.

Fig. 1: White noise process with σ=20.

To illustrate the behavior of the AR(1) process, Fig. 2 plots two simulated AR(1) processes. Each is generated using the white noise process et displayed in Fig. 1.

The plot in Fig. 2(a) sets ρ=0.5 and the plot in Fig. 2(b) sets ρ=0.95.

Remarks

  • Fig. 2(b) is more smooth than Fig. 2(a).
  • The smoothing increases with ρ.
Simulated AR(1) processes with positive $\rho$. (a) $\rho=0.5$, (b) $\rho=0.95$. Each is generated useing the white noise process $w_t$ displayed in Fig. \@ref(fig:WN).

Fig. 2: Simulated AR(1) processes with positive ρ. (a) ρ=0.5, (b) ρ=0.95. Each is generated useing the white noise process wt displayed in Fig. 1.


We have seen the cases when ρ is positive, now let’s consider when ρ is negative. Fig. 3(a) shows an AR(1) process with ρ=0.5, and Fig. 3(a) shows an AR(1) process with ρ=0.95.

We see that the sample path is very choppy when ρ is negative. The different patterns for positive and negative ρ’s are due to their autocorrelation functions (ACFs).

Simulated AR(1) processes with negtive $\rho$. (a) $\rho=-0.5$, (b) $\rho=-0.95$. Each is generated useing the white noise process $w_t$ displayed in Fig. \@ref(fig:WN).

Fig. 3: Simulated AR(1) processes with negtive ρ. (a) ρ=0.5, (b) ρ=0.95. Each is generated useing the white noise process wt displayed in Fig. 1.


Mathematical Representation

Let’s formulate an AR(1) model as follows:

εt=ρεt1+wt where wt is a white noise series with mean zero and variance σ2w. We also assume |ρ|<1.

We can represent the AR(1) model as a linear combination of the innovations wt.

By iterating backwards k times, we get

εt=ρεt1+wt=ρ(ρεt2+wt1)+wt=ρ2εt2+ρwt1+wt=ρkεtk+k1j=0ρjwtj. This suggests that, by continuing to iterate backward, and provided that |ρ|<1 and sup, we can represent \varepsilon_t as a linear process given by

\color{#EE0000FF}{\varepsilon_t = \sum_{j=0}^\infty \rho^j \,w_{t-j}} \,.


Expectation

\varepsilon_t is stationary with mean zero.

E(\varepsilon_t) = \sum_{j=0}^\infty \rho^j \, E(w_{t-j})


Autocovariance

The autocovariance function of the AR(1) process is \begin{aligned} \gamma (h) &= \text{Cov}(\varepsilon_{t+h}, \varepsilon_t) \\ &= E(\varepsilon_{t+h}, \varepsilon_t) \\ &= E\left[\left(\sum_{j=0}^\infty \rho^j \,w_{t+h-j}\right) \left(\sum_{k=0}^\infty \rho^k \,w_{t-k}\right) \right] \\ &= \sum_{l=0}^{\infty} \rho^{h+l} \rho^l \sigma_w^2 \\ &= \sigma_w^2 \cdot \rho^{h} \cdot \sum_{l=0}^{\infty} \rho^{2l} \\ &= \frac{\sigma_w^2 \cdot \rho^{h} }{1-\rho^2}, \quad h>0 \,. \end{aligned} When h=0, \gamma(0) = \frac{\sigma_w^2}{1-\rho^2} is the variance of the process \text{Var}(\varepsilon_t).

Note that

  • \gamma(0) \ge |\gamma (h)| for all h. Maximum value at 0 lag.
  • \gamma (h) is symmetric, i.e., \gamma (-h) = \gamma (h)

Autocorrelation

The autocorrelation function (ACF) is given by

\rho(h) = \frac{\gamma (h)}{\gamma (0)} = \rho^h, which is simply the correlation between \varepsilon_{t+h} and \varepsilon_{t}\,.

Note that \rho(h) satisfies the recursion \rho(h) = \rho\cdot \rho(h-1) \,.

  • For \rho >0, \rho(h)=\rho^h>0 observations close together are positively correlated with each other. The larger the \rho, the larger the correlation.
  • For \rho <0, the sign of the ACF \rho(h)=\rho^h depends on the time interval.
    • When h is even, \rho(h) is positive;
    • when h is odd, \rho(h) is negative.
    This result means that observations at contiguous time points are negatively correlated, but observations two time points apart are positively correlated.
    • For example, if an observation, \varepsilon_t, is positive, the next observation, \varepsilon_{t+1}, is typically negative, and the next observation, \varepsilon_{t+2}, is typically positive. Thus, in this case, the sample path is very choppy.

Another interpretation of \rho(h) is the optimal weight for scaling \varepsilon_t into \varepsilon_{t+h}, i.e., the weight, a, that minimizes E[(\varepsilon_{t+h} - a\,\varepsilon_{t})^2]\,.