# Trends and cycles in unemployment¶

Here we consider three methods for separating a trend and cycle in economic data. Supposing we have a time series $y_t$, the basic idea is to decompose it into these two components:

$$y_t = \mu_t + \eta_t$$

where $\mu_t$ represents the trend or level and $\eta_t$ represents the cyclical component. In this case, we consider a stochastic trend, so that $\mu_t$ is a random variable and not a deterministic function of time. Two of methods fall under the heading of "unobserved components" models, and the third is the popular Hodrick-Prescott (HP) filter. Consistent with e.g. Harvey and Jaeger (1993), we find that these models all produce similar decompositions.

This notebook demonstrates applying these models to separate trend from cycle in the U.S. unemployment rate.

In :
%matplotlib inline

In :
import numpy as np
import pandas as pd
import statsmodels.api as sm
import matplotlib.pyplot as plt

In :
from pandas_datareader.data import DataReader
endog = DataReader('UNRATE', 'fred', start='1954-01-01')


### Hodrick-Prescott (HP) filter¶

The first method is the Hodrick-Prescott filter, which can be applied to a data series in a very straightforward method. Here we specify the parameter $\lambda=129600$ because the unemployment rate is observed monthly.

In :
hp_cycle, hp_trend = sm.tsa.filters.hpfilter(endog, lamb=129600)


### Unobserved components and ARIMA model (UC-ARIMA)¶

The next method is an unobserved components model, where the trend is modeled as a random walk and the cycle is modeled with an ARIMA model - in particular, here we use an AR(4) model. The process for the time series can be written as:

\begin{align} y_t & = \mu_t + \eta_t \\ \mu_{t+1} & = \mu_t + \epsilon_{t+1} \\ \phi(L) \eta_t & = \nu_t \end{align}

where $\phi(L)$ is the AR(4) lag polynomial and $\epsilon_t$ and $\nu_t$ are white noise.

In :
mod_ucarima = sm.tsa.UnobservedComponents(endog, 'rwalk', autoregressive=4)
# Here the powell method is used, since it achieves a
# higher loglikelihood than the default L-BFGS method
res_ucarima = mod_ucarima.fit(method='powell', disp=False)
print(res_ucarima.summary())

/home/travis/build/statsmodels/statsmodels/statsmodels/tsa/base/tsa_model.py:165: ValueWarning: No frequency information was provided, so inferred frequency MS will be used.
% freq, ValueWarning)

                        Unobserved Components Results
==============================================================================
Dep. Variable:                 UNRATE   No. Observations:                  782
Model:                    random walk   Log Likelihood                 249.658
+ AR(4)   AIC                           -487.316
Date:                Mon, 25 Mar 2019   BIC                           -459.352
Time:                        20:22:50   HQIC                          -476.561
Sample:                    01-01-1954
- 02-01-2019
Covariance Type:                  opg
================================================================================
coef    std err          z      P>|z|      [0.025      0.975]
--------------------------------------------------------------------------------
sigma2.level     0.0174      0.003      6.169      0.000       0.012       0.023
sigma2.ar        0.0109      0.003      3.481      0.000       0.005       0.017
ar.L1            1.0444      0.066     15.832      0.000       0.915       1.174
ar.L2            0.4610      0.105      4.372      0.000       0.254       0.668
ar.L3           -0.3313      0.127     -2.617      0.009      -0.579      -0.083
ar.L4           -0.1850      0.077     -2.405      0.016      -0.336      -0.034
===================================================================================
Ljung-Box (Q):                       76.46   Jarque-Bera (JB):                42.01
Prob(Q):                              0.00   Prob(JB):                         0.00
Heteroskedasticity (H):               0.49   Skew:                             0.24
Prob(H) (two-sided):                  0.00   Kurtosis:                         4.03
===================================================================================

Warnings:
 Covariance matrix calculated using the outer product of gradients (complex-step).


### Unobserved components with stochastic cycle (UC)¶

The final method is also an unobserved components model, but where the cycle is modeled explicitly.

\begin{align} y_t & = \mu_t + \eta_t \\ \mu_{t+1} & = \mu_t + \epsilon_{t+1} \\ \eta_{t+1} & = \eta_t \cos \lambda_\eta + \eta_t^* \sin \lambda_\eta + \tilde \omega_t \qquad & \tilde \omega_t \sim N(0, \sigma_{\tilde \omega}^2) \\ \eta_{t+1}^* & = -\eta_t \sin \lambda_\eta + \eta_t^* \cos \lambda_\eta + \tilde \omega_t^* & \tilde \omega_t^* \sim N(0, \sigma_{\tilde \omega}^2) \end{align}
In :
mod_uc = sm.tsa.UnobservedComponents(
endog, 'rwalk',
cycle=True, stochastic_cycle=True, damped_cycle=True,
)
# Here the powell method gets close to the optimum
res_uc = mod_uc.fit(method='powell', disp=False)
# but to get to the highest loglikelihood we do a
# second round using the L-BFGS method.
res_uc = mod_uc.fit(res_uc.params, disp=False)
print(res_uc.summary())

/home/travis/build/statsmodels/statsmodels/statsmodels/tsa/base/tsa_model.py:165: ValueWarning: No frequency information was provided, so inferred frequency MS will be used.
% freq, ValueWarning)

                            Unobserved Components Results
=====================================================================================
Dep. Variable:                        UNRATE   No. Observations:                  782
Model:                           random walk   Log Likelihood                 214.530
+ damped stochastic cycle   AIC                           -421.060
Date:                       Mon, 25 Mar 2019   BIC                           -402.428
Time:                               20:22:52   HQIC                          -413.893
Sample:                           01-01-1954
- 02-01-2019
Covariance Type:                         opg
===================================================================================
coef    std err          z      P>|z|      [0.025      0.975]
-----------------------------------------------------------------------------------
sigma2.level        0.0120      0.005      2.435      0.015       0.002       0.022
sigma2.cycle        0.0195      0.005      4.031      0.000       0.010       0.029
frequency.cycle     0.0653      0.005     12.718      0.000       0.055       0.075
damping.cycle       0.9904      0.004    258.038      0.000       0.983       0.998
===================================================================================
Ljung-Box (Q):                      162.66   Jarque-Bera (JB):                87.15
Prob(Q):                              0.00   Prob(JB):                         0.00
Heteroskedasticity (H):               0.48   Skew:                             0.48
Prob(H) (two-sided):                  0.00   Kurtosis:                         4.33
===================================================================================

Warnings:
 Covariance matrix calculated using the outer product of gradients (complex-step).


### Graphical comparison¶

The output of each of these models is an estimate of the trend component $\mu_t$ and an estimate of the cyclical component $\eta_t$. Qualitatively the estimates of trend and cycle are very similar, although the trend component from the HP filter is somewhat more variable than those from the unobserved components models. This means that relatively mode of the movement in the unemployment rate is attributed to changes in the underlying trend rather than to temporary cyclical movements.

In :
fig, axes = plt.subplots(2, figsize=(13,5));
axes.set(title='Level/trend component')
axes.plot(endog.index, res_uc.level.smoothed, label='UC')
axes.plot(endog.index, res_ucarima.level.smoothed, label='UC-ARIMA(2,0)')
axes.plot(hp_trend, label='HP Filter')
axes.legend(loc='upper left')
axes.grid()

axes.set(title='Cycle component')
axes.plot(endog.index, res_uc.cycle.smoothed, label='UC')
axes.plot(endog.index, res_ucarima.autoregressive.smoothed, label='UC-ARIMA(2,0)')
axes.plot(hp_cycle, label='HP Filter')
axes.legend(loc='upper left')
axes.grid()

fig.tight_layout(); 