\( \newcommand{\mathbbm}[1]{\boldsymbol{\mathbf{#1}}} \)

9.5 Examples of application

Building upon the example with AirPassengers data from Section 8.5.2, we will construct several multiplicative ARIMA models and see which one is the most appropriate for the data. As a reminder, the best additive ARIMA model was SARIMA(0,2,2)(1,1,1)\(_{12}\), which had AICc of 1035.196. We will do something similar here, using Log-Normal distribution, thus working with Log-ARIMA. To understand what model can be used in this case, we can take the logarithm of data and see what happens with the components of time series:

log(AirPassengers) |> plot()

We still have the trend in the data, and the seasonality now corresponds to the additive rather than the multiplicative (as expected). While we might still need the second differences for the non-seasonal part of the model, taking the first differences for the seasonal should suffice because the logarithmic transform will take care of the expanding seasonal pattern in the data. So we can test several models with different options for ARIMA orders:

adamLogSARIMAAir <- vector("list",3)
# logSARIMA(0,1,1)(0,1,1)[12]
adamLogSARIMAAir[[1]] <-
  adam(AirPassengers, "NNN", lags=c(1,12),
       orders=list(ar=c(0,0), i=c(1,1), ma=c(1,1)),
       h=12, holdout=TRUE, distribution="dlnorm")
# logSARIMA(0,2,2)(0,1,1)[12]
adamLogSARIMAAir[[2]] <-
  adam(AirPassengers, "NNN", lags=c(1,12),
       orders=list(ar=c(0,0), i=c(2,1), ma=c(2,2)),
       h=12, holdout=TRUE, distribution="dlnorm")
# logSARIMA(1,1,2)(0,1,1)[12]
adamLogSARIMAAir[[3]] <-
  adam(AirPassengers, "NNN", lags=c(1,12),
       orders=list(ar=c(1,0), i=c(1,1), ma=c(2,1)),
       h=12, holdout=TRUE, distribution="dlnorm")
names(adamLogSARIMAAir) <- c("logSARIMA(0,1,1)(0,1,1)[12]",
                             "logSARIMA(0,2,2)(0,1,1)[12]",
                             "logSARIMA(1,1,2)(0,1,1)[12]")

The thing that is different between the models is the non-seasonal part. Using the connection with ETS (discussed in Section 8.4), the first model should work on local level data, the second should be optimal for the local trend series, and the third one is placed somewhere in between the two. We can compare the models using AICc:

sapply(adamLogSARIMAAir, AICc)
## logSARIMA(0,1,1)(0,1,1)[12] logSARIMA(0,2,2)(0,1,1)[12] 
##                    988.6863                   1310.5471 
## logSARIMA(1,1,2)(0,1,1)[12] 
##                    983.6632

It looks like the logSARIMA(1,1,2)(0,1,1)\(_{12}\) is more appropriate for the data. In order to make sure that we did not miss anything, we analyse the residuals of this model (Figure 9.1):

par(mfcol=c(2,1), mar=c(2,2,2,1))
plot(adamLogSARIMAAir[[3]], which=10:11)
ACF and PACF of logSARIMA(1,1,2)(0,1,1)$_{12}$.

Figure 9.1: ACF and PACF of logSARIMA(1,1,2)(0,1,1)\(_{12}\).

We can see that there are no significant coefficients on either the ACF or PACF, so there is nothing else to improve in this model (we discuss this in more detail in Section 14.5). We can then produce a forecast from the model and see how it performed on the holdout sample (Figure 9.2):

forecast(adamLogSARIMAAir[[3]], h=12, interval="prediction") |>
    plot(main=paste0(adamLogSARIMAAir[[3]]$model,
                     " with Log-Normal distribution"))
Forecast from logSARIMA(1,1,2)(0,1,1)$_{12}$.

Figure 9.2: Forecast from logSARIMA(1,1,2)(0,1,1)\(_{12}\).

The ETS model closest to the logSARIMA(1,1,2)(0,1,1)\(_{12}\) would probably be ETS(M,M,M), because the former has both seasonal and non-seasonal differences (see discussion in Subsection 8.4.4):

adamETSAir <- adam(AirPassengers, "MMM", h=12, holdout=TRUE)
adamETSAir
## Time elapsed: 0.11 seconds
## Model estimated using adam() function: ETS(MMM)
## With optimal initialisation
## Distribution assumed in the model: Gamma
## Loss function type: likelihood; Loss function value: 474.4954
## Persistence vector g:
##  alpha   beta  gamma 
## 0.6906 0.0443 0.0490 
## 
## Sample size: 132
## Number of estimated parameters: 17
## Number of degrees of freedom: 115
## Information criteria:
##       AIC      AICc       BIC      BICc 
##  982.9908  988.3592 1031.9984 1045.1049 
## 
## Forecast errors:
## ME: -7.171; MAE: 14.248; RMSE: 21.331
## sCE: -32.785%; Asymmetry: -43.8%; sMAE: 5.428%; sMSE: 0.66%
## MASE: 0.592; RMSSE: 0.681; rMAE: 0.187; rRMSE: 0.207

Comparing information criteria, Log-ARIMA should be preferred to ETS(M,M,M). Similar conclusion is achieved if we compare holdout accuracy (based on RMSSE) of the two models:

adamLogSARIMAAir[[3]]
## Time elapsed: 0.3 seconds
## Model estimated using adam() function: SARIMA(1,1,2)[1](0,1,1)[12]
## With optimal initialisation
## Distribution assumed in the model: Log-Normal
## Loss function type: likelihood; Loss function value: 469.4387
## ARMA parameters of the model:
##         Lag 1
## AR(1) -0.4546
##         Lag 1  Lag 12
## MA(1)  0.1243 -0.5919
## MA(2) -0.0541      NA
## 
## Sample size: 132
## Number of estimated parameters: 19
## Number of degrees of freedom: 113
## Information criteria:
##       AIC      AICc       BIC      BICc 
##  976.8775  983.6632 1031.6507 1048.2174 
## 
## Forecast errors:
## ME: -12.767; MAE: 13.653; RMSE: 18.998
## sCE: -58.363%; Asymmetry: -92%; sMAE: 5.201%; sMSE: 0.524%
## MASE: 0.567; RMSSE: 0.606; rMAE: 0.18; rRMSE: 0.184

To get a better idea about the holdout performance of the two models, we would need to apply them in the rolling origin fashion (Section 2.4), collect a distribution of errors, and then decide, which one to choose.