\( \newcommand{\mathbbm}[1]{\boldsymbol{\mathbf{#1}}} \)

9.3 Distributional assumptions of ADAM ARIMA

Following the same idea as in pure additive (Section 5.5) and pure multiplicative (Section 6.5) ETS models, we can have state space ARIMA with different distributional assumptions, but with the distributions aligning more appropriately with the types of models. For additive ARIMA:

  1. Normal: \(\epsilon_t \sim \mathcal{N}(0, \sigma^2)\);
  2. Laplace: \(\epsilon_t \sim \mathcal{L}(0, s)\);
  3. Generalised Normal: \(\epsilon_t \sim \mathcal{GN}(0, s, \beta)\);
  4. S: \(\epsilon_t \sim \mathcal{S}(0, s)\).

For multiplicative ARIMA:

  1. Inverse Gaussian: \(\left(1+\epsilon_t \right) \sim \mathcal{IG}(1, \sigma^2)\);
  2. Log-Normal: \(\left(1+\epsilon_t \right) \sim \text{log}\mathcal{N}\left(-\frac{\sigma^2}{2}, \sigma^2\right)\);
  3. Gamma: \(\left(1+\epsilon_t \right) \sim \mathcal{\Gamma}(\sigma^{-2}, \sigma^2)\).

The restrictions imposed on the model parameters are inherited from the ADAM ETS.

9.3.1 Conditional distributions

When it comes to conditional distribution of variables, additive ADAM ARIMA with the assumptions above has closed forms for all of them. For example, if we work with additive ARIMA, then according to recursive relation (9.22) from Section 9.2, the h steps ahead value follows the same distribution but with different conditional mean and variance. For example, if \(\epsilon_t \sim \mathcal{GN}(0, s, \beta)\), then \(y_{t+h} \sim \mathcal{GN}(\mu_{y,t+h}, s_{h}, \beta)\), where \(s_{h}\) is the conditional h steps ahead scale, found from the connection between variance and scale in Generalised Normal distribution via: \[\begin{equation*} s_h = \sqrt{\frac{\sigma^2_h \Gamma(1/\beta)}{\Gamma(3/\beta)}}. \end{equation*}\]

Using similar principles, we can calculate scale parameters for the other distributions.

When it comes to the multiplicative models, the conditional distribution has the closed form in case of Log-Normal (it is Log-Normal as well), but does not have it in case of Inverse Gaussian and Gamma. In the former case, the logarithmic moments can be directly used to define the parameters of distribution, i.e. if \(\left(1+\epsilon_t \right) \sim \text{log}\mathcal{N}\left(-\frac{\sigma^2}{2}, \sigma^2\right)\), then \(y_{t+h} \sim \text{log}\mathcal{N}\left(\mu_{\log y,t+h}, \sigma^2_{\log y,h} \right)\). In the other cases, simulations need to be used in order to get the quantile, cumulative and density functions.