ARIMA methodology proposed by Box Jenkin (1976) is one of the popular methods to uncover the hidden characteristics in the time series data, and to generate forecasts. The model is built based on the plot of the autocorrelation and partial autocorrelation functions of the dependent time series. The plot provides the information to determine which autoregressive or moving average component should be used in the model. Basically, ARIMA(p,q) model includes two independent processes. They are autoregressive process (AR(p)) and Moving average Process (MA(q)).
Autoregressive Process
Autoregressive Process can be represented as AR(p) that can be interpreted as a linear combination of prior observations, AR(p) can be summarized as:xt = ζ + φ1xt−1 + φ2xt−2 + … φkxt−p + ε(1) where p is order of autoregressive model, ζ is the constant of the model, and φ1, φ2, …, φp are the autoregressive model parameters.
3Moving Average Process
It can be represented as MA(q). In the Moving average model, the observation can be affected by its previous error, MA(q) can be written as:xt = μ + εt − θ1εt−1 − θ2εt−2 − … − θ3εt−q(2) where q is the order of the moving average model, μ is a constant and θ1,θ2,…,θq are the moving average model parameters
We suppose the stochastic process {xt : t = 0, ±1, ±2,…}, and its mean function is defined as μt = E(xt) for t = 0, ±1, ±2,…. In general, μt differs at different t. Autocorrelation is defined as:
(3) where
Cov(
xt,
xs ) is autocovariance function,
Cov(
xt,
xs ) =
E[(
xt − μt)(
xs − μs)] =
E(
xtxs)
− μtμsThe partial autocorrelation at lag k (ϕkk) is defined as:ϕkk = Corr(xt, xt− k | xt−1, xt−2, …, xt− k+1) that can be interpreted as the correlation between xt and xt− k after removing the effect of the intervening variables xt− 1, xt− 2, …, xt− k=1