The model
We define a family of models by a discrete-time stochastic process
,
, called the Full Range Auto Regressive (FRAR) model, by the difference equation
(1)
where
,
,
,
,
and
are parameters,
,
,
, … are independent and identically distributed normal random variables with mean zero and variance
. The initial assumptions about the parameters are as follows:
It is assumed that
will influence
for all positive n and the influence of
on
will decrease, at least for large n, and become insignificant as n becomes very large, because more important for the recent observations and less important for an older observations. Hence
must tend to zero as n goes to infinity. This is achieved by assuming that
. The feasibility of
having various magnitudes of influence on
, when n is small, is made possible by allowing k to take any real value. Because of the periodicity of the circular functions sine and cosine, the domain of
and
are restricted to the interval
.
Thus, the initial assumptions are
,
, and
,
. i.e.,
, where
. Further restrictions on the range of the parameters are placed by examining the identifiability of the model.
Identifiability condition
Identifiability ensures that there is a one to one correspondence between the parameter space and set of associated probability models. Without identifiability it is meaningless to proceed to estimate the parameters of a model using a set of given data. In the present context, identifiability is achieved by restricting the parameters space in such a way that no two points in the parameter space could produce the same time series model.
The coefficients
’s in (1) are functions of
,
,
,
as well as n. That is,
,
,
.
Define
,
, (2)
,
.
Since
,
to each
belonging to A there is
belonging to D such that
. So A is omitted. Similarly, it can be shown that B and C can also be omitted.
Define
,
,
,
.
Since
for
(3)
Using (3) it can be shown as before, that the regions
and
can be omitted. Since no further reduction is possible, it is finally deduced that the region of identifiability of the model is given by
.
Stationarity of the FRAR process
The stationarity of the newly developed FRAR time series model is now examined. The model is given by
. That is,
, where B is the backward shift operator, defined by
. Thus, the model is given by
, or
, where
.
Box and Jenkins2 and Priestley3 have shown that a necessary condition for the stationarity of such processes is that the roots of the equation
must all lie outside the unit circle. So, it is now proposed to investigate the nature of the zeros of
.
The power series
may be rewritten as
, where
,
,
and
. Therefore,
.
The above two series are separately evaluated below.
, where
and IP stands for imaginary part.
Similarly, it can be shown that
, where
.
Therefore,
.
Thus,
implies that
where
,
,
,
. After simplifying, the above equation becomes
. Thus,
(4)
or
(5)
where
,
, and
. This equation (of degree 4) reduces to
where
.
The roots of this equation are, say
and
, are given by
.
Since
, one finally gets the four roots of the equation (4), as
,
,
and
.
The equation (5) implies that, if
is a root of the equation (5) then
is also a root. This implies that
and
are roots of equation (4). Therefore the process is stationary for sufficiently large values of
. But when
is small it seems difficult to examine the stationarity of the process by this approach. Hence, it is proposed to study the asymptotic stationarity of the process in the following section.
Asymptotic stationarity of the FRAR process
In this section we derive the condition for asymptotic stationarity of the FRAR process. For which one has to solve the difference equation (1), so as to obtain an expression for
in terms of
,
,
,
, .... The precise solution of this equation depends on the initial conditions. So to investigate the nature of the first and second moments of
, following Priestley,3 it is assumed that
for
, N being the number of observations in the time series. Then solving (1) by repeated substitutions one obtains
,
where
;
,
,
where
;
.
Similarly proceeding one finally gets
where
with
. Thus, if it is assumed that
for
, which implies has
, then,
.
Further, it can be shown that
and in general
where
. Therefore, allowing
, we get
,
and
provided the series on the right converges. Thus, it is seen that if
exists then it is a function of s only. In order to examine the convergence of
and
, first the behaviour of
, as j tends infinity, is investigated. Since
,
. Similarly,
;
. Thus, in general
, for
.
Since
, the above relation implies that
as
, for any fixed n. Thus
will converge if
.
If we assume that
, then one can show that
and
.
Therefore, the auto-correlation function of the process exists and, as shown earlier, it is a function of s only. Finally allowing
, it is seen that
-
and
exist finitely;
-
exists finitely and is a function of ‘s’ only.
Thus, the condition for
to be asymptotically stationary is that
. Therefore, we summarized the above results by the following theorem 1.
Theorem 1: The Full Range Auto Regressive (FRAR) process
is asymptotically stationary and identifiable if and only if the domain of the parameter space
is
.
Thus, the new FRAR model incorporates long range dependence, involves only four parameters and is totally free from order determination problems.
The posterior analysis
The Bayesian approach to the analyses of the new model consists in determining the posterior distribution of the parameters of the FRAR model and the predictive distribution of future observations. From the former, one makes posterior inferences about the parameters of the FRAR model including the variance of the white noise. From the latter, one may forecast future observations. All these techniques are illustrated by Broemeling4 for autoregressive models.
We shall consider the FRAR model and assume that it is asymptotically stationary and identifiable.
The problem is to estimate the unknown parameters
,
,
,
and
, using the Bayesian methodology on the basis of a past random realization of
say
.
The joint probability density of
is given by
(6)
where
,
and
.
The notation
is used as a general notation for the probability density function of the random variables given within the parentheses following
and
are the past realizations on
which are unknown. Following Priestley2 and Broemeling,3 these are assumed to be zero for the purpose of deriving the posterior distribution of
. Therefore, the range for the index r, viz., 1 through ∞, reduces to 1 through N and so, in the joint probability density function of the observations given by (6), the range of the summation 1 through ∞ can be replaced by 1 through N. By expanding the square in the exponent and simplifying, one gets
(7)
where
,
,
,
.
To find the posterior distribution of
we first have to specify the prior distribution for the parameters.
is distributed as the displaced exponential distribution(since it is bigger than 1) with parameter
;
has the inverted gamma distribution with parameter v and δ;
,
and
are uniformly distributed over their domain.
Thus, the joint prior density function of
is given by
(8)
Using (7), (8), and Bayes’ theorem, the joint posterior density of
,
,
,
and
is obtained as
(9)
(10)
Integrating
out of this joint posterior distribution, we obtain the joint posterior distribution of
,
,
and
,
(11)
where
;
;
;
;
;
.
Thus, the posterior distribution of k conditional on
,
and
is a t-distribution located at
with
degrees of freedom.
Thus, the joint posterior density function of
,
and
can be obtained by integrating with respect to k. Thus,
; with
,
and
. (12)
The above joint posterior density of
,
and
is a very complicated expression and is analytically intractable. One way of solving the problem is to find the marginal posterior density of
,
and
from the joint density (12) using ordinary numerical integration, using FORTRAN.
One-step-ahead prediction
In order to forecast
using the random realization
on
, one must find the conditional distribution of
given the past observations. This is the predictive distribution of
and will be derived by multiplying the conditional density of
given
,
and the posterior density of
given
and then integrating with respect to
. That is,
.
Thus, we obtain
,
. (13)
The square in the exponent in the above expression, say
, can be rewritten, after expanding the square, as
, where
and
. Now multiplying (13) by the joint posterior density of
and integrating over the parameter space
, we obtain,
(14)
First, integrating out
in (14), one gets the joint distribution of
,
,
,
and
as
(15)
where
,
,
,
;
.
Thus,
(16)
where
,
,
.
Further, integrating out <
from (16) we get
(17)
with
which is the conditional predictive distribution of
given
,
and
. Further elimination of the parameters
,
and
from (17) is not possible analytically. So the marginal posterior density of
cannot be expressed in a closed form. Since the distribution in (17) is analytically not tractable, a complete Bayesian analysis is possible by numerical integration technique or simulation based approach, viz., MCMC technique.
Suppose one wants a point estimate (posterior mean) of
, then one should compute the marginal posterior density of
from (17) and use it to calculate the marginal posterior mean of
. Thus four dimensional numerical integration is necessary in order to estimate
. But it is a very difficult problem.
Practically, to perform four dimensional numerical integration is very difficult and therefore to reduce the dimensions of the numerical integration one may substitute the estimators, posterior means,
,
and
respectively in the place of
,
and
and then perform one dimensional numerical integration to find the conditional mean of
. That is, one may eliminate the parameters as much as possible by analytical methods and then use the conditional estimates for the remaining parameters to compute the marginal posterior mean of the future observation.