455 Long Memory Discrete-Valued Time Series
The setup here is similar to Brockwell (2007) and MacDonald and Zucchini (2015; Chap-
ter 12 in this volume) and proceeds via a conditional specication. For simplicity, we
restrict attention to a conditional Poisson model where the logarithm of a latent intensity
parameter is modeled as a Gaussian autoregressive fractionally integrated moving-average
(ARFIMA) (p, d, q) process. It is assumed that the data are conditionally independent given
the underlying latent Poisson intensity parameter. Specically, we posit that the conditional
distribution of X
t
given λ
t
is
ind
X
t
|λ
t
, ∼ Poisson(λ
t
), t = 1, 2, .... (21.15)
Let λ
∗
= log(λ
t
). We model {λ
∗
}
∞
1
with a zero-mean Gaussian ARFIMA(p, d, q) process
t t
t=
satisfying
φ(B)(1 − B)
d
λ
∗
t
= θ(B)
t
, t = 1, 2, ...,
where (1 −B)
d
= 1 − Bd −B
2
d(d −1)/2!−··· is the general binomial expansion, p, q ∈ Z
+
,
d ∈ (−1/2, 1/2), {
t
} is zero-mean white noise, and the AR and MA polynomials are as in
(21.3) and (21.4). With λ
∗
n
= (λ
∗
1
, ..., λ
∗
n
)
and = (, , d, σ
2
), where = (φ
1
, ..., φ
p
) and
= (θ
1
, ..., θ
q
), the Gaussian ARFIMA supposition implies that
λ
∗
| ∼ N(0,
n
), (21.16)
n
where
n
is the autocovariance matrix of λ
∗
n
. As in Brockwell (2007), it is straightforward
to specify a nonzero mean in (21.16); that is, deterministic regressors could be added to
(21.16) in a straightforward manner.
With xed values of the ARFIMA parameters in , {λ
∗
}
∞
1
is a strictly stationary
t
t=
Gaussian series. It follows that {λ
t
}
∞
1
and {X
t
}
∞
1
are also strictly stationary. However,
t= t=
the marginal distribution of X
t
is unclear. Some computations provide the form
�
∞
−λ
λ
k
exp{−
1
ln(λ)
2
}
P(X
t
= k) =
e
√
2 γ
∗
(0)
dλ, k = 0, 1, ...,
k!
λ 2πγ
∗
(0)
0
where γ
∗
(0) = Var(λ
∗
t
). This is a difcult integral to explicitly evaluate, although numeri-
cal approximations can be made—see Asmussen et al. (2014) for the latest. The covariance
Cov(X
t
, X
t+h
) also appears intractable. It seems logical that {X
t
}
∞
1
will also have long
t=
memory, but this has not been formally veried.
In a Bayesian setting, the time series parameters are typically treated as random. For
example, the distributions of and could be taken as uniform over their respective AR
and MA stationarity and invertibility regions, d could be uniform over (−1/2, 1/2),and σ
2
would have a distribution supported on (0, ∞). One could take these components to be
independent, although formulations allowing dependence between these components are
also possible.
In practice, it is convenient to work with an autoregressive setup (i.e., q = 0) for λ
∗
.
Even with this simplifying assumption, several open research questions arise. For esti-
mation, it would be useful to derive efcient MCMC sampling algorithms. One such
algorithm is provided by Brockwell (2007). Also, for large n, it might be advantageous
to consider approximate Bayesian inference by a Whittle likelihood in lieu of an exact
Gaussian likelihood (see Palma 2007, McElroy and Holan 2012). Further computational