16 Handbook of Discrete-Valued Time Series
1.5 Prediction
Following Gneiting et al. (2007), we take the point of view that predictions should be proba-
bilistic in nature. In addition, they should strive to maximize the sharpness of the predictive
distribution subject to calibration. Calibration refers to the statistical consistency between
the predictive distribution and the observations. The notion of sharpness refers to the con-
centration of the predictive distribution and is a property of the forecasts only. It follows
that if the predictive distribution is more concentrated, then the forecasts are sharper. In
this section, we provide diagnostic tools to evaluate the predictive performance. Note that
calculation of all these measures requires an assumption on the conditional distribution of
the process; hence general processes of the form (1.2) cannot be tted without simulating
from the mixing variables Z
t
. Predictive performance based on the following diagnostic
tools has been examined recently by Jung and Tremayne (2011) and Christou and Fokianos
(2015).
1.5.1 Assessment of Probabilistic Calibration
To ascertain whether or not the negative binomial distribution is a better choice than the
Poisson distribution, we use the diagnostic tool of the Probability Integral Transformation
(PIT) histogram, as explained below. This tool is used for checking the statistical consis-
tency between the predictive distribution and the distribution of the observations. If the
observation is drawn from the predictive distribution, then the PIT has a standard uniform
distribution. In the case of count data, the predictive distribution is discrete and therefore
the PIT is no longer uniform. To remedy this, several authors have suggested a randomized
PIT. However, Czado et al. (2009) recently proposed a nonrandomized uniform version of
the PIT. We explain their approach in the context of count time series models. Note that the
approach is quite general and can accommodate various data generating processes.
In our context, we t any model discussed earlier to the data by using the quasi-
likelihood function (1.14). After obtaining consistent estimators for the regression param-
eters, we estimate the mean process λ
t
by
λ
ˆ
t
= λ
t
(
θ
ˆ
) and the parameter ν by νˆ. Then, the
PIT is based on the conditional cumulative distribution
⎧
⎪
0 u ≤ P
y−1
,
⎨
F(u|Y
t
= y) =
(u − P
y−1
)/(P
y
− P
y−1
) P
y−1
≤ u ≤ P
y
,
⎪
⎩
1 u ≥ P
y
,
where P
y
is equal to the conditional c.d.f. either of the Poisson distribution (1.3) evaluated
at
λ
ˆ
t
, or of the negative binomial p.m.f. (1.4) evaluated at
λ
ˆ
t
and νˆ. Subsequently, we form
the mean PIT by
1
n
F
¯
(u) =
F
(t)
(u|y
t
),0 ≤ u ≤ 1.
n
t=1
The mean PIT is compared to the c.d.f. of the standard uniform distribution. The compar-
ison is performed by plotting a nonrandomized PIT histogram, which can be used as a
diagnostic tool. After selecting the number of bins, say J, we compute