440 Handbook of Discrete-Valued Time Series
Own drug
Observed
Prediction
0
20
25
10
15
5
0 20
40 60 80 100
Physician
Observed
Prediction
Challenger drug
0
20
25
10
15
5
Leader drug
0 20
40 60 80 100
Physician
30
20
10
0
Physician
Observed
Prediction
0 20 40 60 80 100
FIGURE 20.3
One-month-ahead predictions for all physicians.
possible to construct other models for multivariate time series of counts. For instance, Ma
et al. (2008) describe a multivariate Poisson-lognormal framework for modeling multivari-
ate count time series with an application to transportation safety, while Li et al. (1999)
describe a framework for tting a multivariate ZIP model.
One consideration in tting multivariate Poisson models is the computation time for
calculating the likelihood function, which can increase rapidly with increasing compo-
nent dimension m. In such cases, it is useful to explore approximations for carrying out
Bayesian inference, such as the integrated nested Laplace approximation (INLA), or the
variational Bayes approach. The INLA approach has been discussed in Rue and Martino
(2007) and Rue et al. (2009); also see Gamerman et al. (2015; Chapter 8 in this volume)
Serhiyenho et al. (2015) described an application to trafc safety using R-INLA. The vari-
ational Bayes approach has been discussed in several papers including McGrory and
Titterington (2007), Ormerod and Wand (2010), and Wand et al. (2011), among others. A
study of these approaches for modeling multivariate time series of counts and a comparison
with the MCMC approach will be useful.
20.A Appendix
Details on the complete conditional distributions for the HMDM and MDFM models are
shown below.
441 Dynamic Models for Time Series of Counts with a Marketing Application
20.A.1 Complete Conditional Distributions for the HMDM Model
The complete conditional density of β
it
for i = 1, ..., N and t = 1, ..., T is
1
f (β
it
|γ
t
, η, V
i
, W, Y) MP
m
(y
it
|λ
it
) exp
(β
it
γ
t
)
V
i
1
(β
it
γ
t
)
.
2
The complete conditional density of γ
t
for t = 1, ..., T is
N
1
f (γ
t
|β
it
, η, V
i
, W, Y) exp
2
(β
it
γ
t
)
V
i
1
(β
it
γ
t
)
i=1
× exp
1
(γ
t
Gγ
t1
)
W
1
(γ
t
Gγ
t1
)
2
× exp
1
(γ
t+1
Gγ
t
)
W
1
(γ
t+1
Gγ
t
)
.
2
The complete conditional density of η is
T
N
1
1
f (η|β
it
, γ
t
, V
i
, W, Y) MP
m
(y
it
|λ
it
) exp
(η μ
η
)
η
(η μ
η
)
.
2
t=1
i=1
The complete conditional density of V
i
for i = 1, ..., N is
T
1
V
1
f (V
i
|β
it
, γ
t
, η, W, Y) |V
i
|
1/2
exp
2
(β
it
γ
t
)
i
(β
it
γ
t
)
t=1
×|V
i
|
n
v
/2
exp
1
2
tr(V
i
1
S
v
)
.
The complete conditional density of W is
T
1
f (W|β
it
, γ
t
, η, V
i
, Y) |W|
1/2
exp
(γ
t
Gγ
t1
)
W
1
(γ
t
Gγ
t1
)
2
t=1
×|W|
1/2
exp
1
(γ
0
m
0
)
C
0
1
(γ
0
m
0
)
2
×|W|
n
w
/2
exp
1
tr(W
1
S
w
)
.
2
442 Handbook of Discrete-Valued Time Series
20.A.2 Complete Conditional Distributions and Sampling Algorithms for the MDFM Model
For t = 2, ..., T and h = 1, ..., H,
N
z
i,t,h
1
W
1
f (β
0,t,h
) π
h
[p(y
i,t
|λ
i,t,h
)]
z
i,t,h
exp
2
(β
0,t,h
Gβ
0,t1,h
)
h
(β
0,t,h
Gβ
0,t1,h
)
i=1
1
W
1
.
× exp
2
(β
0,t+1,h
Gβ
0,t,h
)
h
(β
0,t+1,h
Gβ
0,t,h
)
For i = 1, ..., N and h = 1, ...H,
T
f
β
1,i,h
π
h
z
i,t,h
p
y
i,t
|λ
i,t,h

z
i,t,h
exp
1
β
1,i,h
a
1,i,h
R
1,i
1
,h
β
1,i,h
a
1,i,h
,and
2
t=1

z
i,t,h
f
β
2,i,h
T
π
z
h
i,t,h
p
y
i,t
|λ
i,t,h
exp
1
2
β
2,i,h
a
2,i,h
R
2,i
1
,h
β
2,i,h
a
2,i,h
.
t=1
For h = 1, ..., H,
T
exp
1
W
1
t=2
f (W
h
)
2
(β
0,t,h
Gβ
0,t1,h
)
h
(β
0,t,h
Gβ
0,t1,h
)
× exp
1
(α μ
α
)
σ
1
|W
h
|
n
h
/2
exp
1
tr(W
1
.
α
(α μ
α
)
h
S
h
)
2 2
The complete conditional density of π is
H
N
T
f (π)
π
d
h
1+
i=1
t=1
z
i,t,h
,
h
h=1
d
1
1+
N
T
which reduces to f (π
1
) π
1
i=1
t=1
z
i,t,1
(1 π
1
)
d
2
1
when H = 2.
The sampling algorithms are shown in the following. At the lth Gibbs iteration, (a) we
generate z
(l)
f (z|y,
(l)
) and (b) we then generate
(l+1)
π(|y, z)). Specically, in step
(a), for i = 1, ..., N and t = 1, ..., T, we simulate z
i,t
from a multinomial distribution with
probabilities
H
p
j
= π
(
j
l)
[p(y
i,t
|λ
i,t,h
)]
z
i,t,h
/ π
(
h
l)
[p(y
i,t
|λ
i,t,h
)]
z
i,t,h
h=1
for j = 1, ..., H. When H = 2, we simulate z
i,t,1
from Bernoulli distribution with weight
π
1
(l)
[p(y
i,t
|λ
i,t,h
)]
z
i,t,h
p
z
=
(l) (l)
.
π
1
[p(y
i,t
|λ
i,t,h
)]
z
i,t,h
+ π
2
[p(y
i,t
|λ
i,t,h
)]
z
i,t,h

443 Dynamic Models for Time Series of Counts with a Marketing Application
In step (b), the complete conditional densities of W
h
for h = 1, ..., H and the mixture
proportion π have closed forms; we directly sample W
h
from IW(n
h
, S
h
), where
n
h
= n
h
+ T
T
S
h
= S
h
+
β
0,t,h
Gβ
0,t1,h
β
0,t,h
Gβ
0,t1,h
.
t=2
N
T
N
T
We sample π from a Dirichlet
d
1
+
i=1 t=1
z
i,t,1
, ..., d
H
+
i=1 t=1
z
i,t,h
distribution
N
T
when H > 2andfromaBeta d
1
+
i=1 t=1
z
i,t,1
, d
2
distribution when H = 2. The draws
are made using a random walk Metropolis–Hastings algorithm.
References
Ahearne, M., Jelinek, R., and Jones, E. (2007). Examining the effect of salesperson service behavior in
a competitive context. Journal of Academy of Marketing Science, 35(4):603–616.
Aktekin, T., Soyer, R., and Xu, F. (2014). Assessment of mortgage default risk via Bayesian state space
models. Annals of Applied Statistics, 7(3):1450–1473.
Albert, J. H. and Chib, S. (1993). Bayesian analysis of binary and polychotomous response data.
JournalofAmericanStatisticalAssociation, 88:669–679.
Blattberg, R. C., Kim, B.-D., and Neslin, S. A. (2008). Database Marketing: Analyzing and Managing
Customers. Springer Science+ Business Media LLC: New York.
Carlin, B. P., Polson, N., and Stoffer, D. S. (1992). Bayes estimates for the linear model (with
discussion). Journal of the Royal Statistical Society, Series B, 34:1–41.
Carter, C. and Kohn, R. (1994). On Gibbs sampling for state space models. Biometrika, 81:541–553.
Chen, M.-H., Shao, A., and Ibrahim, J. (2000). Monte Carlo Methods. Springer-Verlag: New York.
Davis, R. A., Dunsmuir, W. T. M., and Streett, S. B. (2003). Observation-driven models for Poisson
counts. Biometrika, 90(4):777–790.
Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977). Maximum likelihood from incomplete data
via the EM algorithm. Journal of the Royal Statistical Society, Series B, 39:1–38.
Diebolt, J. and Robert, C. P. (1994). Estimation of nite mixture distributions through Bayesian
sampling. Journal of the Royal Statistical Society, Series B, 56:363–375.
Fahrmeir, L. (1992). Posterior mode estimation by extended Kalman ltering for multivariate
dynamic generalized linear models. Journal of the American Statistical Association, 87:501–509.
Fahrmeir, L. and Kaufmann, H. (1991). On Kalman ltering, posterior mode estimation and Fisher
scoring in dynamic exponential family regression. Metrika, 38:37–60.
Fruhwirth-Schnatter, S. (1994). Data augmentation and dynamic linear models. Journal of Time Series
Analysis, 24:295–320.
Gamerman, D. (1998). Markov chain Monte Carlo for dynamic generalized linear models. Biometrika,
85:215–227.
Gamerman, D., Abanto-Valle, C. A., Silva, R. S., and Martins, T. G. (2015). Dynamic Bayesian models
for discrete-valued time series. In R. A. Davis, S. H. Holan, R. Lund and N. Ravishanker, eds.,
Handbook of Discrete-Valued Time Series, pp. 165–186. Chapman & Hall, Boca Raton, FL.
Gamerman, D. and Migon, H. (1993). Dynamic hierarchical models. Journalofthe RoyalStatistical
Society Series B, 3:629–642.
Harvey, A. C. and Fernandes, C. (1989). Time series models for count or qualitative observations.
Journal of Business and Economic Statistics, 7(4):407–417.
444 Handbook of Discrete-Valued Time Series
Hu, S. (2012). Dynamic modeling of discrete-valued time series with applications. PhD thesis,
University of Connecticut: Storrs, CT.
Johnson, N., Kotz, S., and Balakrishnan, N. (1997). Discrete Multivariate Distributions. Wiley:
New York.
Kalman, R. E. (1960). A new approach to linear ltering and prediction problems. Transactions ASME
Journal of Basic Engineering, 82(Series D):35–45.
Kalman, R. E. and Bucy, R. S. (1961). New results in linear ltering and prediction theory. Tran s a c tions
ASME-Journal of Basic Engineering, 83(Series D):95–108.
Karlis, D. and Meligkotsidou, L. (2005). Multivariate Poisson regression with covariance structure.
Statistics and Computing, 15:255–265.
Karlis, D. and Meligkotsidou, L. (2007). Finite mixtures of multivariate Poisson regression with
application. Journal of Statistical Planning and Inference, 137:1942–1960.
Kedem, B. and Fokianos, K. (2002). Regression Models for Time Series Analysis. Wiley: Hoboken, NJ.
Lambert, D. (1992). Zero-inated Poisson regression, with an application to defects in manufacturing.
Technometrics, 34(1):1–13.
Lancaster, T. (2000). The incidental parameter problem since 1948. Journal of Econometrics, 95:
391–413.
Landim, F. and Gamerman, D. (2000). Dynamic hierarchical models: An extension to matrix-variate
observations. Computational Statistics and Data Analysis, 35:11–42.
Li, C., Lu, J., Park, J., Kim, K., Brinkley, P., and Peterson, J. (1999). Multivariate zero-inated Poisson
models and their applications. Technometrics, 41:29–38.
Lindley, D. and Smith, A. (1972). Bayes estimates for the linear model (with discussion). Journal of
the Royal Statistical Society, Series B, 34:1–41.
Ma, J., Kockelman, K. M., and Damien, P. (2008). A multivariate Poisson-lognormal regression
model for prediction of crash counts by severity, using Bayesian methods. Accident Analysis
and Prevention, 40:964–975.
Mahamunulu, D. (1967). A note on regression in the multivariate Poisson distribution. Journal of the
American Statistical Association, 62:251–258.
McCullagh, P. and Nelder, J. (1989). Generalized Linear Models, 2nd edn. Chapman & Hall:
London, U.K.
McGrory, C. A. and Titterington, D. M. (2007). Variational approximations in Bayesian model selec-
tion for nite mixture distributions. Computational Statistics and Data Analysis, 51:5352–5367.
Mizik, N. and Jacobson, R. (2004). Are physicians easy marks? Quantifying the effects of detailing
and sampling on new prescriptions. Management Science, 50(12):1704–1715.
Montoya, R., Netzer, O., and Jedidi, K. (2010). Dynamic allocation of pharmaceutical detailing and
sampling for long-term protability. Marketing Science, 29(5):909–924.
Netzer, O., Lattin, J. M., and Srinivasan, V. (2008). A hidden Markov model of customer relationship
dynamics. Marketing Science, 27(2):185–204.
Ormerod, J. T. and Wand, M. P. (2010). Explaining variational approximations. The American
Statistician, 64(2):140–153.
Ravishanker, N., Serhiyenko, V., and Willig, M. (2014). Hierarchical dynamic models for multivariate
time series of counts. Statistics and Its Interface, 7(4):559–570.
Rue, H. and Martino, S. (2007). Approximate Bayesian inference for hierarchical Gaussian Markov
random elds models. Journal of Statistical Planning and Inference, 137:3177–3192.
Rue, H., Martino, S., and Chopin, N. (2009). Approximate Bayesian inference for latent Gaussian
models by using integrated nested Laplace approximations. JournaloftheRoyalStatisticalSociety:
Series B, 71:319–392.
Serhiyenko, V., Mamun, S. A., Ivan, J. N. and Ravishanker, N. (2015). Fast Bayesian inference for
modeling multivariate crash counts on Connecticut limited access highways. Technical Report,
Department of Statistics, University of Connecticut: Storrs, CT.
Tsiamyrtzis, P. and Karlis, D. (2004). Strategies for efcient computation of multivariate Poisson
probabilities. Communications in Statistics - Simulation and Computation, 33:271–292.
..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset