86 ADVANCED TECHNIQUES USING COALESCENCE
dx|X
t
= x)=
π
(dy)P(X
t+1
∈dx|X
t
= y),whichiswhyinDefinition 1.34 suc h chains
were labe led reversible.
Okay, so now we know how to run the chain backwards in time. Suppose in our
simulation, X
5
= 0, X
4
= 1, X
3
= 2, X
2
= 1, X
1
= 0, and X
0
= 0. Now look at how
the chain can be run forward in time conditioned on the path (0,0,1,2,1,0) using the
updates X
t
=
φ
(X
t−1
,U
t
).
In the first step X
1
= 0givenX
0
= 0. That means that U
1
≤ 2/3. So generate U
1
uniformly from [0, 2/3]. Similarly, since X
1
= 0andX
2
= 1, U
2
> 2 /3. The rest of
the U
i
can be bounded similarly.
Note that regardless of the value of x
0
, setting x
t
=
φ
(x
t−1
,U
t
) using U
1
≤ 2 /3,
U
2
> 2/3, U
3
> 2/3, U
4
≤2/3, and U
5
≤2/3givesx
5
= 0 = X
5
. Hence this gives an
S
x
block.
On the other hand, suppose (X
0
,X
1
,X
2
,X
3
,X
4
,X
5
)=(1,0 , 1, 0,1,0). Then gener-
ating the U
i
gives a final state of 1 if x
0
= 1and0ifx
0
= 0. So that is an example o f
an F block.
In practice, it is not necessary to retain the entire path (X
0
,...,X
t
). As the pair
(X
t
−1
,X
t
) the value of U
t
can be imputed and stored. So only the (U
0
,...,U
t
) need
be re tained.
5.2.1 Reversing a Markov chain
In the random walk on {0, 1,2}example, it is was easy to determine how the reversed
walk process behaved. For instance, if X
t+1
= X
t
+ 1, then in the reverse process
X
t
= X
t+1
−1, and so on. Now the general method for reversing a Markov chain is
discussed.
Recall Definitio n 1.27 of a Markov kernel K requires that for all x ∈ Ω, K(x,·) is
a probability distribution, and that for all measurable A and a ∈[0,1], the set of states
x such that K(x,A) ≤ a is also measurable.
In other words, what the kernel K gives us is a family of probability distributions.
Given the current state x, K(x,·) is the probability distribution for the next state y.
(The second part of the definition is necessary for technical reasons.)
Now the notion of a reverse kernel with respect to a distribution
π
can be created.
Definition 5.1. Let K be a kernel from (Ω,F ) to itself, and
π
a probability mea-
sur e on (Ω,F ).ThenK
rev
is a reverse kernel associated with
π
if for all bounded
measurable functions f defined on Ω ×Ω
(x,y)∈Ω×Ω
f (x, y)
ν
(dx)K(x, dy)=
(x,y)∈Ω×Ω
f (x, y)
ν
(dy)K
rev
(y,dx).
Note that in general the reverse kernel might not even exist, and if it does, is not
always uniquely defined!
While the definitio n above is presented in an abstract fashion, the notion that it is
capturing is more straightforward. Suppose X
t
∼
π
,andthenX
t+1
comes from taking
one step in the Markov chain. Now suppose the value of X
t
is lost. What is the new
distribution of X
t
given X
t+1
? This is what the reverse kernel gives us.
As was seen in the previous section, for the biased simple random walk on