Figure 6.22 shows the unit circles using the l2 norm, l1 norm, and l0 quasi-norm, respectively. Except for the l2 norm unit circle, the other two unit circles do not look like circles at all. For a “measure” to be qualified as a norm, it needs to satisfy a set of axioms. A quasi-norm is almost a norm except that it does not satisfy an axiom called the triangle inequality.

image

Fig. 6.22: A unit circle is a trajectory of the points that have a distance 1 from the origin. Left: l2 norm’s unit circle. Middle: l1 norm’s unit circle. Right: l0 quasi-norm’s unit circle.

In theory, the l0 quasi-norm is the best measure to promote the sparsity of ψX and should be used. However, the optimization procedure is not tractable for any lp norm with 0 ≤ p < 1, because its associated objective function is not convex any more. The closest norm that produces a convex objective function is the l1 norm. The l1 norm minimization is nearly optimal, in the sense that its solution is not far from the solution with the much more complicated l0 quasi-norm minimization.

If the derivative of the Bayesian term exists, we can use the one-step late algorithm introduced in Section 6.9.6 to reconstruct the image. Unfortunately, the derivative of the Bayesian term in the objective function (6.88) does not exit. Rather complicated numerical methods are needed to minimize the objective function (6.88), even if the l0 quasi-norm is replaced by the l1 norm.

There is hope. Instead of using l0 quasi-norm and l1 norm, you may have also heard of the TV minimization. The TV (total variation) norm of an image v(x, y) is defined as

TV(v)=||v(x,y)||2dxdy=(vx)2+(vy)2dxdy.

The TV norm minimization enforces a flat image with the gradient being zero in most places. The resultant image tends to be piecewise constant. In practice, a small number ε is introduced to calculate the TV norm as

TV(v)(vx)2+(vy)2+ε2dxdy

so that the TV norm is differentiable. In order to get some intuition about the modified TV norm (6.90) expressed above, let us assume ε2 = 1. Then the above expression is nothing but the surface area formula!

We see that minimizing the modified TV norm is almost equivalent to minimizing the surface area. If a solution has a minimum surface area, it has the least oscillations. TV norm minimization avoids the l0 quasi-norm and l1 norm minimization.

Using Green’s one-step-late method presented in Section 6.9.6, a TV-regulated EM (TV-EM) algorithm can be obtained as

xinext=xjcurrentjaij+βTV(Xcurrent)xjcurrentiaijpiΣj^aij^xj^current.

For a 2D image X, if we express each pixel with double indices as xk,l, then the partial derivative in the above TV-EM algorithm can be given as

TV(X)xk,l=xk,lxk1,l(xk,lxk1,l)2+(xk1,l+1xk1,l)2+ε2+xk,lxk,l1(xk+1,l1xk,l1)2+(xk,lxk,l1)2+ε2xk+1,l+xk,l+12xk,l(xk+1,lxk,l)2+(xk,l+1xk,l)2+ε2.

Reconstruction with highly undersampled data is a Bayesian reconstruction problem. Performing the sparsifying transform ψX is one way to extract the prior knowledge from the image X. If other prior information about the image X is available, it should be included in the objective function as well.

6.11Worked examples

Example 1: For an arbitrary matrix A, its generalized inverse matrix A+ must satisfy the following four properties:

AA+A=A,A+AA+=A,(A+A)*=A+A,(AA+)*=AA+.

Here M is the Hermitian transpose (also called conjugate transpose) of a matrix M. For a real matrix, it is simply the transpose. In the following, we assume that matrix A is real.

Please verify that if (ATA)–1 exists, then A+ =(ATA)–1 AT is a generalized inverse matrix of A. Please also verify that if (AAT)–1 exists, then A+ = AT(AAT)–1 is a generalized inverse matrix of A.

Proof.

Case 1 : If(ATA)1exists,A+=(ATA)1AT.

Property 1:

Left=AA+A=A(ATA)1ATA=A(ATA)1(ATA)=A=Right

Property 2:

Left=A+AA+=(ATA)1ATA(ATA)1AT=(ATA)1(ATA)(ATA)1AT=(ATA)1AT=A+=Right

Property 3:

Left=(A+A)T=((ATA)1ATA)T=((ATA)1(ATA))T=IRight=A+A=(ATA)1ATA=(ATA)1(ATA)=I

Property 4:

Left=(AA+)T=(A(ATA)1AT)T=A((ATA)1)TAT=A((ATA)T)1AT=A(ATA)1AT=AA+=Right

Case 2 : If(AAT)1exits,A+=AT(AAT)1.

Property 1:

Left=AA+A=AAT(AAT)1A=(AAT)(AAT)A=A=Right

Property 2:

Left=A+AA=AT(AAT)1AAT(AAT)=AT(AAT)1(AAT)(AAT)1=AT(AAT)1=A+=Right

Property 3:

Left=(A+A)T=(AT(AAT)1A)T=AT((AAT)1)TA=AT((AAT)T)1A=AT(AAT)A=A+A=Right

Property 4:

Left=(AA+)T=(AAT(AAT)1)T=((AAT)(AAT)1)T=IRight=AA+=AAT(ATA)1=(AAT)(ATA)1=I

Example 2: In Section 6.1, we used SVD to decompose matrix Am×nintoAm×n=Um×mΣVn×nT and defined A+=VΣ+UT and Dr,=diag{1σ1,1σ2,,1σr,0,,0}.IsA+=VΣ+UT a generalized inverse of matrix A according the four properties stated in Example 1? Find the condition under which A+ = VG+UT is a generalized inverse of matrix A.

Solution

Let us apply the first property to A+=VΣ+UTandAm×n=Um×mΣVn×nT:

AA+A=UΣVTVΣ+UTUΣVT=UΣ(VTV)Σ+(UTU)ΣVT=UΣΣ+ΣVT.

In order to have AA+A = A+, we must have

ΣΣ+Σ=Σ,

which is equivalent to

diag{σ1,σ2,,σr,,σm}diag{1σ1,1σ2,,1σr,0,,0}diag{σ1,σ2,,σr,,σm}=diag{σ1,σ2,,σr,σm},

or

σr+1==σm=0.

Since

σ1σ2σl0,

Property 1 is equivalent to

diag{1σ1,1σ2,,1σr,0,,0}=diag{1σ1,1σ2,,1σr,0,,0}.

Now let us look at Property 2:

A+AA+=A+,

which is equivalent to

diag{1σ1,1σ2,,1σr,0,,0}=diag{1σ1,1σ2,,1σr,0,,0}.

This property is always satisfied.

Property 3 (A+A ) = A+A and Property 4 AA+ ∗ = AA+ both imply that

diag{1,1,,1,0,,0}=diag{1,1,,1,0,,0},

which is always satisfied.

To summarize the above discussion, the condition under which A+=VΣ+UT generalized inverse of matrix A is

σr+1=0andσr>0.

However, in practice, a cutoff index much smaller than this r is used to obtain a stable solution.

Example 3: None of the following images match the measured projections. Which one is the best solution among the three solutions in Figure 6.23? (Compare χ2)

Solution

i)χ2=02+02+12+12=2,(Best)ii)χ2=02+02+22+02=4,iii)χ2=12+02+12+22=6.

Example 4: Find a least-squares reconstruction to the imaging problem shown in Figure 6.24.

Solution

The imaging matrix A for this problem AX = P is given as

A=[1010010111000011].

The rank of A is 3. The system AX = P is also not consistent. Therefore, there is no solution for this problem. We can use SVD-based method to find a pseudo-solution. Using Matlab with X = pinv(A) ∗ P, we get x1 =2.25, x2 =1.75, x3 =1.75, and x4 =1.25.

Example 5: Let A=[2001]. Find the conjugate direction u1ofu0=[11].

Solution

The conjugate direction is defined by this relationship: u0(ATA)u1=0,, where

ATA=[2001][2001]=[4001].

image

Fig. 6.23: Three images try to match the projections.

image

Fig. 6.24: A2 × 2 image reconstruction problem.

image

Fig. 6.25: Conjugate directions.

We also know

[11][11]=0.

Thus [4001]u1=[11],which leads to

u1=[1/4001][11]=[1/41].

(If you want, you could normalize 1 and make it a unit vector.)

The matrix ATA defines a quadratic form, which is an ellipse in our case. The point of this example is as follows: For any initial direction 0, draw an ellipse specified by the quadratic form ATA such that 0 is a tangential direction. If you travel along the conjugate direction 1, you will reach the center of the ellipse (see Figure 6.25).

Example 6: Compute one iteration of the ML-EM algorithm with the initial image and projections given in Figure 6.26.

image

Fig. 6.26: Initial image and projection data for Example 6.

Solution

Step 1: Find the forward projections.

image

Step 2: Find the ratios of given (i.e., measured) projections and the forward projections of the current image.

image

Step 3: Backproject the ratios.

image

Step 4: Backproject a constant 1 from all rays.

image

Step 5: Find pixel-by-pixel ratio of the backprojected image obtained from Step 3 and the backprojected image from Step 4.

image

Step 6: Pixel-by-pixel update the current image (using point-by-point multiplication, not matrix multiplication).

image

Example 7: Use a Bayesian method to find a stable solution of the system

{x1+0.01x2=1.2x1+0.001x2=1.

Solution

If we consider the potential measurement errors, the system of equations can be written as

{x1+0.01x2=1.2+δ1x2+0.001x2=1+δ2.

The solution of this modified system is given as

{x1=0.9780.111δ1+1.11δ2x2=22.22+111.1δ1111.1δ2.

It is seen that x2 is sensitive to measurement noise. Let us assume a priori that “x1 and x2 are close” and solve this problem using the Bayesian method. Here, a prior is a probability distribution representing knowledge or belief about an unknown quantity a priori, that is, before any data have been observed. First, we set up an objective function and use the prior knowledge as a penalty term:

F(x1,x2)=(x1+0.01x21.2δ1)2+(x1+0.001x21δ2)2+β(x1+x2)2.

Note that β in this expression is not a Lagrange multiplier, but a preassigned constant. To minimize This results in a different problem:

[2+β0.011β0.011β0.012+0.0012+β][x1x2]=[(1.2+δ1)+(1+δ2)(0.01)(1.2+δ1)+(0.001)(1+δ2)]

The solution of this system depends on the value of β, as well as the values of δ1 and δ2. Some MAP solutions are listed in the following two tables:

Case 1: Noiseless (δ1=δ2=0)

image

Case 2: Noisy (δ1=δ2=0.2)

image

The system becomes more stable with an additional prior term in the objective function. The stability is reflected by much smaller condition numbers. This example also tells us that using a prior changes the original problem. Even for noiseless data, the MAP solutions may be different from the true solution. The Bayesian constraint β(x1x2)2 in eq. (6.124) is not a good one, and a better constraint should have been used. When you plan to use the Bayesian method, be careful and be sure that your prior knowledge is reasonable.

Example 8: Show that for the ML-EM algorithm at each iteration, the total number of the counts of the forward projections is the same as the total number of the counts of the original projection data:

i(jaijxj)=jpi.

Proof.

ijaijxj=ijaij(xjoldnanjnanjpnkankxkold)=jxjoldiaijnanjnanjpnkankxkold[changetheoderofsummations.]=jxjoldnanjpnkankxkold[iaij=nanj]=npnjanjxjoldkankxkold[iaij=nanj]=jpn.

Example 9: Use a computer simulation to demonstrate that the resolution recover rate is location dependent in the ML-EM algorithm.

Solution

A 2D computer-generated phantom shown in Figure 6.27 is used to generate projection data with and without Poisson noise. In data generation, attenuation, system blurring, and photon scattering are not included. That is, the projections are ideal line integrals of the object. The image size is 256 × 256, and 256 views uniformly distributed over 360° are used for projection data generation.

image

Fig. 6.27: The true image (a mathematical phantom).

image

Fig. 6.28: Iterative ML-EM reconstructions.

The iterative ML-EM algorithm is used for image reconstruction. Two reconstructed images are shown in Figure 6.28, one of which is obtained after 25 iterations, and the other is obtained after 250 iterations.

After 250 iterations, the algorithm has almost converged, and the resolution throughout the image is uniform. When the algorithm is stopped early at the 25th iteration, the resolution is not uniformly recovered. Higher resolution can be observed at the edge of the object. The resolution recovery rate is slower at the center. The motivation of early stopping is to regulate the noise. This is demonstrated with the noisy data reconstructions as shown in Figure 6.29, where two images are displayed. One of the images is obtained after 25 iterations and the other is obtained after 250 iterations.

The image with 250 iterations is noisier than that with 25 iterations. If we apply a low-pass filter to the noisy image obtained with 250 iterations, the image becomes less noisy and still maintains the uniform resolution throughout the image (see Figure 6.30). Therefore, it is a good strategy to over-iterate and then to apply a low-pass filter to control noise.

On the other hand, if the projections are ideal line integrals, the 2D parallel-beam FPB algorithm can provide images with uniform resolution (see Figure 6.31). However, if the imaging system has a spatially variant resolution and sensitivity, the conventional FBP algorithms are unable to model the system accurately and cannot provide uniform resolution in their reconstructions.

image

Fig. 6.29: Iterative ML-EM reconstructions with the same noisy projection data.

image

Fig. 6.30: Applying a low-pass filter to the image reconstructed with 250 iterations of ML-EM results in uniform resolution and reduced noise.

image

Fig. 6.31: Filtered backprojection reconstructions with noise-less and noisy data.

image

Fig. 6.32: Illustration of 2-pixel solutions with the minimum l2 norm, l1 norm, and l0 quasi-norm.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset