Figure 6.22 shows the unit circles using the l2 norm, l1 norm, and l0 quasi-norm, respectively. Except for the l2 norm unit circle, the other two unit circles do not look like circles at all. For a “measure” to be qualified as a norm, it needs to satisfy a set of axioms. A quasi-norm is almost a norm except that it does not satisfy an axiom called the triangle inequality.
Fig. 6.22: A unit circle is a trajectory of the points that have a distance 1 from the origin. Left: l2 norm’s unit circle. Middle: l1 norm’s unit circle. Right: l0 quasi-norm’s unit circle.
In theory, the l0 quasi-norm is the best measure to promote the sparsity of ψX and should be used. However, the optimization procedure is not tractable for any lp norm with 0 ≤ p < 1, because its associated objective function is not convex any more. The closest norm that produces a convex objective function is the l1 norm. The l1 norm minimization is nearly optimal, in the sense that its solution is not far from the solution with the much more complicated l0 quasi-norm minimization.
If the derivative of the Bayesian term exists, we can use the one-step late algorithm introduced in Section 6.9.6 to reconstruct the image. Unfortunately, the derivative of the Bayesian term in the objective function (6.88) does not exit. Rather complicated numerical methods are needed to minimize the objective function (6.88), even if the l0 quasi-norm is replaced by the l1 norm.
There is hope. Instead of using l0 quasi-norm and l1 norm, you may have also heard of the TV minimization. The TV (total variation) norm of an image v(x, y) is defined as
The TV norm minimization enforces a flat image with the gradient being zero in most places. The resultant image tends to be piecewise constant. In practice, a small number ε is introduced to calculate the TV norm as
so that the TV norm is differentiable. In order to get some intuition about the modified TV norm (6.90) expressed above, let us assume ε2 = 1. Then the above expression is nothing but the surface area formula!
We see that minimizing the modified TV norm is almost equivalent to minimizing the surface area. If a solution has a minimum surface area, it has the least oscillations. TV norm minimization avoids the l0 quasi-norm and l1 norm minimization.
Using Green’s one-step-late method presented in Section 6.9.6, a TV-regulated EM (TV-EM) algorithm can be obtained as
For a 2D image X, if we express each pixel with double indices as xk,l, then the partial derivative in the above TV-EM algorithm can be given as
Reconstruction with highly undersampled data is a Bayesian reconstruction problem. Performing the sparsifying transform ψX is one way to extract the prior knowledge from the image X. If other prior information about the image X is available, it should be included in the objective function as well.
Example 1: For an arbitrary matrix A, its generalized inverse matrix A+ must satisfy the following four properties:
Here M∗ is the Hermitian transpose (also called conjugate transpose) of a matrix M. For a real matrix, it is simply the transpose. In the following, we assume that matrix A is real.
Please verify that if (ATA)–1 exists, then A+ =(ATA)–1 AT is a generalized inverse matrix of A. Please also verify that if (AAT)–1 exists, then A+ = AT(AAT)–1 is a generalized inverse matrix of A.
Proof.
Case 1 :
Property 1:
Property 2:
Property 4:
Case 2 :
Property 1:
Property 2:
Property 3:
Property 4:
Example 2: In Section 6.1, we used SVD to decompose matrix and defined and a generalized inverse of matrix A according the four properties stated in Example 1? Find the condition under which A+ = VG+UT is a generalized inverse of matrix A.
Let us apply the first property to
In order to have AA+A = A+, we must have
which is equivalent to
or
Since
Property 1 is equivalent to
Now let us look at Property 2:
which is equivalent to
This property is always satisfied.
Property 3 (A+A ∗) = A+A and Property 4 AA+ ∗ = AA+ both imply that
which is always satisfied.
To summarize the above discussion, the condition under which generalized inverse of matrix A is
However, in practice, a cutoff index much smaller than this r is used to obtain a stable solution.
Example 3: None of the following images match the measured projections. Which one is the best solution among the three solutions in Figure 6.23? (Compare χ2)
Solution
Example 4: Find a least-squares reconstruction to the imaging problem shown in Figure 6.24.
Solution
The imaging matrix A for this problem AX = P is given as
The rank of A is 3. The system AX = P is also not consistent. Therefore, there is no solution for this problem. We can use SVD-based method to find a pseudo-solution. Using Matlab with X = pinv(A) ∗ P, we get x1 =2.25, x2 =1.75, x3 =1.75, and x4 =1.25.
Example 5: Let Find the conjugate direction
Solution
The conjugate direction is defined by this relationship: , where
Fig. 6.23: Three images try to match the projections.
Fig. 6.24: A2 × 2 image reconstruction problem.
Fig. 6.25: Conjugate directions.
We also know
Thus which leads to
(If you want, you could normalize 1 and make it a unit vector.)
The matrix ATA defines a quadratic form, which is an ellipse in our case. The point of this example is as follows: For any initial direction 0, draw an ellipse specified by the quadratic form ATA such that 0 is a tangential direction. If you travel along the conjugate direction 1, you will reach the center of the ellipse (see Figure 6.25).
Example 6: Compute one iteration of the ML-EM algorithm with the initial image and projections given in Figure 6.26.
Fig. 6.26: Initial image and projection data for Example 6.
Solution
Step 1: Find the forward projections.
Step 2: Find the ratios of given (i.e., measured) projections and the forward projections of the current image.
Step 3: Backproject the ratios.
Step 4: Backproject a constant 1 from all rays.
Step 5: Find pixel-by-pixel ratio of the backprojected image obtained from Step 3 and the backprojected image from Step 4.
Step 6: Pixel-by-pixel update the current image (using point-by-point multiplication, not matrix multiplication).
Example 7: Use a Bayesian method to find a stable solution of the system
Solution
If we consider the potential measurement errors, the system of equations can be written as
The solution of this modified system is given as
It is seen that x2 is sensitive to measurement noise. Let us assume a priori that “x1 and x2 are close” and solve this problem using the Bayesian method. Here, a prior is a probability distribution representing knowledge or belief about an unknown quantity a priori, that is, before any data have been observed. First, we set up an objective function and use the prior knowledge as a penalty term:
Note that β in this expression is not a Lagrange multiplier, but a preassigned constant. To minimize This results in a different problem:
The solution of this system depends on the value of β, as well as the values of δ1 and δ2. Some MAP solutions are listed in the following two tables:
Case 1: Noiseless
Case 2: Noisy
The system becomes more stable with an additional prior term in the objective function. The stability is reflected by much smaller condition numbers. This example also tells us that using a prior changes the original problem. Even for noiseless data, the MAP solutions may be different from the true solution. The Bayesian constraint β(x1 – x2)2 in eq. (6.124) is not a good one, and a better constraint should have been used. When you plan to use the Bayesian method, be careful and be sure that your prior knowledge is reasonable.
Example 8: Show that for the ML-EM algorithm at each iteration, the total number of the counts of the forward projections is the same as the total number of the counts of the original projection data:
Proof.
Example 9: Use a computer simulation to demonstrate that the resolution recover rate is location dependent in the ML-EM algorithm.
Solution
A 2D computer-generated phantom shown in Figure 6.27 is used to generate projection data with and without Poisson noise. In data generation, attenuation, system blurring, and photon scattering are not included. That is, the projections are ideal line integrals of the object. The image size is 256 × 256, and 256 views uniformly distributed over 360° are used for projection data generation.
Fig. 6.27: The true image (a mathematical phantom).
Fig. 6.28: Iterative ML-EM reconstructions.
The iterative ML-EM algorithm is used for image reconstruction. Two reconstructed images are shown in Figure 6.28, one of which is obtained after 25 iterations, and the other is obtained after 250 iterations.
After 250 iterations, the algorithm has almost converged, and the resolution throughout the image is uniform. When the algorithm is stopped early at the 25th iteration, the resolution is not uniformly recovered. Higher resolution can be observed at the edge of the object. The resolution recovery rate is slower at the center. The motivation of early stopping is to regulate the noise. This is demonstrated with the noisy data reconstructions as shown in Figure 6.29, where two images are displayed. One of the images is obtained after 25 iterations and the other is obtained after 250 iterations.
The image with 250 iterations is noisier than that with 25 iterations. If we apply a low-pass filter to the noisy image obtained with 250 iterations, the image becomes less noisy and still maintains the uniform resolution throughout the image (see Figure 6.30). Therefore, it is a good strategy to over-iterate and then to apply a low-pass filter to control noise.
On the other hand, if the projections are ideal line integrals, the 2D parallel-beam FPB algorithm can provide images with uniform resolution (see Figure 6.31). However, if the imaging system has a spatially variant resolution and sensitivity, the conventional FBP algorithms are unable to model the system accurately and cannot provide uniform resolution in their reconstructions.
Fig. 6.29: Iterative ML-EM reconstructions with the same noisy projection data.
Fig. 6.30: Applying a low-pass filter to the image reconstructed with 250 iterations of ML-EM results in uniform resolution and reduced noise.
Fig. 6.31: Filtered backprojection reconstructions with noise-less and noisy data.
Fig. 6.32: Illustration of 2-pixel solutions with the minimum l2 norm, l1 norm, and l0 quasi-norm.