Loading...
 
PDF Print

Non-statistical iterative reconstructions

With the previously introduced notations, the fundamental equation of the image reconstruction reads:
$\mathbf{y}=\mathbf{A}\mathbf{x}
Do not forget that the A matrix is a sparse matrix, not sqare and the number of its elements may be around 1012.

A Moore-Penrose generalized/psuedoinverse

 
As the matrix A is not square, most of the common methods for solving the equality (pl. Gauss-Siedel iteration) cannot directly be applied. Either when our system is overdetermined or underdetermined the criteria for the solution can be stated as:
$ \underset{\mathbf{x}}{min}\left \| \mathbf{y}-\mathbf{A}\boldsymbol{\mathbf{x}} \right \|

where ||.|| is a norm of our choice. Let us assume that $ \widetilde{\mathbf{x}} minimizes the above expression and let us choose the scalar product for the norm. Now for an arbitrary
$ \mathbf{v} vector and a t real number it must be true that
$ \partial _{t}\left \| \left (\mathbf{y}-\mathbf{A}\left (\widetilde{\mathbf{x}} -t\mathbf{v} \right )   \right )\right \|\mid _{t=0} =0
Calculating the derivative
$  \mathbf{A}\mathbf{v} \left (\mathbf{y}-\mathbf{A}\left (\widetilde{\mathbf{x}} -t\mathbf{v} \right )   \right )\mid _{t=0} =0
thus
$ \mathbf{v} \mathbf{A}^{T}\left (\mathbf{y}-\mathbf{A}\widetilde{\mathbf{x}}  \right )  =0
This must be true for every v so
$ \mathbf{A}^{T}\mathbf{y}= \mathbf{A}^{T}\mathbf{A}\widetilde{\mathbf{x}}

The marix ATA is a square matrix now,and if it is invertible:
$ \widetilde{\mathbf{x}}= \left (\mathbf{A}^{T}\mathbf{A}   \right )^{-1}\mathbf{A}^{T}\mathbf{y}
The A+ Moore-Penrose psudoinverse:
$ \mathbf{A}^{+}= \left (\mathbf{A}^{T}\mathbf{A}   \right )^{-1}\mathbf{A}^{T}

We can use for the calculationof A+ an UV decomposition, or any other regular method, can be calcuated by a simple though not too efficient iteration (Ben-Israel and Cohen iteration):
$\mathbf{A}_{(k)}^{+}=2\mathbf{A}_{(k-1)}^{+} +\mathbf{A}_{(k-1)}^{+}\mathbf{A}_{}\mathbf{A}_{(k-1)}^{+}

A Kaczmarz iteration

 
The system response matrix of currently available PET, SPECT and CT scanners is large compared to the easily accessable computer hardware, or/and computations on sparse matrices are not very efficient. A simple, stable and fast algorithm is the Kaczmarz iteration. The underlying idea is that a nxm matrix A is actually a
n number of equation of mdimensional hyperplanes and their crossing point is the x solution. If the system is overdetermined and some noise is added, perhaps the intersection point will not be really a point. If it is underdetermined it is possible that we can narrow down the solution only to a certain domain.
The essence of the Kaczmarz-iteráció is that we are trying to tend towards the domain of the solution by projecting perpendicularly to the hyperplane given as a single line of the matrix our previous x guess for a solution.
For example let us take a 2x2-es A matrix with elements aii.
Now the following two-equations system is given:

$ \left.\begin{matrix}
y_{1}=a_{11}x_{1}+a_{12}x_{1}
\\ 
y_{2}=a_{21}x_{2}+a_{22}x_{2}
\end{matrix}\right\}
where the crossing point is our solution. We have sketched on the graph the first 4 points of the iterations.

Image
The Kaczmarz iteration in two dimensions for 4 steps

 
In this 2D example it can be readily seen that

  • if the lines were perpendicular to each other i.e. the lines of the matrix were linearly indipendent, we would get the exact solution in two steps
  • if the lines were parallel, i.e. the set of eqations were contraversial the iteration would oscillate between the two lines
  • if the two lines were inclined only very narrowly, i.e. the rows were much correlated with each other the iteration would converge very slowly. The more independent they are the faster the method converges

 
A serious advantage of the Kaczmarz iteration that we only use a single line of the matrix for an iteration step, we do not need to handle the whole matrix together.

Now let us look at the steps of the projection now for a general case. Index i nominates which line of the matrix is being used, which row vector comes to play.
Let xk be teh solution obtained in the k-th step. Inthe k+1.
iteration step we should project this point on the hyperplane determined by equation i.The equation of this hyperplane reads:
$ \mathbf{A}^{i}\mathbf{x}=y_{i}
Earlier we hav seen, that such an equation describes a plane perpendicular to the direction vector $  \mathbf{A}^{i}/\sqrt{\mathbf{A}^{i}\mathbf{A}^{i}} , with distance from the origin of $  y_{i}/\sqrt{\mathbf{A}^{i}\mathbf{A}^{i}}. So, when in the kth step
we would like to project our solution to the i-th hyperplane, we need to go in the direction of $  \mathbf{A}^{i}/\sqrt{\mathbf{A}^{i}\mathbf{A}^{i}} , according to the convention of distance $  -\alpha.For simpler formulas let us include in this expression the normalisation of the unit vector, so now we can write the for next point:$ \mathbf{x}^{k+1}=\mathbf{x}^{k}-\alpha  \mathbf{A}^{i}

The next point must be on the i. next hyperplane, therefore:
$  \mathbf{A}^{i}\left (\mathbf{x}^{k}-\alpha  \mathbf{A}^{i}  \right )=y_{i}
Ebből kifejezhető $  \alpha
$  \alpha=\frac{\mathbf{A}^{i}\mathbf{x}^{k}-y_{i}}{\mathbf{A}^{i}\mathbf{A}^{i}}
The solution in the k+1. iteration steps therefore reads:
$  \mathbf{x}^{k+1}=\mathbf{x}^{k}-\frac{\mathbf{A}^{i}\mathbf{x}^{k}-y_{i}}{\mathbf{A}^{i}\mathbf{A}^{i}}   \mathbf{A}^{i}

The Kaczmarz-iteration in the medical diagnostic literature is mentioned also as ART (Algebraic Reconstruction Technique). The method is simple, but when applied only advanced versions are used. It is interesting that the Kaczmarz-iteration can be seen as an extremely low subset version of the OSEM iteration scheme.

Our next chapter would extend the matrix equation to probalistic interpretation of the elements and the measure of the distance.

 


Site Language: English

Log in as…