[Rtk-users] Model-based image reconstruction based on the RTK modules

Cyril Mory cyril.mory at creatis.insa-lyon.fr
Thu Nov 3 02:23:25 EDT 2016


Hello Vahid,

Thank you for this insight on Newton's methods. Yes, the output of the 
backprojection filter, in SART, is the gradient of the cost function. 
You may have noticed that the pipeline performs a division of the 
forward projection by something coming from "RayBoxIntersectionFilter". 
This is to normalize the forward projection, so that R^T R f ~= blurry 
f. If you don't do it, you'll have R^T R f ~= alpha * blurry f, with 
alpha that can reach 200 or so, and your algorithm will quickly diverge.

You could try to extract the gradient from the conjugate gradient filter 
as well, but it is significantly more tricky: first, the CG filter was 
implemented from the following algorithm, taken from wikipedia (which 
embeds the normalization I was mentioning earlier):

{\begin{aligned}&\mathbf {r} _{0}:=\mathbf {b} -\mathbf {Ax} 
_{0}\\&\mathbf {p} _{0}:=\mathbf {r} 
_{0}\\&k:=0\\&{\hbox{repeat}}\\&\qquad \alpha _{k}:={\frac {\mathbf {r} 
_{k}^{\mathsf {T}}\mathbf {r} _{k}}{\mathbf {p} _{k}^{\mathsf 
{T}}\mathbf {Ap} _{k}}}\\&\qquad \mathbf {x} _{k+1}:=\mathbf {x} 
_{k}+\alpha _{k}\mathbf {p} _{k}\\&\qquad \mathbf {r} _{k+1}:=\mathbf 
{r} _{k}-\alpha _{k}\mathbf {Ap} _{k}\\&\qquad {\hbox{if 
}}r_{k+1}{\hbox{ is sufficiently small then exit loop}}\\&\qquad \beta 
_{k}:={\frac {\mathbf {r} _{k+1}^{\mathsf {T}}\mathbf {r} 
_{k+1}}{\mathbf {r} _{k}^{\mathsf {T}}\mathbf {r} _{k}}}\\&\qquad 
\mathbf {p} _{k+1}:=\mathbf {r} _{k+1}+\beta _{k}\mathbf {p} 
_{k}\\&\qquad k:=k+1\\&{\hbox{end repeat}}\\&{\hbox{The result is 
}}\mathbf {x} _{k+1}\end{aligned}}

In this algorithm, it is not clear to me what exactly is the gradient of 
the cost function. I would say it is something like "- r_k", but I'm not 
sure. Second, as you see, the CG filter needs an operator A, which may 
differ from one problem to another, so this operator is implemented in a 
separate filter, which in your case would be 
rtkReconstructionConjugateGradientOperator, with the laplacian 
regularization parameter gamma set to 0.

Note that we never actually store the system matrix R. Instead, the 
interpolation coefficient it contains are re-computed on the fly 
everytime we forward project. And the same holds for backprojection, i.e 
the matrix R^T.

Best,

Cyril


On 11/03/2016 03:38 AM, vahid ettehadi wrote:
> Hello Simon and Cyril,
> Thanks for the reply.
> You are right Simon. I did not notice it too in the literature. The 
> main problem as you said is the storage. Actually I developed  the 
> conjugate gradient (CG), quasi-Newton and Newton optimization methods 
> for optical tomography and I intended to apply them to the CT 
> reconstruction as well. I implemented the Newton's methods 
> (Gauss-Newton and Levenberg-Marquardt) in a 
> Jacobian-Free-Newton-Krylov approaches to avoid the matrix 
> multiplication of Jacobians (sensitivity). It means we only need to 
> store the Jacobian matrix for the these methods (the matrix R that 
> Cyril was mentioned), that is still a big matrix for practical 
> problems in CT reconstruction. For the quasi-Newton I adapted an 
> L-BFGS algorithm that only need the 3 or 8 last iterations of the 
> gradient vector to calculate the Hessian matrix. In my case, the 
> L-BFGS and Newton's methods was much faster than the CG as you know 
> because of using the second order derivative (hessian matrix). I saw 
> in your last paper you implement the conjugate gradient method, so I 
> thought it might be easy to extract the gradient vector from CG 
> modules and solve the cost function within the quasi-Newton/Newton 
> methods. I will look at the codes to see what I can do.
> Thanks again for the reply.
>
> @Cyril:
> Please correct me if I am wrong. you mean the output of 
> backProjectionFilter is the gradient of defined cost function?
>
> Regards,
> Vahid
>
>
> On Wednesday, November 2, 2016 2:53 AM, Cyril Mory 
> <cyril.mory at creatis.insa-lyon.fr> wrote:
>
>
> Hi Vahid,
> Welcome to RTK :)
> Indeed, there are several iterative methods already implemented in 
> RTK, but none of the filters allows you to easily extract the gradient 
> of the least squares function there are minimizing.
> If you need to minimize the classical non-regularized tomographic cost 
> function, ie || R f - p ||², with R the forward projection operator, f 
> the volume you are looking for, and p the measured projections, my 
> best advice would be to copy some part of the pipeline of 
> rtkSARTConeBeamReconstructionFilter to get the job done, ie the 
> following part (copy-paste this into webgraphviz.com)
>
> digraph SARTConeBeamReconstructionFilter {
>
> Input0 [ label="Input 0 (Volume)"];
> Input0 [shape=Mdiamond];
> Input1 [label="Input 1 (Projections)"];
> Input1 [shape=Mdiamond];
>
> node [shape=box];
> ForwardProject [ label="rtk::ForwardProjectionImageFilter" URL="\ref 
> rtk::ForwardProjectionImageFilter"];
> Extract [ label="itk::ExtractImageFilter" URL="\ref 
> itk::ExtractImageFilter"];
> MultiplyByZero [ label="itk::MultiplyImageFilter (by zero)" URL="\ref 
> itk::MultiplyImageFilter"];
> AfterExtract [label="", fixedsize="false", width=0, height=0, shape=none];
> Subtract [ label="itk::SubtractImageFilter" URL="\ref 
> itk::SubtractImageFilter"];
> MultiplyByLambda [ label="itk::MultiplyImageFilter (by lambda)" 
> URL="\ref itk::MultiplyImageFilter"];
> Divide [ label="itk::DivideOrZeroOutImageFilter" URL="\ref 
> itk::DivideOrZeroOutImageFilter"];
> GatingWeight [ label="itk::MultiplyImageFilter (by gating weight)" 
> URL="\ref itk::MultiplyImageFilter", style=dashed];
> Displaced [ label="rtk::DisplacedDetectorImageFilter" URL="\ref 
> rtk::DisplacedDetectorImageFilter"];
> ConstantProjectionStack [ label="rtk::ConstantImageSource" URL="\ref 
> rtk::ConstantImageSource"];
> ExtractConstantProjection [ label="itk::ExtractImageFilter" URL="\ref 
> itk::ExtractImageFilter"];
> RayBox [ label="rtk::RayBoxIntersectionImageFilter" URL="\ref 
> rtk::RayBoxIntersectionImageFilter"];
> ConstantVolume [ label="rtk::ConstantImageSource" URL="\ref 
> rtk::ConstantImageSource"];
> BackProjection [ label="rtk::BackProjectionImageFilter" URL="\ref 
> rtk::BackProjectionImageFilter"];
> OutofInput0 [label="", fixedsize="false", width=0, height=0, shape=none];
> OutofBP [label="", fixedsize="false", width=0, height=0, shape=none];
> BeforeBP [label="", fixedsize="false", width=0, height=0, shape=none];
> BeforeAdd [label="", fixedsize="false", width=0, height=0, shape=none];
> Input0 -> OutofInput0 [arrowhead=none];
> OutofInput0 -> ForwardProject;
> ConstantVolume -> BeforeBP [arrowhead=none];
> BeforeBP -> BackProjection;
> Extract -> AfterExtract[arrowhead=none];
> AfterExtract -> MultiplyByZero;
> AfterExtract -> Subtract;
> MultiplyByZero -> ForwardProject;
> Input1 -> Extract;
> ForwardProject -> Subtract;
> Subtract -> MultiplyByLambda;
> MultiplyByLambda -> Divide;
> Divide -> GatingWeight;
> GatingWeight -> Displaced;
> ConstantProjectionStack -> ExtractConstantProjection;
> ExtractConstantProjection -> RayBox;
> RayBox -> Divide;
> Displaced -> BackProjection;
> BackProjection -> OutofBP [arrowhead=none];
> }
>
> As you can see, it is a very large part of the SART reconstruction 
> filter, so yoiu might be better off just copying the whole 
> SARTConeBeamReconstructionFilter and modifying it.
>
> Of course, you could also look into ITK's cost function class, and see 
> if one of the classes inherited from it suits your needs, implement 
> your cost function this way, and use ITK's off-the-shelf solvers to 
> minimize it. See the inheritance diagram in 
> https://itk.org/Doxygen/html/classitk_1_1CostFunctionTemplate.html if 
> you want to try this approach.
>
> Best regards,
> Cyril
>
> On 11/01/2016 05:50 PM, vahid ettehadi via Rtk-users wrote:
>> Hello RTK users and developers,
>>
>> I already implemented the RTK and reconstructed some images with the 
>> FDK algorithm implemented in RTK. It works well. Thanks to RTK 
>> developers.
>> Now, I am trying to develop a model-based image reconstruction for 
>> our cone-beam micro-CT. I see already that some iterative algorithms 
>> like ART and its modifications and conjugate-gradient (CG) method are 
>> implemented in the RTK. I want to develop a model-based 
>> reconstruction through the Newton/quasi-Newton optimizations methods. 
>> I was wondering is it possible to extract the gradient of least 
>> square function from implemented algorithms like CG module? Any 
>> recommendation will be appreciated.
>>
>> Best Regards,
>> Vahid
>>
>>
>>
>> _______________________________________________
>> Rtk-users mailing list
>> Rtk-users at public.kitware.com <mailto:Rtk-users at public.kitware.com>
>> http://public.kitware.com/mailman/listinfo/rtk-users
>
>
> _______________________________________________
> Rtk-users mailing list
> Rtk-users at public.kitware.com <mailto:Rtk-users at public.kitware.com>
> http://public.kitware.com/mailman/listinfo/rtk-users
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://public.kitware.com/pipermail/rtk-users/attachments/20161103/1a63c9c0/attachment-0010.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: e300dfefdbd374cdee765397528a65a5736a50d3
Type: image/svg+xml
Size: 29828 bytes
Desc: not available
URL: <http://public.kitware.com/pipermail/rtk-users/attachments/20161103/1a63c9c0/attachment-0010.svg>


More information about the Rtk-users mailing list