1
Precision Learning Computed Tomography Tobias Würfl, Mathis Hoffmann, Christopher Syben, Katharina Breininger, Yixing Huang, Andreas K.Maier Pattern Recognition Lab, Friedrich-Alexander University Erlangen-Nürnberg, Erlangen,Germany Precision Learning Concept [1] Figure 1: Known operators embedded in a neural network Idea: Include known operators in learnable mappings Deduce network topologies using domain knowledge Benefit: Faster convergence Domain knowledge restricts every block of the network Precision Learning for Computed Tomography [2] Most reconstruction algorithms are chains of operators. Architectures can be derived and initialized to be identical. sinogram projection C A T pb non-neg. constraint convolution layer fully connected layer rectified linear unit reconstruction loss function Figure 2: Neural network architecture equivalent to the filtered backprojection algorithm An advantage is that weights can be adapted in a data-driven fashion. Comparison to Deep-learning-based Reconstruction [3] Competing deep learning networks as post-processing show great results, but lack interpretability and can produce unreliable results. (a) (b) (c) (d) (e) (f) Figure 3: Fig. 3a: Limited-angle reconstruction; Fig. 3b: Corresponding groundtruth; Fig. 3c: U-net reconstruction; Figures 3d, 3e and 3f: Same images using different window Application to Limited Angle Cone-beam CT [4] Loss of mass in limited angle CT can be compensated using heuristic weig- hts. Using precision learning a data-optimal solution can be learned. sinogram projection W cos 2D W red C A T cb non-neg. constraint weighting layer weighting layer conv. layer FC layer ReLU reconstruction loss function Figure 4: Compensation weights can be learned to compensate for missing data. Benefits are the increased robustness and interpretability by design. Precision learning methods are less prone to disturbances like noise or different object appearance when compared to deep learning methods. (a) (b) Figure 5: Fig. 5a: Limited-angle reconstruction with simulated noise; Fig. 5b: Reconstruction using the learned compensation weights Deriving an Architecture for Rebinning [5] We can also hypothesize whole algorithms and learn unknown elements. parallel-beam sinogram projection F K F H A T p A f S non-neg. constraint Fourier transform layer weighting layer inverse Fourier transform layer parallel Back- projection layer fan-beam projection layer weighting layer ReLU fan-beam projection loss function Figure 6: Parallel-beam to fan-beam projection rebinning can be formulated as an efficient convolution-based solution, with a filter learned in a data-driven manner Learned filters produce sharper images faster than classical algorithms. References [1] A. Maier, F. Schebesch, C. Syben, T. Würfl, S. Steidl, J.-H. Choi, and R. Fahrig, “Precision Learning: Towards Use of Known Operators in Neural Networks,” [2] T. Würfl, F. C. Ghesu, V. Christlein, and A. Maier, “Deep learning computed tomography,” [3] Y. Huang, T. Würfl, K. Breininger, L. Liu, G. Lauritsch, and A. Maier, “Some investigations on robustness of deep learning in limited angle tomography,” [4] T. Würfl, M. Hoffmann, V. Christlein, K. Breininger, Y. Huang, M. Unberath, and A. K. Maier, “Deep learning computed tomography: Learning projection-domain weights from image domain in limited angle problems,” IEEE transactions on medical imaging. [5] C. Syben, B. Stimpel, J. Lommen, T. Würfl, A. Dörfler, and A. Maier, “Deriving neural network architectures using precision learning: Parallel-to-fan beam conversion,” Acknowledgements Contact Tobias Würfl Pattern Recognition Lab Friedrich-Alexander University Erlangen-Nürnberg Erlangen, Germany +49 9131 85 27799 tobias.wuerfl@fau.de

0.94!Precision Learning Computed Tomography

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: 0.94!Precision Learning Computed Tomography

Precision Learning Computed TomographyTobias Würfl, Mathis Hoffmann, Christopher Syben, Katharina Breininger, Yixing Huang, Andreas K.MaierPattern Recognition Lab, Friedrich-Alexander University Erlangen-Nürnberg, Erlangen,Germany

Precision Learning Concept [1]

Figure 1: Known operators embedded in a neural network

Idea:• Include known operators in learnable mappings• Deduce network topologies using domain knowledge

Benefit:• Faster convergence• Domain knowledge restricts every block of the network

Precision Learning for Computed Tomography [2]

Most reconstruction algorithms are chains of operators. Architectures canbe derived and initialized to be identical.

sin

ogra

m proj

ecti

on

C ATpb non-neg. constraint

convolution layer fully connected layer rectified linear unit

reco

nst

ruct

ion

lossfunction

Figure 2: Neural network architecture equivalent to the filtered backprojection algorithm

An advantage is that weights can be adapted in a data-driven fashion.

Comparison to Deep-learning-based Reconstruction [3]

Competing deep learning networks as post-processing show great results,but lack interpretability and can produce unreliable results.

(a) (b) (c)

(d) (e) (f)

Figure 3: Fig. 3a: Limited-angle reconstruction; Fig. 3b: Corresponding groundtruth;Fig. 3c: U-net reconstruction; Figures 3d, 3e and 3f: Same images using different window

Application to Limited Angle Cone-beam CT [4]

Loss of mass in limited angle CT can be compensated using heuristic weig-hts. Using precision learning a data-optimal solution can be learned.

sinogram projection

Wcos2D Wred C ATcb non-neg. constraint

weighting layer weighting layer conv. layer FC layer ReLU

reconstruction

lossfunction

Figure 4: Compensation weights can be learned to compensate for missing data.Benefits are the increased robustness and interpretability by design.

Precision learning methods are less prone to disturbances like noise ordifferent object appearance when compared to deep learning methods.

(a) (b)

Figure 5: Fig. 5a: Limited-angle reconstruction with simulated noise; Fig. 5b:Reconstruction using the learned compensation weights

Deriving an Architecture for Rebinning [5]

We can also hypothesize whole algorithms and learn unknown elements.

parallel-beam

sinogram

projection

F K FH ATp Af S non-neg. constraint

Fourier transformlayer

weighting layer inverse Fouriertransform layer

parallel Back-projection layer

fan-beamprojection layer

weighting layer ReLU

fan-beam

projection

lossfunction

Figure 6: Parallel-beam to fan-beam projection rebinning can be formulated as anefficient convolution-based solution, with a filter learned in a data-driven manner

Learned filters produce sharper images faster than classical algorithms.

References

[1] A. Maier, F. Schebesch, C. Syben, T. Würfl, S. Steidl, J.-H. Choi, and R. Fahrig,“Precision Learning: Towards Use of Known Operators in Neural Networks,”

[2] T. Würfl, F. C. Ghesu, V. Christlein, and A. Maier, “Deep learning computedtomography,”

[3] Y. Huang, T. Würfl, K. Breininger, L. Liu, G. Lauritsch, and A. Maier, “Someinvestigations on robustness of deep learning in limited angle tomography,”

[4] T. Würfl, M. Hoffmann, V. Christlein, K. Breininger, Y. Huang, M. Unberath, and A. K.Maier, “Deep learning computed tomography: Learning projection-domain weightsfrom image domain in limited angle problems,” IEEE transactions on medical imaging.

[5] C. Syben, B. Stimpel, J. Lommen, T. Würfl, A. Dörfler, and A. Maier, “Deriving neuralnetwork architectures using precision learning: Parallel-to-fan beam conversion,”

AcknowledgementsContact Tobias Würfl

Pattern Recognition LabFriedrich-Alexander University Erlangen-NürnbergErlangen, Germany� +49 9131 85 27799� [email protected]