5
A SPLITTING-BASED ITERATIVE ALGORITHM FOR GPU-ACCELERATED STATISTICAL DUAL-ENERGY X-RAY CT RECONSTRUCTION Fangda Li Ankit Manerikar Tanmay Prakash Avinash Kak School of Electrical and Computer Engineering, Purdue University {li1208, amanerik, tprakash, kak}@purdue.edu ABSTRACT When dealing with material classification in baggage at airports, Dual-Energy Computed Tomography (DECT) allows characterization of any given material with coeffi- cients based on two attenuative effects: Compton scatter- ing and photoelectric absorption. However, straightforward projection-domain decomposition methods for this charac- terization often yield poor reconstructions due to the high dynamic range of material properties encountered in an ac- tual luggage scan. Hence, for better reconstruction quality under a timing constraint, we propose a splitting-based, GPU- accelerated, statistical DECT reconstruction algorithm. Com- pared to prior art, our main contribution lies in the significant acceleration made possible by separating reconstruction and decomposition within an ADMM framework. Experimental results, on both synthetic and real-world baggage phantoms, demonstrate a significant reduction in time required for con- vergence. Index TermsDECT, ADMM, GPU 1. INTRODUCTION X-ray Computed Tomography (CT) is a widely deployed technique for threat detection in baggage at airport security checkpoints. However, using a single X-ray spectrum lim- its its reconstruction to only that of LAC (linear attenuation coefficient) or HU (Hounsfield Unit) images which are, at best, an approximation to the underlying energy-dependent characteristics of the materials. While LAC might suffice for discriminating between different types of tissues in medical imaging, in the adversarial case of airport baggage scan- ning, materials used commonly for homemade explosives can possess LAC values that are nearly the same as for benign materials. As a result, the technique of DECT, where pro- jections from two X-ray spectra are collected simultaneously at each angle, has been proposed for material discrimination because it allows for the material property to be recovered in an additional dimension by using an energy-dependent attenuation model. This research is funded by the BAA 17-03 AATR contract with Depart- ment of Homeland Security. The most commonly used DECT model represents X-ray attenuation as the combined effect of Compton scattering and photoelectric absorption, which can be written as: μ(E,x)= x c f KN (E)+ x p f p (E), (1) where f KN (E) and f p (E) denote the energy-dependent mul- tipliers 1 to the Compton coefficient x c and the photoelec- tric (PE) coefficient x p , respectively. Therefore, the goal of DECT is to recover both the Compton and the PE coefficients of the object using the projections m h , m l R M measured at two different X-ray spectra. This amounts to solving simul- taneously for x c , x p R N from the following two equations that can be shown compactly by: m h/l = - ln Z S h/l (E)e -(E) dE + ln Z S h/l (E)dE, (2) where R denotes the forward projection matrix, and S h (E) and S l (E) denote, respectively, the photon distribution across energy levels in the high- or low-energy spectra. Solving the two equations shown above is made challeng- ing by the fact the two phenomena that contribute to X-ray attenuation — Compton and PE, occur at grossly different scales. At most applicable energy levels, Compton scattering is the dominant contributor to attenuation and this disparity between the two phenomena becomes worse with increasing energy since f p (E) has a cubic decay with E. As a result, stable recovery of PE coefficients requires more sophisticated inversion algorithms, such as those described in [2, 3, 4, 5]. Unfortunately, the algorithms cited above tend to be it- erative and, with run-of-the-mill computing hardware, take a long time to return the results. Therefore, a straightforward implementation of these algorithms is not appropriate for the end-goals that motivate our research — high-throughput bag- gage screening at airports. The focus of the concepts pre- sented in this paper is on improving the computational effi- ciency of an ADMM-based statistical inversion algorithm. 2. RELATED WORK DECT involves two key steps: dual-energy decomposition and tomographic reconstruction. The two tasks can be done 1 See [1] for a detailed formulation of f KN (E) and fp(E). arXiv:1905.00934v1 [eess.IV] 2 May 2019

A SPLITTING-BASED ITERATIVE ALGORITHM FOR GPU ...projection-domain decomposition methods for this charac-terization often yield poor reconstructions due to the high dynamic range of

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: A SPLITTING-BASED ITERATIVE ALGORITHM FOR GPU ...projection-domain decomposition methods for this charac-terization often yield poor reconstructions due to the high dynamic range of

A SPLITTING-BASED ITERATIVE ALGORITHM FOR GPU-ACCELERATED STATISTICALDUAL-ENERGY X-RAY CT RECONSTRUCTION

Fangda Li Ankit Manerikar Tanmay Prakash Avinash Kak

School of Electrical and Computer Engineering, Purdue University{li1208, amanerik, tprakash, kak}@purdue.edu

ABSTRACT

When dealing with material classification in baggageat airports, Dual-Energy Computed Tomography (DECT)allows characterization of any given material with coeffi-cients based on two attenuative effects: Compton scatter-ing and photoelectric absorption. However, straightforwardprojection-domain decomposition methods for this charac-terization often yield poor reconstructions due to the highdynamic range of material properties encountered in an ac-tual luggage scan. Hence, for better reconstruction qualityunder a timing constraint, we propose a splitting-based, GPU-accelerated, statistical DECT reconstruction algorithm. Com-pared to prior art, our main contribution lies in the significantacceleration made possible by separating reconstruction anddecomposition within an ADMM framework. Experimentalresults, on both synthetic and real-world baggage phantoms,demonstrate a significant reduction in time required for con-vergence.

Index Terms— DECT, ADMM, GPU

1. INTRODUCTIONX-ray Computed Tomography (CT) is a widely deployedtechnique for threat detection in baggage at airport securitycheckpoints. However, using a single X-ray spectrum lim-its its reconstruction to only that of LAC (linear attenuationcoefficient) or HU (Hounsfield Unit) images which are, atbest, an approximation to the underlying energy-dependentcharacteristics of the materials. While LAC might suffice fordiscriminating between different types of tissues in medicalimaging, in the adversarial case of airport baggage scan-ning, materials used commonly for homemade explosives canpossess LAC values that are nearly the same as for benignmaterials. As a result, the technique of DECT, where pro-jections from two X-ray spectra are collected simultaneouslyat each angle, has been proposed for material discriminationbecause it allows for the material property to be recoveredin an additional dimension by using an energy-dependentattenuation model.

This research is funded by the BAA 17-03 AATR contract with Depart-ment of Homeland Security.

The most commonly used DECT model represents X-rayattenuation as the combined effect of Compton scattering andphotoelectric absorption, which can be written as:

µ(E, x) = xcfKN(E) + xpfp(E), (1)

where fKN(E) and fp(E) denote the energy-dependent mul-tipliers1 to the Compton coefficient xc and the photoelec-tric (PE) coefficient xp, respectively. Therefore, the goal ofDECT is to recover both the Compton and the PE coefficientsof the object using the projections mh,ml ∈ RM measuredat two different X-ray spectra. This amounts to solving simul-taneously for xc,xp ∈ RN from the following two equationsthat can be shown compactly by:

mh/l = − ln

∫Sh/l(E)e−Rµ(E)dE + ln

∫Sh/l(E)dE,

(2)where R denotes the forward projection matrix, and Sh(E)and Sl(E) denote, respectively, the photon distribution acrossenergy levels in the high- or low-energy spectra.

Solving the two equations shown above is made challeng-ing by the fact the two phenomena that contribute to X-rayattenuation — Compton and PE, occur at grossly differentscales. At most applicable energy levels, Compton scatteringis the dominant contributor to attenuation and this disparitybetween the two phenomena becomes worse with increasingenergy since fp(E) has a cubic decay with E. As a result,stable recovery of PE coefficients requires more sophisticatedinversion algorithms, such as those described in [2, 3, 4, 5].

Unfortunately, the algorithms cited above tend to be it-erative and, with run-of-the-mill computing hardware, take along time to return the results. Therefore, a straightforwardimplementation of these algorithms is not appropriate for theend-goals that motivate our research — high-throughput bag-gage screening at airports. The focus of the concepts pre-sented in this paper is on improving the computational effi-ciency of an ADMM-based statistical inversion algorithm.

2. RELATED WORKDECT involves two key steps: dual-energy decompositionand tomographic reconstruction. The two tasks can be done

1See [1] for a detailed formulation of fKN(E) and fp(E).

arX

iv:1

905.

0093

4v1

[ee

ss.I

V]

2 M

ay 2

019

Page 2: A SPLITTING-BASED ITERATIVE ALGORITHM FOR GPU ...projection-domain decomposition methods for this charac-terization often yield poor reconstructions due to the high dynamic range of

either sequentially, as in projection-wise decomposition, or ina unified step using iterative statistical approaches.Projection-wise Decomposition: One of the earliest ap-proaches for DECT decomposition, the Constrained Decom-position Method (CDM) [1] involves directly decomposingthe dual energy projections to Compton and PE line integralsfollowed by the FBP reconstructions of the two. In [6], CDMwas also extended to operate for Multi-Energy CT. A majordisadvantage of this method is that it guards itself poorlyagainst artifacts, especially in PE coefficients but it is still apreferred approach as it enables parallel implementation.Iterative Statistical Approaches: The statistical methods forDECT solve for the MAP estimates, finding the Compton andPE coefficients that best correspond to the measurements andany prior knowledge. The literature on Multi-Energy CT hasfocused largely on designing the models and the priors thatbest leverage the structural similarity across bases. In [4],Compton/PE images are reconstructed on a set of exponentialbasis functions with an edge-correlation penalty term. Thisidea of encouraging geometric similarity between the morestable Compton image and the PE image is further exploredin [5]. For this purpose, the authors have proposed a newNon-Local Mean (NLM) regularizer on the PE image and laidout an ADMM formation that scaled up to a problem sizeof practical interest. By treating the energy level as a thirddimension, the contributions in [2, 3] adopt a tensor-basedmodel for reconstruction using sparse priors.

Despite being able to produce high-quality DECT, statis-tical reconstructions generally fail drastically with regard tothe timing constraints of practical applications. Compared toLAC reconstructions, DECT has to deal with the added com-putational burden associated with the decomposition step. Asa result, in existing approaches, solving decomposition andreconstruction in one step becomes inefficient due to the com-bined complexity and high-dimensionality of the problem.

We therefore propose a statistical DECT approach thatcombines the best of the projection-wise decompositionmethods and the iterative statistical methods. More specifi-cally, we employ a splitting-based MAP estimation methodembedded in an ADMM framework. As we will show inSection 4, our new splitting scheme not only provides a betterconvergence rate, but also allows for powerful hardware-enabled acceleration.

3. PROPOSED METHOD3.1. Problem Formulation

To describe our proposed method, we first define the forwardmodel for X-ray attenuation, i.e., the nonlinear transform f(·)from Compton/PE coefficients to logarithmic projections:

fh/l(Rx) = Ch/l − ln

∫Sh/le

−fKNRxc−fpRxpdE (3)

where Ch/l are constants, x =[xTc ;xTp

]Tand R =[

R;R]. Given a pair of line-integral measurements m =

[mTh ;mT

l

]Tcorrupted by Poisson photon noise and possibly

other artifacts, we construct a MAP estimate with the TotalVariation (TV) term and the non-negativity prior:

xMAP = argminx

1

2

∥∥∥f(Rx)−m∥∥∥2

Σ+ λ|x|TV + g(x), (4)

where Σ is a diagonal matrix with photon counts as trace el-

ements as proposed in [7], g(xi) =

{0, xi ∈ R+

∞, xi /∈ R+ and λ is

the regularization parameter. While the unconstrained opti-mization in (4) is highly inefficient to solve directly, the Al-ternating Direction Method of Multipliers (ADMM) methodprovides a flexible splitting-based framework for both opti-mality and convergence [5].

3.2. Formulation for ADMM

ADMM begins with the conversion of (4) to its constrainedequivalent. By introducing two new auxiliary variables y andz for the TV and the non-negativity terms, respectively, andby posing x as the primal variable, the MAP minimizationcan be transformed into:

xMAP = argminx,y,z

1

2

∥∥∥f(Rx)−m∥∥∥2

Σ+ λ|y|+ g(z),

s.t.[yc/pzc/p

]=

[DI

]xc/p = Cxc/p, (5)

where D denotes the finite difference operator. Subsequently,the corresponding Augmented Lagrangian (AL) can be writ-ten as:L(x,y, z) =

1

2

∥∥∥f(Rx)−m∥∥∥2

Σ+ λ|y|+ g(z)

+∑

β∈{c,p}

ρβ2

∥∥∥∥[yβzβ]−Cxβ + uβ

∥∥∥∥2

,(6)

where u denotes the dual variable, or the scaled Lagrangianmultiplier, and ρ denotes the penalty parameter. ADMMsplits the minimization of AL into subproblems that aresolved separately in an iterative framework. One such in-tuitive splitting scheme proposed in [5] is for the iterativeupdates to be carried out according to (7) through (11) shownbelow. First, the primal variables are updated using:

xc = argminxc

1

2‖f(Rxc,Rxp)−m‖2Σ

+ρc2

∥∥∥∥[yczc]−Cxc + uc

∥∥∥∥2

,

(7)

xp = argminxp

1

2‖f(Rxc,Rxp)−m‖2Σ

+ρp2

∥∥∥∥[ypzp]−Cxp + up

∥∥∥∥2

.

(8)

Note that the update to xp is made subsequent to xc sincewe can expect xc to stabilize the recovery of the more noise-prone xp. Secondly, we update the auxiliary variables y and

Page 3: A SPLITTING-BASED ITERATIVE ALGORITHM FOR GPU ...projection-domain decomposition methods for this charac-terization often yield poor reconstructions due to the high dynamic range of

z corresponding to the TV and non-negativity terms, respec-tively, by using

yc/p = shrinkage(Dxc/p − uyc/p,λc/p

ρc/p), (9)

zc/p = max(0, xc/p − uzc/p), (10)

where the shrinkage function is defined in [8, Eq. 9.35]. Thenthe dual variable u is updated with

uc/p = uc/p +

[yc/pzc/p

]−Cxc/p. (11)

Lastly, we update ρ adaptively using a method described in [9]that is based on the primal and dual residual. While (9)-(11)can be realized by straightforward element-wise operations,solving (7)-(8) requires a nonlinear least squares algorithmsuch as Levenberg-Marquardt (LM). Despite its robustness,LM, is a sequential algorithm, offering little room for paral-lelization. Intuitively, the overall computational inefficiencycan be attributed to (7) or (8) which at once addresses twoproblems of very different nature: nonlinear dual-energy de-composition and regular linear tomographic reconstruction.

3.3. Proposed Splitting Scheme

For the new sped-up implementation presented in this paper,we further decompose (7)-(8) into two simpler subproblems:tomographic reconstruction followed by dual-energy decom-position. This is achieved by viewing measurement fitting asan additional constraint: a = Rx, where a is a new auxil-iary variable for that purpose. Incorporating this constraintwith the others, for the first subproblem we pose tomographicreconstruction as an unweighted unconstrained least squaresproblem:

xc/p = argminxc/p

ρc/p

2

∥∥∥∥∥∥∥ac/pyc/pzc/p

− [RC

]xc/p +

uac/p

uyc/puzc/p

∥∥∥∥∥∥∥

2

.

(12)and we can now separately deal with the decomposition, i.e.solving for a, in the following subproblem. We have cho-sen to use a CG solver for the minimization shown above be-cause our system is huge and sparse. Additionally, since theprojection-wise weights Σ are now absorbed by the decompo-sition step to be described later, the linear system correspond-ing to (12) is shift-invariant. Therefore, for improved conver-gence rate, each CG iteration lends itself well to precondition-ing using a high-passing filter [10]. In our implementation,we experimented with the ramp filter as the preconditioner.

The second subproblem, that of dual-energy decomposi-tion, can be solved by finding the pair of Compton/PE coeffi-cients that minimizes the cost function at each ray projectionindependently. Therefore, the decomposition is achieved byusing the Unconstrained Decomposition Method (UDM):[

acap

]= argmin

ac,ap

1

2

∥∥∥∥f(

[acap

])−m

∥∥∥∥2

Σ

+∑

β∈{c,p}

ρβ2

∥∥aβ −Rxβ + uaβ∥∥2

(13)

Table 1. Minimum number of operations per ADMM itera-tion. In LM(m)×CG(n), n CG iterations are taken to com-pute the update per LM iteration. Note that LM may requiremore operations than listed to tune the damping parameter.

R RT f(·)LM(m)×CG(n) 2m(n+ 1) 2m(n+ 1) 4mUDM(m)-PCG(n) 2n 2(n+ 1) 4m

Compared to (7) and (8), the update formulas in (12) and (13)have the following advantages: First, treating reconstructionand decomposition separately allows us to use two entirelydifferent algorithms, with each tailored for the correspondingsubproblem, which in our case are Preconditioned CG (PCG)and UDM. As an additional benefit, while LM is still usedin UDM, backtracking of the damping parameter, which isnecessary for dealing with nonlinear optimization, no longerinvolves expensive tomographic operations. Lastly, compu-tational efficiency is improved in terms of both the numberof operations required, as listed in Table 1, and paralleliza-tion. We can now not only independently execute (12) forboth bases, but also leverage massively parallelized hardwaresuch as GPUs to solve (13).

4. EXPERIMENTAL RESULTSWe now present qualitative and quantitative results on threephantoms: a simulated phantom (sim18), a real-world wa-ter bottle phantom (Water), and an actual baggage phantom(Clutter). We have implemented the following algorithmsand evaluated them on these three phantoms: (i) CDM-FBP:CDM [1] with 16 CPU threads; (ii) LM(m)×CG(n): ADMMas in [5]; and (iii) UDM(m)-(P)CG(n): proposed ADMMmethod with m UDM iterations and n (P)CG iterations.

Our Python implementations use the Astra toolbox [11]for GPU-based calculations of forward and backward projec-tions. Additionally, as a key factor for acceleration, we useGpufit [12] to parallelize the decomposition step. All experi-ments are carried out on a single computing cluster node with16 cores, 20GB RAM and an Nvidia 1080Ti GPU. Regard-ing initialization, in our experiment we have chosen the CDMoutput as the initial Compton estimate, since it is generallystable. For initial PE coefficients, we use a scaled version ofthe Compton estimate, similar to what is done in [5]. For bothCompton and PE, we used λ = 10−5 as the TV regularizationparameter and ρ = 10−3 as the initial penalty parameter. Fur-thermore, for scale-invariant evaluation, quantitative recon-struction quality is assessed with the normalized `2-distancebetween x and ground truth x∗:

ξ(x) = 20 log10

(‖x− x∗‖2‖x∗‖2

). (14)

The first phantom, sim18, contains seven circular re-gions of different materials. The reconstructed images, of

Page 4: A SPLITTING-BASED ITERATIVE ALGORITHM FOR GPU ...projection-domain decomposition methods for this charac-terization often yield poor reconstructions due to the high dynamic range of

Fig. 1. Compton and PE reconstructions by CDM-FBP and UDM1-PCG1. Note that we clipped the values in PE imagesof sim18 to show the streaking artifacts in the CDM-FBP reconstruction. Such artifacts, appearing also in the CDM-FBPreconstructions of Water and Clutter, are significantly suppressed by the UDM1-PCG1 method.

Fig. 2. ξ(x) vs. iteration for PE and average seconds periteration for (a) sim18 and (b) Water. The same plots forCompton are omitted since they exhibit similar trends.

size 512 × 512, are constructed from 720 angles, each con-taing 725 parallel ray projections. The first row in Figure 1shows the reconstructions by CDM-FBP and the proposedUDM-PCG algorithm. In Figure 2a, we display the aver-age computational time per iteration inside parentheses in theinset box and plot ξ(x) versus iteration to compare the con-vergence rates. In general, the ADMM algorithm with theproposed splitting scheme results in significantly shorter totalexecution time than LM×CG to reach the same error.

For results on the two real-world phantoms, the parallelbeam projection data for the two was collected on the Ima-tron C300 CT scanner. The X-ray source emits 1.8× 105 and

1.7 × 105 photons per ray with two energy spectra at 95keVand 130keV, respectively. The high- and the low-energy sino-grams are subsampled by two, resulting in 360 angles for eachand with 512 bins for each angle. The reconstruction resultsfor the Water phantom are shown in the second row of Fig-ure 1. For quantitative evaluation, in Figure 2b we plot ξ(x)versus iteration within the ROI – the central circular regionthat is occupied by distilled water. Our proposed UDM-PCGalgorithm is significantly more efficient not only in terms ofthe number of the iterations needed, but also in terms of thetime per iteration as compared to LM×CG.

The real-world baggage phantom Clutter is particu-larly challenging because it contains metallic objects, and il-lustrates well the need for a statistical reconstruction. Asshown in the third row of Figure 1, while the CDM-FBP PEreconstruction is completely overshadowed by streaking ar-tifacts, UDM1-PCG1 is able to recover object shapes to areasonable degree.

5. CONCLUSIONS AND FUTURE WORKIn this paper, we have proposed a new splitting scheme forimplementing ADMM to reconstruct Compton and PE co-efficient images using dual-energy projection data. By sep-arating the the reconstruction and decomposition steps, theproposed GPU-accelerated ADMM algorithm achieves a sig-nificant speedup in time when compared to the prior state-of-the-art. Future work will be aimed at generalization to 3Dreconstruction and improving initialization heuristics.

Page 5: A SPLITTING-BASED ITERATIVE ALGORITHM FOR GPU ...projection-domain decomposition methods for this charac-terization often yield poor reconstructions due to the high dynamic range of

6. REFERENCES

[1] Zhengrong Ying, Ram Naidu, and Carl R Crawford,“Dual energy computed tomography for explosive de-tection,” Journal of X-ray Science and Technology, vol.14, no. 4, pp. 235–256, 2006.

[2] Oguz Semerci, Ning Hao, Misha E Kilmer, and Eric LMiller, “Tensor-based formulation and nuclear normregularization for multienergy computed tomography,”IEEE Transactions on Image Processing, vol. 23, no. 4,pp. 1678–1693, 2014.

[3] Yanbo Zhang, Xuanqin Mou, Ge Wang, and HengyongYu, “Tensor-based dictionary learning for spectral ctreconstruction,” IEEE transactions on medical imaging,vol. 36, no. 1, pp. 142–154, 2017.

[4] Oguz Semerci and Eric L Miller, “A parametric level-setapproach to simultaneous object identification and back-ground reconstruction for dual-energy computed tomog-raphy,” IEEE transactions on image processing, vol. 21,no. 5, pp. 2719–2734, 2012.

[5] Brian H Tracey and Eric L Miller, “Stabilizing dual-energy x-ray computed tomography reconstructions us-ing patch-based regularization,” Inverse Problems, vol.31, no. 10, pp. 105004, 2015.

[6] Yaoshen Yuan, Brian Tracey, and Eric Miller, “Robustx-ray based material identification using multi-energysinogram decomposition,” in Anomaly Detection andImaging with X-Rays (ADIX). International Society forOptics and Photonics, 2016, vol. 9847, p. 98470V.

[7] Charles A Bouman and Ken Sauer, “A unified approachto statistical tomography using coordinate descent opti-mization,” IEEE Transactions on image processing, vol.5, no. 3, pp. 480–492, 1996.

[8] Charles A Bouman, “Model based image processing,”2013.

[9] Stephen Boyd, Neal Parikh, Eric Chu, Borja Peleato,Jonathan Eckstein, et al., “Distributed optimization andstatistical learning via the alternating direction methodof multipliers,” Foundations and Trends R© in Machinelearning, vol. 3, no. 1, pp. 1–122, 2011.

[10] Jeffrey A Fessler and Scott D Booth, “Conjugate-gradient preconditioning methods for shift-variant petimage reconstruction,” IEEE transactions on image pro-cessing, vol. 8, no. 5, pp. 688–699, 1999.

[11] Wim van Aarle, Willem Jan Palenstijn, Jeroen Cant,Eline Janssens, Folkert Bleichrodt, Andrei Dabravolski,Jan De Beenhouwer, K Joost Batenburg, and Jan Sijbers,

“Fast and flexible x-ray tomography using the astra tool-box,” Optics express, vol. 24, no. 22, pp. 25129–25147,2016.

[12] Adrian Przybylski, Bjorn Thiel, Jan Keller-Findeisen,Bernd Stock, and Mark Bates, “Gpufit: An open-sourcetoolkit for gpu-accelerated curve fitting,” Scientific re-ports, vol. 7, no. 1, pp. 15722, 2017.