11
70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, NO. 1, JANUARY 1, 2015 Ef cient Gaussian Sampling for Solving Large-Scale Inverse Problems Using MCMC Clément Gilavert, Saïd Moussaoui, Member, IEEE, and Jérôme Idier, Member, IEEE Abstract—The resolution of many large-scale inverse problems using MCMC methods requires a step of drawing samples from a high dimensional Gaussian distribution. While direct Gaussian sampling techniques, such as those based on Cholesky factor- ization, induce an excessive numerical complexity and memory requirement, sequential coordinate sampling methods present a low rate of convergence. Based on the reversible jump Markov chain framework, this paper proposes an efcient Gaussian sampling algorithm having a reduced computation cost and memory usage, while maintaining the theoretical convergence of the sampler. The main feature of the algorithm is to perform an approximate resolution of a linear system with a truncation level adjusted using a self-tuning adaptive scheme allowing to achieve the minimal computation cost per effective sample. The connection between this algorithm and some existing strategies is given and its performance is illustrated on a linear inverse problem of image resolution enhancement. Index Terms—Adaptive MCMC, conjugate gradient, Gibbs al- gorithm, multivariate Gaussian sampling, reversible jump Monte Carlo. I. INTRODUCTION A common inverse problem arising in many signal and image processing applications is to estimate a hidden object (e.g., an image or a signal) from a set of measurements given an observation model [1], [2]. The most frequent case is that of a linear model between and according to (1) with the known observation matrix and an ad- ditive noise term representing measurement errors and model uncertainties. Such a linear model covers many real problems such as, for instance, denoising [3], deblurring [4], and recon- struction from projections [5], [6]. The statistical estimation of in a Bayesian simulation frame- work [7], [8] rstly requires the formulation of the posterior distribution , with a set of unknown hyper-param- eters. Pseudo-random samples of are then drawn from this Manuscript received October 18, 2013; revised May 27, 2014; accepted October 06, 2014. Date of publication November 04, 2014; date of current version November 27, 2014. The associate editor coordinating the review of this manuscript and approving it for publication was Prof. Francesco Verde. This work was supported by the CNRS and the French Région des Pays de la Loire (France). The authors are with IRCCyN, CNRS UMR 6597, Ecole Centrale Nantes, 44321, Nantes Cedex 3, France (e-mail: [email protected]). Digital Object Identier 10.1109/TSP.2014.2367457 posterior distribution. Finally, a Bayesian estimator (posterior mean, maximum a posteriori) is computed from the set of gen- erated samples. Other quantities of interest, such as posterior variances, can be estimated likewise. Within the standard Monte Carlo framework, independent realizations of the posterior law must be generated, which is rarely possible in realistic cases of inverse problems. One rather resorts to Markov Chain Monte Carlo (MCMC) schemes, where Markovian dependencies be- tween successive samples are allowed. A very usual sampling scheme is then to iteratively draw realizations from the condi- tional posterior densities and , according to a Gibbs sampler [11]. In such a context, when independent Gaussian models and are assigned to the noise statistics and to the unknown object distribution, respectively, the set of hyper-parameters determines the mean and the covariance of the latter two distributions. This framework also covers the case of priors based on hierarchical or latent Gaussian models such as Gaussian scale mixtures [9], [10] and Gaussian Markov elds [11], [12]. The additional parameters of such models are then included in . According to such modeling, the conditional posterior distribution is also Gaussian, , with a precision matrix (i.e., the inverse of the covariance matrix) given by: (2) and a mean vector such that: (3) Let us remark that the precision matrix generally depends on the hyper-parameter set through and , so that is a varying matrix along the Gibbs sampler iterations. Moreover, the mean vector is expressed as the solution of a linear system where is the normal matrix. In order to draw samples from the conditional posterior distri- bution , a usual way is to rstly perform the Cholesky factorization of the covariance matrix [13], [14]. Since (2) yields the precision matrix rather than the covariance matrix, Rue [15] proposed to compute the Cholesky decomposition of , i.e., , and to solve the triangular system , where is a vector of independent Gaussian variables of zero mean and unit variance. Moreover, the Cholesky factorization is exploited to calculate the mean from (3) by solving two tri- angular systems sequentially. The Cholesky factorization of generally requires operations. Spending such a numerical cost at each iteration of the sampling scheme rapidly becomes prohibitive for large 1053-587X © 2014 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information. 2015 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.

70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

Embed Size (px)

Citation preview

Page 1: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, NO. 1, JANUARY 1, 2015

Efficient Gaussian Sampling for Solving Large-ScaleInverse Problems Using MCMC

Clément Gilavert, Saïd Moussaoui, Member, IEEE, and Jérôme Idier, Member, IEEE

Abstract—The resolution of many large-scale inverse problemsusing MCMC methods requires a step of drawing samples froma high dimensional Gaussian distribution. While direct Gaussiansampling techniques, such as those based on Cholesky factor-ization, induce an excessive numerical complexity and memoryrequirement, sequential coordinate sampling methods present alow rate of convergence. Based on the reversible jump Markovchain framework, this paper proposes an efficient Gaussiansampling algorithm having a reduced computation cost andmemory usage, while maintaining the theoretical convergenceof the sampler. The main feature of the algorithm is to performan approximate resolution of a linear system with a truncationlevel adjusted using a self-tuning adaptive scheme allowing toachieve the minimal computation cost per effective sample. Theconnection between this algorithm and some existing strategiesis given and its performance is illustrated on a linear inverseproblem of image resolution enhancement.

Index Terms—Adaptive MCMC, conjugate gradient, Gibbs al-gorithm, multivariate Gaussian sampling, reversible jump MonteCarlo.

I. INTRODUCTION

A common inverse problem arising in many signal andimage processing applications is to estimate a hidden

object (e.g., an image or a signal) from a set ofmeasurements given an observation model [1], [2].The most frequent case is that of a linear model between andaccording to

(1)

with the known observation matrix and an ad-ditive noise term representing measurement errors and modeluncertainties. Such a linear model covers many real problemssuch as, for instance, denoising [3], deblurring [4], and recon-struction from projections [5], [6].The statistical estimation of in a Bayesian simulation frame-

work [7], [8] firstly requires the formulation of the posteriordistribution , with a set of unknown hyper-param-eters. Pseudo-random samples of are then drawn from this

Manuscript received October 18, 2013; revised May 27, 2014; acceptedOctober 06, 2014. Date of publication November 04, 2014; date of currentversion November 27, 2014. The associate editor coordinating the reviewof this manuscript and approving it for publication was Prof. FrancescoVerde. This work was supported by the CNRS and the French Région desPays de la Loire (France).The authors are with IRCCyN, CNRS UMR 6597, Ecole Centrale Nantes,

44321, Nantes Cedex 3, France (e-mail: [email protected]).Digital Object Identifier 10.1109/TSP.2014.2367457

posterior distribution. Finally, a Bayesian estimator (posteriormean, maximum a posteriori) is computed from the set of gen-erated samples. Other quantities of interest, such as posteriorvariances, can be estimated likewise.Within the standardMonteCarlo framework, independent realizations of the posterior lawmust be generated, which is rarely possible in realistic cases ofinverse problems. One rather resorts to Markov Chain MonteCarlo (MCMC) schemes, where Markovian dependencies be-tween successive samples are allowed. A very usual samplingscheme is then to iteratively draw realizations from the condi-tional posterior densities and , accordingto a Gibbs sampler [11].In such a context, when independent Gaussian models

and are assigned to the noise statisticsand to the unknown object distribution, respectively, the set ofhyper-parameters determines the mean and the covarianceof the latter two distributions. This framework also coversthe case of priors based on hierarchical or latent Gaussianmodels such as Gaussian scale mixtures [9], [10] and GaussianMarkov fields [11], [12]. The additional parameters of suchmodels are then included in . According to such modeling, theconditional posterior distribution is also Gaussian,

, with a precision matrix (i.e., the inverse of thecovariance matrix) given by:

(2)

and a mean vector such that:

(3)

Let us remark that the precision matrix generally dependson the hyper-parameter set through and , so that isa varying matrix along the Gibbs sampler iterations. Moreover,the mean vector is expressed as the solution of a linear systemwhere is the normal matrix.In order to draw samples from the conditional posterior distri-

bution , a usual way is to firstly perform the Choleskyfactorization of the covariancematrix [13], [14]. Since (2) yieldsthe precision matrix rather than the covariance matrix, Rue[15] proposed to compute the Cholesky decomposition of ,i.e., , and to solve the triangular system ,where is a vector of independent Gaussian variables of zeromean and unit variance. Moreover, the Cholesky factorizationis exploited to calculate the mean from (3) by solving two tri-angular systems sequentially.The Cholesky factorization of generally requires

operations. Spending such a numerical cost at each iterationof the sampling scheme rapidly becomes prohibitive for large

1053-587X © 2014 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

2015 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising orpromotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrightedcomponent of this work in other works must be obtained from the IEEE.

1

Page 2: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

GILAVERT et al.: LARGE-SCALE INVERSE PROBLEMS USING MCMC 71

values of . In specific cases where belongs to certain fam-ilies of structured matrices, the factorization can be obtainedwith a reduced numerical complexity, e.g., when isToeplitz [16] or even when is circulant [17].Sparse matrices can be also factorized at a reduced cost [15],[18]. Alternative approaches to the Cholesky factorization arebased on using an iterative method for the calculation of the in-verse square root matrix of using Krylov subspace methods[19]–[21]. In practice, even in such favorable cases, the factor-ization often remains a burdensome operation to be performedat each iteration of the Gibbs sampler.The numerical bottleneck represented by the factorization

techniques can be removed by using alternate schemes thatbypass the step of exactly sampling . For instance,a simple alternative solution is to sequentially sample eachentry of given the other variables according to a scalar Gibbsscheme. However, such a scalar approach reveals extremelyinefficient when is strongly correlated, since eachconditional sampling step will produce a move of very smallvariance. As a consequence, a huge number of iterations will berequired to reach convergence [22]. A better trade-off betweenthe numerical cost of each iteration and the overall convergencespeed of the sampler must be found.In this paper, we focus on a two-step approach named In-

dependent Factor Perturbation in [12] and Perturbation-Opti-mization in [23] (see also [18], [24]). It consists in• drawing a sample from ,• solving the linear system .

It can be easily checked that, when the linear system issolved exactly, the new sample is distributed according to

. Hereafter, we refer to this method as Exact Per-turbation Optimization (E-PO). However, the numerical costof E-PO is typically as high as the Cholesky factorization of .Therefore, an essential element of the Perturbation Optimiza-tion approach is to truncate the linear system solving [12], [23],[24], by running a limited number of iterations of the conjugategradient method (CG). For the sake of clarity, let us call theresulting version Truncated Perturbation Optimization (T-PO).Skipping from E-PO to T-PO allows to strongly reduce the

numerical cost of each iteration. However, let us stress that noconvergence analysis of T-PO exists, to our best knowledge. Itis only argued that a well-chosen truncation level will induce asignificant reduction of the numerical cost and a small estima-tion error. The way the latter error alters the convergence to-wards the target distribution remains a fully open issue, that hasnot been discussed in existing contributions. Moreover, how theresolution accuracy should be chosen in practice is also an openquestion.A first contribution of the present paper is to bring practical

evidence that the T-PO does not necessarily converge towardsthe target distribution (see Section IV). In practice, the implicittrade-off within T-PO is between the computational cost andthe error induced on the target distribution, depending on theadopted truncation level. Our second contribution is to proposea new scheme similar to T-PO, but with a guarantee of conver-gence to the target distribution, whatever the truncation level.We call the resulting scheme Reversible Jump Perturbation Op-timization (RJPO), since it incorporates an accept-reject step de-

rived within the Reversible Jump MCMC (RJ-MCMC) frame-work [25], [26]. The numerical cost of the proposed test is mar-ginal, so that RJPO has nearly the same cost per iteration asT-PO. Finally, we propose an unsupervised tuning of the trun-cation level allowing to automatically achieve a pre-specifiedoverall acceptance rate or even to minimize the computationcost per effective sample. Consequently, the resulting algorithmcan be viewed as an adaptive MCMC sampler [27]–[29].The rest of the paper is organized as follows: Section II intro-

duces the global framework of RJ-MCMC and presents a gen-eral scheme to sample Gaussian vectors. Section III considers aspecific application of the previous results, which finally boilsdown to the proposed RJPO sampler. Section IV analyses theperformance of RJPO compared to T-PO on simple toy prob-lems and presents the adaptive RJPO which incorporates an au-tomatic tuning of the truncation level. Finally, in Section V, anexample of linear inverse problem, the unsupervised image res-olution enhancement is presented to illustrate the applicabilityof the method. These results confirm the superiority of the RJPOalgorithm over the usual sampling approach, based on Choleskyfactorization, in terms of computational cost andmemory usage.

II. THE REVERSIBLE JUMP MCMC FRAMEWORK

The sampling procedure consists on iteratively constructinga Markov chain whose distribution asymptotically converges tothe target distribution . Let be the current sampleof the Markov chain and the new sample obtained accordingto a transition kernel derived in the reversible jump framework[25], [26].

A. General Framework

In the constant dimension case, the Reversible Jump MCMCstrategy introduces an auxiliary variable , obtained froma distribution and a deterministic move according to adifferentiable transformation

This transformation must be reversible, that is .The new sample is thereby obtained by submitting (resultingfrom the deterministic move) to an accept-reject step with anacceptance probability given by

with the Jacobian determinant of the transformationat . In fact, the choice of the conditional distributionand the transformation must be adapted to the target distri-bution and affects the resulting Markov chain propertiesin terms of correlation and convergence rate.

B. Gaussian Case

To draw samples from a Gaussian distribution ,we generalize the scheme adopted in [30]. We consider ,and take an auxiliary variable sampled from

(4)

Page 3: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

72 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, NO. 1, JANUARY 1, 2015

where , and denote a real matrix, a realpositive definite matrix and a real vector, respectively.The choice of the latter three quantities will be discussed later.The proposed deterministic move is performed using the

transformation such that

(5)

with functions , and.

Proposition 1: Let an auxiliary variable be obtained ac-cording to (4) and a proposed sample resulting from (5). Thenthe acceptance probability is

(6)

with

(7)

In particular, the acceptance probability equals one whenis defined as the exact solution of the linear system

(8)

Proof: See Appendix A.Let us remark that is a dummy parameter, since the residual(and thus ) depends on through only.

However, choosing a specific expression of jointly with andwill lead to a simplified expression of in the next section.Proposition 1 plays a central role in our proposal. When the

exact resolution of (8) is numerically costly, it allows to derivea procedure where the resolution is performed only approxi-mately, at the expense of a lowered acceptance probability. Theconjugate gradient algorithm stopped before convergence, is atypical example of an efficient algorithm to approximately solve(8).Proposition 2: Let an auxiliary variable be obtained ac-

cording to (4), a proposed sample resulting from (5) andbe the exact solution of (8). The correlation between two succes-sive samples is zero if and only if matrices and are chosensuch that

(9)

Proof: See Appendix B.Many couples fulfill condition (9):• Consider the Cholesky factorization and take

, . It leads to with. According to (8), the next sample

, will be obtained as

Such an update scheme is exactly the same as the one pro-posed by Rue in [15].

• The particular configuration

(10)

is retained in the sequel, since:i) is a condition of Proposition 2,ii) simplifies (8) to a linear system .

In particular, it allows to make a clear connection betweenour RJ-MCMC approach and the E-PO algorithm in thecase of an exact resolution of the linear system. It also al-lows to simplify the accept-reject step that must be consid-ered when an approximate resolution is retained.

III. RJ-MCMC ALGORITHMS FOR SAMPLINGGAUSSIAN DISTRIBUTIONS

A. Sampling the Auxiliary Variable

According to the configuration (10), the auxiliary variableis distributed according to . It can then be

expressed as , being distributed according to. Consequently, the auxiliary variable sampling step

is reduced to the simulation of , which is the perturbation stepin the PO algorithm. In [12], [23], a subtle way of sampling isproposed. It consists in exploiting (3) and perturbing each factorseparately:1) Sample ,2) Sample ,3) Set , a sample of .It is important to notice that such a tricky method is interestingsince matrices and have often a simple structure if notdiagonal.We emphasize that this perturbation step can be applied

more generally for the sampling of any Gaussian distribution,for which a factored expression of the precision matrix isavailable under the form , with matrix .In such a case, , where .

B. Exact Resolution Case: The E-PO Algorithm

As stated by proposition 1, the exact resolution of system (8)implies an acceptance probability equal to one. The resultingsampling procedure is thus based on the following steps:1) Sample ,2) Set ,3) Take .Let us remark that , so the

handling of variable can be skipped and Steps 2 and 3 can bemerged to an equivalent but more direct step:

2) Set .In the exact resolution case, the obtained algorithm is thus iden-tical to the E-PO algorithm [23].According to Proposition 2, E-PO enjoys the property that

each sample is totally independent from the previous ones.However, a drawback is that the exact resolution of the linearsystem often leads to an excessive numerical com-plexity and memory usage in high dimensions [12]. In practice,early stopping of an iterative solver such as the linear conjugategradient algorithm is used, yielding the Truncated PerturbationOptimization (T-PO) version. The main point is that, up to our

Page 4: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

GILAVERT et al.: LARGE-SCALE INVERSE PROBLEMS USING MCMC 73

knowledge, there is no theoretical analysis of the efficiencyof T-PO and of its convergence to the target distribution.Actually, the simulation tests provided in Section IV indicatethat convergence to the target distribution is not guaranteed.However, as shown in the next subsection, two slight butdecisive modifications of T-PO lead us to the RJPO version,which is a provably convergent algorithm.

C. Approximate Resolution Case: The RJPO Algorithm

In the case of configuration (10), (7) reduces to

(11)

Therefore, a first version of the RJPO algorithm is as follows:1) Sample .2) Set . Solve the linear system , inan approximate way. Let denote the obtained solution,

and propose .3) With probability , set , other-wise let .

An important point concerns the initialization of the linearsolver in Step 2: in the case of an early stopping, the computedapproximate solution may depend on the initial point . Onthe other hand, must not depend on , otherwise the re-versibility of the deterministic move (5) would not be ensured.Hence, the initial point must not depend on either. In therest of the paper, is the default choice.A more compact and direct version of the sampler can be ob-

tained by substituting in (11). The latter reducesto the solving of the system . Step 2 of the RJPO algo-rithm is then simplified to:

2) Solve the linear system in an approximate way.Let denote the obtained solution and .

For the reason just discussed above, the initial point of thelinear solver must be such that does not dependon . Hence, as counterintuitive as it may be, choices such as

or are not allowed, while is the defaultchoice corresponding to .It is remarkable that both T-PO and the proposed RJPO al-

gorithm rely on the approximate resolution of the same linearsystem . However, RJPO algorithm incorporates twoadditional ingredients that make the difference in terms of math-ematical validity:• RJPO relies on an accept-reject strategy to ensure the sam-pler convergence in the case of an approximate systemsolving.

• There is a constraint on the initial point of the linearsolver: must not depend on .

D. Implementation Issues

Let us stress that there is no constraint on the choice of thelinear solver, nor on its initialization and the early stopping rule,except that they must not depend on the value of . Indeed,any linear system solver, or any quadratic programming methodcould be employed. In the sequel, we have adopted the linearconjugate gradient algorithm for two reasons:• Early stopping (i.e., truncating) the conjugate gradient it-erations is a very usual procedure to approximately solve a

Fig. 1. Acceptance rate of the RJPO algorithm for different values of therelative residual norm in the case of a small size problem .

linear system, with well-known convergence properties to-wards the exact solution [31]. Moreover, a preconditionedconjugate gradient could well be used to accelerate the con-vergence speed.

• It lends itself to a matrix-free implementation with reducedmemory requirements, as far as matrix-vector products in-volving matrix can be performed without explictly ma-nipulating such a matrix.

On the other hand, we have selected a usual stopping rulebased on a threshold on the relative residual norm:

(12)

IV. PRACTICAL PERFORMANCE ANALYSIS AND OPTIMIZATION

The aim of this section is to analyze the performance of theRJPO algorithm and to discuss the influence of the relativeresidual norm (and hence, the truncation level of the conjugategradient algorithm) on the practical efficiency of both RJPOand T-PO schemes.

A. Performance Analysis

A Gaussian distribution with randomly generated positivedefinite precision matrix and mean vector is considered.First, we focus on a small size problem to discuss theinfluence of the truncation level on the numerical performancein terms of acceptance rate and estimation error. For the retainedGaussian sampling schemes, both RJPO and T-PO are run fora number of CG iterations allowing to reach a predefined valueof relative residual norm (12). We also discuss the influence ofthe problem dimension on the best value of the truncation levelallowing to minimize the total number of CG iterations beforeconvergence.1) Acceptance Rate: Fig. 1 shows the average acceptance

probability (acceptance rate) obtained over itera-tions of the sampler for different relative residual norm values.It can be noted that the acceptance rate is almost zero when

the relative residual norm is larger than and monotonicallyincreases for higher resolution accuracies. Moreover, a relativeresidual norm lower than leads to an acceptance proba-bility almost equal to one. Such a curve indicates that the stop-ping criterion of the CGmust be chosen carefully in order to run

Page 5: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

74 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, NO. 1, JANUARY 1, 2015

the RJPO algorithm efficiently and to get non-zero acceptanceprobabilities. Finally, we emphasize that this curve mainly de-pends on the condition number of the precision matrix . Evenif the shape of the acceptance curve stays the same for differentproblems, it happens to be difficult to determine the value of therelative residual norm that corresponds to a given acceptancerate.2) Estimation Error: The estimation error is assessed as

the relative mean square error (RMSE) on the estimated meanvector and covariance matrix

(13)

where and represent the Frobenius and the norms,respectively. , , , and are respectively the mean and thecovariance matrix of the Gaussian vector, and their empiricalestimates using the generated Markov chain samples:

and

with iterations of burn-in and total iterations.As expected, Fig. 2 indicates that the estimation error is very

high if the acceptance rate is zero (when the relative residualnorm is lower than ), even for RJPO after it-erations. This is due to the very low acceptance rate which slowsdown the chain convergence. However, as soon as new samplesare accepted, RJPO leads to the same performance as when thesystem is solved exactly (E-PO algorithm). On the other hand,T-PO keeps a significant error for small and moderate resolu-tion accuracies. Naturally, both methods present similar perfor-mance when the relative residual norm is sufficiently low sincethese methods tend to provide almost the same samples withan acceptance probability equal to one. This experimental re-sult clearly highlights the deficiency of T-PO: the system mustbe solved with a relatively high accuracy to avoid an importantestimation error. On the other hand, in the RJPO algorithm theacceptance rate is a good indicator whether the value of the rela-tive residual norm threshold is appropriate to ensure a sufficientmixing of the chain.3) Computation Cost: Since the CG iterations correspond to

the only burdensome task, the numerical complexity of the sam-pler can expressed in terms of the total number of CG iter-ations to be performed before convergence and the number ofrequired samples to get efficient empirical approximation of theestimators. The Markov chain convergence is firstly assessedusing the Gelman-Rubin criterion, which requires the runningof several chains [32]. It consists in computing a scale reduc-tion factor based on the between and within-chain variances. Inthis experiment, 100 parallel chains are considered. The resultsare summarized in Fig. 3. It can be noted that a lower acceptance

Fig. 2. Estimation error for different values of the truncation level afteriterations of E-PO, T-PO and RJPO algorithms: (a) mean vector,

(b) covariance matrix. (a) Mean vector; (b) covariance matrix.

Fig. 3. Number of CG iterations before convergence and acceptance proba-bility of the RJPO algorithm for different values of relative residual norm for asmall size problem .

rate induces a higher number of iterations since the Markovchain converges more slowly towards its stationary distribution.One can also see that a minimal cost can be reached and, ac-

cording to Fig. 1, it corresponds to an acceptance rate of almostone. As the acceptance rate decreases, even a little, the computa-tional cost rises very quickly. Conversely, if the relative residualis too small, the computation effort per sample will decreasebut additional sampling iterations will be needed before conver-gence, which naturally increases the overall computation cost.The latter result points out the need to appropriately choose the

Page 6: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

GILAVERT et al.: LARGE-SCALE INVERSE PROBLEMS USING MCMC 75

truncation level to jointly avoid a low acceptance probabilityand a high resolution accuracy of the linear system since bothinduce unnecessary additional computations.4) Statistical Efficiency: The performance of the RJPO sam-

pler can also be analyzed using the effective sample size (ESS)[33, p. 125]. This indicator gives the number of independentsamples, , that would yield the same statistical efficiency inapproximating the Bayesian estimator as successive sam-ples of the simulated chain [34]. It is related to the chain auto-correlation function according to

(14)

where the autocorrelation coefficient at lag . In the Gaussiansampling context, such a relation allows to define how manyiterations are needed for each resolution accuracy in orderto get estimators with comparable statistical performance (e.g.,using chains having the same effective sample size). Under thehypothesis of a first-order autoregressive chain, , (14)leads to the ESS ratio

(15)

It can be noted that the ESSR is equal to one when the samplesare independent and decreases as the correlation be-tween successive samples grows. In the RJPO case, we proposeto define the computing cost per effective sample (CCES) as

(16)

where is the average number of CG iterationsper sample. Fig. 4 shows the ESSR and the CCES in the caseof a Gaussian vector of dimension . It can be seen thatan early stopped CG algorithm induces a very small ESSR, dueto a large sample correlation value, and thus a high effectivecost to produce accurate estimates. On the contrary, a very pre-cise resolution of the linear system induces a larger number ofCG iterations per sample but a shorter Markov chain since theESSR is almost equal to 1. The best trade-off is produced by in-termediate values of the relative residual norm threshold around

.To conclude, the Gelman-Rubin convergence diagnostic and

the ESS approach both confirm that the computation cost of theRJPO can be reduced by appropriately truncating the CG iter-ations. Although the Gelman-Rubin convergence test is prob-ably more accurate, since it is based on several independentsequences, the CCES based test is far simpler and providesnearly the same trade-off in the tested example. Such encour-aging results have motivated us to develop a self-tuning strategyto automatically adjust the threshold parameter by tracking theminimizer of the CCES. Our proposed strategy is presented inSection IV-B.5) Influence of the Dimension: Fig. 5 summarizes the op-

timal values of the truncation level that allows to minimize theCCES for different values of . The best trade-off is reached

Fig. 4. Computing cost per effective sample of the RJPO algorithm for differentrelative residual norm values on a small size problem estimated from

samples.

Fig. 5. Influence of the problem dimension on the optimal values of the relativeresidual norm and the acceptance rate.

for decreasing values of as grows. More generally, the sameobservation can be made as the problem conditioning deterio-rates. In practice, predicting the appropriate truncation level fora given problem is difficult. Fortunately, Fig. 5 also indicatesthat the optimal setting is obtained for an acceptance probabilitythat remains almost constant. The best trade-off is clearly ob-tained for an acceptance rate lower than one ( corre-sponds to , i.e., to the exact solving of ). In thetested example, the optimal truncation level rather correspondsto an acceptance rate around 0.99. However, finding an explicitmathematical correspondence between and is not a simpletask. In the next subsection, we propose an unsupervised tuningstrategy of the relative residual norm allowing either to achievea predefined target acceptance rate, or even to directly optimizethe computing cost per effective sample.

B. Adaptive Tuning of the Resolution Accuracy

The suited value of the relative residual norm to achievea desired acceptance rate can be adjusted recursively usinga Robbins-Monro type algorithm [35], [36]. Such an adaptive

Page 7: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

76 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, NO. 1, JANUARY 1, 2015

scheme is formulated in the stochastic approximation frame-work [37] in order to solve a non-linear equation of the form

using an update

(17)

where is a random variable traducing the uncertainty on eachevaluation of function and is a sequence of step-sizesensuring stability and convergence [38]. Such a procedure hasbeen widely used for the optimal scaling of adaptive MCMC al-gorithms [27], [28] since it does not alter the chain convergencetowards the target distribution. For instance, it is used in [39],[40] to set adaptively the scale parameters of a Random walkMetropolis-Hastings (RWMH) algorithm in order to reach theoptimal acceptance rate suggested by theoretical or empiricalanalysis [41], [42]. The same procedure was also used by [43]for the adaptive tuning of a Metropolis-adjusted Langevin algo-rithm (MALA) to reach the optimal acceptance rate proposedby [44].1) Tuning to Achieve a Target Acceptance Rate: In order to

ensure the positivity of the relative residual norm , the updateis performed on its logarithm. At each iteration of the sampler,the relative residual norm is adjusted according to

(18)

where is a given target acceptance probability and isa sequence of step-sizes decaying to 0 as grows in order toensure the convergence of the Markov chain to the target distri-bution. As suggested in [28], the step-sizes are chosen accordingto , with . We emphasize that more so-phisticated methods, such as those proposed in [36] could beused to approximate the acceptance rate curve and to derive amore efficient adaptive strategy for choosing this parameter.The adaptive RJPO is applied to the sampling of the previ-

ously described Gaussian distribution using the adopted step-size with parameters and . Fig. 6 presents theevolution of the average acceptance probability and the obtainedrelative residual norm for three different values of the target ac-ceptance rate . One can note that the average acceptance rateconverges to the desired value. Moreover, the relative residualnorm also converges to the expected value according to Fig. 1(for example, the necessary relative residual norm to get an ac-ceptance probability is equal to ).In practice, it remains difficult to a priori determine which

acceptance rate should be targeted to achieve the faster conver-gence. The next subsection proposes to modify the target of theadaptive strategy to directly minimize the CCES.2) Tuning to Optimize the Numerical Efficiency: A given

threshold on the relative residual norm induces an averagetruncation level and an ESSR value, fromwhich the CCES canbe deduced according to (16). Our goal is to adaptively adjustthe threshold value in order to minimize the CCES. Let bethe average number of CG iterations per sample correspondingto the optimal threshold value. In the plane , it is easyto see that is the abscissa of the point at which the tangentof the ESSR curve intercepts the origin (see Fig. 7).The ESSR is expressed by (15) as a function of the chain cor-

relation , the latter being an implicit function of the acceptance

Fig. 6. Behavior of the adaptive RJPO for 1000 iterations and three values ofthe target acceptance probability: (a) Evolution of the average acceptance prob-ability and (b) Evolution of the computed relative residual norm. (a) Acceptanceprobability; (b) relative residual norm.

Fig. 7. Influence of the CG truncation level on the overall computation costand the statistical efficiency of the RJPO for sampling a Gaussian of dimension

.

rate . For , according to Proposition 2. For ,since no new sample can be accepted. For intermediate

values of , the correlation lies between 0 and 1, and it is typi-cally decreasing. It can be decomposed on two terms:• With a probability , the accept-reject procedure pro-duces identical (i.e., maximally correlated) samples in caseof rejection.

Page 8: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

GILAVERT et al.: LARGE-SCALE INVERSE PROBLEMS USING MCMC 77

Fig. 8. Evolution of the relative residual norm and the acceptance rate for aGaussian sampling problem of size . The adaptive algorithm leads toa relative residual norm leading to .

• In case of acceptance, the new sample is slightly correlatedwith the previous one, because of the early stopping of theCG algorithm.

While it is easy to express the correlation induced by rejection,it is difficult to find an explicit expression for the correlationbetween accepted samples. However, we have checked that thelatter source of correlation is negligible compared to the cor-relation induced by rejection. If we approximately assume thataccepted samples are independent, we get . Conse-quently, the best tuning of the relative residual norm leading tothe lowest computation cost per effective sample is the mini-mizer of . By necessary condition, we get

Finally, a similar procedure as in the previous section is appliedto adaptively adjust the optimal value of the relative residualnorm according to

(19)

where is evaluated numerically. Fig. 8 illustrates that theproposed adaptive scheme efficiently adjusts to minimize theCCES where is around 26, which is in agreement withFig. 7.

V. APPLICATION TO UNSUPERVISED SUPER-RESOLUTION

In the linear inverse problem of unsupervised image super-resolution, several images are observed with a low spatial res-olution. In addition, the measurement process presents a pointspread function (PSF) that introduces a blur on the images. Thepurpose is then to reconstruct the original image with a higherresolution using an unsupervised method. Such an approach al-lows to also estimate the model hyper-parameters and the PSF[23], [45], [46]. In order to discuss the relevance of the pre-viously presented Gaussian sampling algorithms we apply aBayesian approach andMCMCmethods for solving this inverseproblem.

A. Problem Statement

The observation model is given by , where, with the vector containing the pixels of the

observed images in a lexicographic order, the soughthigh resolution image, the circulant convolution ma-trix associated with the blur, the decimation matrixand the additive noise. This linear model also includes clas-sical image deconvolution problems [1], [2]. The noise is as-sumed to follow a zero-mean Gaussian distribution with an un-known precision matrix . We also assume a zero-meanGaussian distribution for the prior of the sought variable , witha precisionmatrix , where is the circulant convo-lution matrix associated to a Laplacian filter. Non-informativeJeffrey’s priors [47] are also assigned to the two hyper-parame-ters and .According to Bayes’ theorem, the posterior distribution is

given by

To explore this posterior distribution, a Gibbs sampler itera-tively draws1) from given as

2) from given as

3) from which is

with

The third step of the sampler requires an efficient sampling ofa multivariate Gaussian distribution whose parameters changealong the sampling iterations. In the sequel, direct sampling withCholesky factorization [15] is firstly employed as a referencemethod. It yields the same results as the E-PO algorithm. Forthe inexact resolution case, the T-PO algorithm using a CG con-trolled by the relative residual norm, and the adaptive RJPO di-rectly tuned with the acceptance probability are performed. Forthese two methods, the product matrix-vector used in the CG al-gorithm is done by exploiting the structure of the precision ma-trix and thus only implies circulant convolutions, performedby FFT, and decimations.

B. MCMC Results

We consider the observation of five images of dimension 128128 pixels and we reconstruct the original

one of dimension 256 256 . The convolutionpart has a Laplace shape with of full width at half maximum(FWHM) of 4 pixels. A white Gaussian noise is added to get a

Page 9: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

78 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, NO. 1, JANUARY 1, 2015

Fig. 9. Unsupervised super-resolution—image reconstruction using the adap-tive RJPO algorithm with .

TABLE ICOMPARISON BETWEEN DIFFERENT GAUSSIAN SAMPLING STRATEGIES: THECHOLESKY FACTORIZATION BASED APPROACH, THE T-PO CONTROLLEDBY THE RELATIVE RESIDUAL NORM AND THE ADAPTIVE RJPO TUNED BYTHE ACCEPTANCE RATE. THE PERFORMANCES ARE EXPRESSED IN TERMSOF EMPIRICAL MEAN AND STANDARD DEVIATION OF HYPER-PARAMETERS

AND ONE RANDOMLY CHOSEN PIXEL

signal-to-noise ratio (SNR) equal to 20 dB. The original imageand one of the observations are shown in Fig. 9.The Gibbs sampler is run for 1000 iterations and a burn-in pe-

riod of 100 iterations is considered after a visual inspection ofthe chains. The performances are evaluated in terms of the meanand standard deviation of both hyper-parameters , and onerandomly chosen pixel of the reconstructed image. Table Ipresents the mean and standard deviation of the variable of in-terest. Aswe can see, the T-PO algorithm is totally inappropriateeven with a precision of . Conversely, the estimation fromthe samples given by the adaptive RJPO and Cholesky methodare very similar, which demonstrates the correct behavior of theproposed algorithm.Fig. 10 shows the evolution of the acceptance rate with re-

spect to the number of CG iterations. We can notice that at least400 iterations are required to have a nonzero acceptance prob-ability. Moreover, more than 800 iterations seems unnecessary.For this specific problem, the E-PO algorithm needs theoreti-cally iterations to have a new sample while theadaptive RJPO only requires around 700. Concerning the com-putation time, on a Intel Core i7-3770 with 8 GB of RAM and a64 bit system, it took about 20.3 s on average and about 6 GB ofRAM for the Cholesky sampler to generate one sample and only15.1 s and less than 200 MB for the RJPO. This last result is dueto the use of a conjugate gradient on which each matrix-vectorproduct is performed without explicitly writing the matrix .Finally, note that if we consider images of higher resolution, forinstance , the Cholesky factorization wouldrequire around 1 TB of RAM and the adaptive RJPO only about3 GB (when using double precision floating-point format).

VI. CONCLUSION

The sampling of high dimensional Gaussian distributions ap-pears in the resolution of many linear inverse problems using

Fig. 10. Evolution of the acceptance rate with respect to average conjugategradient iterations for sampling a Gaussian of dimension .

MCMC methods. Alternative solutions to the Cholesky fac-torization are needed to reduce the computation time and tolimit the memory usage. Based on the theory of reversible jumpMCMC,we derived a samplingmethod allowing to introduce anapproximate solution of a linear system during the sample gen-eration step. The approximate resolution of a linear system wasalready adopted in methods like IFP and PO to reduce the nu-merical complexity, but without any guarantee of convergenceto the target distribution. The proposed algorithm RJPO is basedon an accept-reject step that is absent from the existing PO al-gorithms. Indeed, the difference between RJPO and existing POalgorithms is much comparable to the difference between theMetropolis-adjusted Langevin algorithm (MALA) [48] and aplainly discretized Langevin diffusion [49].Our results pointed out that the required resolution accuracy

in these methods must be carefully tuned to prevent a significanterror. It was also shown that the proposed RJ-MCMC frame-work allows to ensure the convergence through the accept-rejectstep whatever the truncation level. In addition, thanks to the sim-plicity of the acceptance probability, the resolution accuracy canbe adjusted automatically using an adaptive scheme allowing toachieve a pre-defined acceptation rate. We have also proposed asignificant improvement of the same adaptive tuning approach,where the target is directly formulated in terms of minimal com-puting cost per effective sample.Finally, the linear system resolution using the conjugate gra-

dient algorithm offers the possibility to implement the matrix-vector products with a limited memory usage by exploiting thestructure of the forward model operators. The adaptive RJPOhas thus proven to be less consuming in both computationalcost and memory usage than any approach based on Choleskyfactorization.This work opens some perspectives in several directions.

Firstly, preconditioned conjugate gradient or alternativemethods can be envisaged for the linear system resolution withthe aim to reduce the computation time per iteration. Such anapproach will highly depend on the linear operator and theability to compute a preconditioning matrix. A second directionconcerns the connection between the RJ-MCMC frameworkand other sampling methods such as those based on Krylovsubspace [19], [20], particularly with appropriate choices ofthe parameters , , and defined in Section II. Another

Page 10: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

GILAVERT et al.: LARGE-SCALE INVERSE PROBLEMS USING MCMC 79

perspective of this work is to analyze more complex situationsinvolving non-gaussian distributions with the aim to be able toformulate the perturbation step and to perform an approximateoptimization allowing to reduce the computation cost. Finally,the proposed adaptive tuning scheme allowing to optimize thecomputation cost per effective sample could be generalized toother Metropolis adjusted sampling strategies.

APPENDIX

A. Expression of the Acceptance Probability

According to the RJ-MCMC theory, the acceptance proba-bility is given by

with and . The Jacobian determinant ofthe deterministic move is . Since

and

the acceptance probability can be written as

with and

Since , we get

Finally

Finally, when the system is solved exactly, and thus.

B. Correlation Between Two Successive Samples

Since

and is sampled from , we have

with totally independent of . One can firstly check that. Consequently, the correlation between two

successive samples is given by

which is zero if and only if .

REFERENCES

[1] M. Bertero and P. Boccacci, Introduction to Inverse Problems inImaging. London, UK: IOP, 1998.

[2] J. Idier, Bayesian Approach to Inverse Problems. New York, NY,USA: ISTE Ltd./Wiley, 2008.

[3] B. Frieden, “Image enhancement and restoration,” in Picture Pro-cessing and Digital Filtering, ser. Topics in Appl. Phys. New York,NY, USA: Springer-Verlag, 1975, vol. 6, pp. 177–248.

[4] G. Demoment, “Image reconstruction and restoration: Overview ofcommon estimation structure and problems,” IEEE Trans. Acoust.Speech, Signal Process., vol. 37, no. 12, pp. 2024–2036, Dec. 1989.

[5] G. T. Gordon and R. Herman, “Reconstruction of pictures from theirprojections,” Commun. ACM, vol. 14, no. 12, pp. 759–768, 1971.

[6] R. M. Lewitt and S. Matej, “Overview of methods for image recon-struction from projections in emission computed tomography,” Proc.Inst. Electr. Eng., vol. 91, no. 1, pp. 1588–1611, 2003.

[7] W. Gilks, S. Richardson, and D. Spiegehalter, Markov Chain MonteCarlo in Practice. London, UK: Chapman & Hall, 1999.

[8] C. Robert, The Bayesian Choice, 2nd ed. New York, NY, USA:Springer-Verlag, 2001.

[9] D. Andrews and C. Mallows, “Scale mixtures of normal distributions,”J. R. Statist. Soc. B, pp. 99–102, 1974.

[10] F. Champagnat and J. Idier, “A connection between half-quadratic cri-teria and EM algorithms,” IEEE Signal Process. Lett., vol. 11, no. 9,pp. 709–712, 2004.

[11] S. Geman and D. Geman, “Stochastic relaxation, Gibbs distributions,the Bayesian restoration of images,” IEEE Trans. Pattern Anal. Mach.Intell., vol. PAMI-6, no. 6, pp. 721–741, Nov. 1984.

[12] G. Papandreou and A. Yuille, “Gaussian sampling by local perturba-tions,” Adv. Neural Inf. Process. Syst., vol. 23, pp. 1858–1866, 2010.

[13] E. M. Scheuer and D. S. Stoller, “On the generation of normal randomvectors,” Technometr., vol. 4, no. 2, pp. 278–281, 1962.

[14] D. R. Barr and N. L. Slezak, “A comparison ofmultivariate normal gen-erators,” Commun. ACM, vol. 15, no. 12, pp. 1048–1049, Dec. 1972.

[15] H. Rue, “Fast sampling of Gaussian Markov random fields,” J. R.Statist. Soc. B, vol. 63, no. 2, pp. 325–338, 2001.

[16] W. F. Trench, “An algorithm for the inversion of finite Toeplitz ma-trices,” J. Soc. Indust. Appl. Math., vol. 12, no. 3, pp. 515–522, 1964.

[17] D. Geman and C. Yang, “Nonlinear image recovery with half-quadraticregularization,” IEEE Trans. Image Process., vol. 4, no. 7, pp.932–946, 1995.

[18] P. Lalanne, D. Prévost, and P. Chavel, “Stochastic artificial retinas:Algorithm, optoelectronic circuits, implementation,” Appl. Opt., vol.40, no. 23, pp. 3861–3876, 2001.

[19] A. Parker and C. Fox, “Sampling Gaussian distributions in Krylovspaces with conjugate gradients,” SIAM J. Sci. Comput., vol. 34, no.3, pp. B312–B334, 2012.

[20] E. Aune, J. Eidsvik, and Y. Pokern, “Iterative numerical methodsfor sampling from high dimensional Gaussian distributions,” Statist.Comp., pp. 1–21, 2013.

[21] E. Chow and Y. Saad, “Preconditioned Krylov subspace methods forsampling multivariate Gaussian distributions,” SIAM J. Sci. Comput.,vol. 36, no. 2, pp. A588–A608, 2014.

[22] Y. Amit and U. Grenander, “Comparing sweep strategies for stochasticrelaxation,” J. Multivar. Anal., vol. 37, no. 2, pp. 197–222, 1991.

[23] F. Orieux, O. Féron, and J. Giovannelli, “Sampling high-dimensionalGaussian distributions for general linear inverse problems,” IEEESignal Process. Lett., vol. 19, no. 5, p. 251, 2012.

[24] X. Tan, J. Li, and P. Stoica, “Efficient sparse Bayesian learning viaGibbs sampling,” in Proc. ICASSP, 2010, pp. 3634–3637.

Page 11: 70 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, …pagesperso.ls2n.fr/~idier-j/pub/pubC/Gilavert15C.pdf · lems and presents the adaptive RJPO which incorporates an au-tomatic

80 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 63, NO. 1, JANUARY 1, 2015

[25] P. J. Green, “Reversible jump Markov chain Monte Carlo computationand Bayesian model determination,” Biometrika, vol. 82, pp. 711–732,1995.

[26] R. Waagepetersen and D. Sorensen, “A tutorial on reversible jumpMCMCwith a view toward applications in QTL-mapping,” Int. Statist.Rev., vol. 69, no. 1, pp. 49–61, 2001.

[27] C. Andrieu and C. Robert, “Controlled MCMC for optimal sampling,”Cahiers de Mathématiques du Ceremade, Univ. Paris-Dauphine, Rep.No. 0125, 2001.

[28] C. Andrieu and J. Thoms, “A tutorial on adaptive MCMC,” Statist.Comp., vol. 18, no. 4, pp. 343–373, 2008.

[29] Y. Atchade, G. Fort, E. Moulines, and P. Priouret, “Adaptive Markovchain Monte Carlo: Theory and methods,” in Bayesian Time SeriesModels. Cambridge, U.K.: Cambridge Univ. Press., 2011, pp. 33–53.

[30] P. De Forcrand, “Monte Carlo quasi-heatbath by approximate inver-sion,” Phys. Rev. D, vol. 59, no. 3, pp. 3698–3701, 1999.

[31] R. Barrett, M. Berry, T. F. Chan, J. Demmel, J. Donato, J. Dongarra, V.Eijkhout, R. Pozo, C. Romine, and H. van der Vorst, Templates for theSolution of Linear Systems: Building Blocks for Iterative Methods.Philadelphia, PA, USA: SIAM, 1994.

[32] A. Gelman and D. B. Rubin, “Inference from iterative simulation usingmultiple sequences,” Statist. Comp., pp. 457–472, 1992.

[33] J. S. Liu, Monte Carlo Strategies in Scientific Computing, ser. SpringerSeries in Statistics, 2nd ed. New York, NY, USA: Springer, 2008.

[34] J. Goodman and A. D. Sokal, “Multigrid Monte Carlo method. Con-ceptual foundations,” Phys. Rev. D, vol. 40, no. 6, 1989.

[35] H. Robbins and S. Monro, “A stochastic approximation method,” Ann.Math. Statist., vol. 22, pp. 400–407, 1951.

[36] B. Bercu and P. Fraysse, “A Robbins-Monro procedure for estimationin semiparametric regression models,” Ann. Statist., vol. 40, no. 2, pp.666–693, 2012.

[37] A. Benveniste, M. Métivier, and P. Priouret, Adaptive Algorithms andStochastic Approximations. New York, NY, USA: Springer, 2012.

[38] C. Andrieu, E. Moulines, and P. Priouret, “Stability of stochastic ap-proximation under verifiable conditions,” SIAM J. Contr. Optimiz., vol.44, no. 1, pp. 283–312, 2006.

[39] H. Haario, E. Saksman, and J. Tamminen, “An adaptive Metropolisalgorithm,” Bernouilli, vol. 7, no. 2, pp. 223–242, 2001.

[40] Y. F. Atchadé and J. S. Rosenthal, “On adaptive Markov chain MonteCarlo algorithms,” Bernouilli, vol. 11, no. 5, pp. 815–828, 2005.

[41] G. O. Roberts, A. Gelman, and W. R. Gilks, “Weak convergence andoptimal scaling of random walk Metropolis algorithms,” Ann. Appl.Probabil., vol. 7, no. 1, pp. 110–120, 1997.

[42] A. Gelman, G. O. Roberts, and W. R. Gilks, “Efficient Metropolisjumping rules",” in Bayesian Statistics 5. Oxford, U.K.: OxfordUniv. Press, 1996, pp. 599–607.

[43] Y. F. Achadé, “An adaptive version for the metropolis adjustedLangevin algorithm with a truncated drift,” Methodol. Comput. Appl.Probabil., vol. 8, no. 2, pp. 235–254, 2006.

[44] G. O. Roberts and J. S. Rosenthal, “Optimal scaling of discrete approx-imations to Langevin’s diffusions,” J. R. Statist. Soc. B, vol. 60, no. 1,pp. 255–268, 1998.

[45] S. C. Park, M. K. Park, and M. G. Kang, “Super-resolution image re-construction: A technical overview,” IEEE Trans. Signal Process., vol.20, no. 3, pp. 21–36, 2003.

[46] G. Rochefort, F. Champagnat, G. Le Besnerais, and J.-F. Giovannelli,“An improved observation model for super-resolution under affine mo-tion,” IEEE Trans. Image Process., vol. 15, no. 11, pp. 3325–3337,2006.

[47] H. Jeffreys, “An invariant form for the prior probability in estimationproblems,” Proc. Royal Soc. London. Ser. A. Math. Phys. Sci., vol. 186,no. 1007, pp. 453–461, 1946.

[48] G. O. Roberts and R. L. Tweedie, “Exponential convergence ofLangevin distributions and their discrete approximations,” Bernoulli,vol. 2, no. 4, pp. 341–363, 1996.

[49] G. Parisi, “Correlation functions and computer simulations,” Nucl.Phys. B, vol. 180, pp. 378–384, 1981.

Clément Gilavert received the engineering degreefrom École Centrale, Nantes, France, in 2011.From October 2011 to March 2014, he was a Ph.D.

student with the Institut de Recherche en Commu-nications et Cybernétique, Nantes (IRCCYN, UMRCNRS 6597). His Ph.D. thesis work is related tostatistical inference methods for large-scale inverseproblems in signal and image processing.

Saïd Moussaoui (M’13) received the State engi-neering degree from Ecole Nationale Polytechnique,Algiers, Algeria, in 2001, and the Ph.D. degreein automatic control and signal processing fromUniversité Henri Poincaré, Nancy, France, in 2005.He is currently an Assistant Professor with

École Centrale de Nantes, Nantes, France. SinceSeptember 2006, he has been with the Institut deRecherche en Communications et Cybernétique,Nantes (IRCCYN, UMR CNRS 6597). His researchinterests are in statistical signal and image processing

methods including source separation, image restoration, and their applicationsto spectrometry and hyperspectral imaging.

Jérôme Idier (M’09) was born in France in 1966. Hereceived the Diploma degree in electrical engineeringfrom École Supérieure d’Électricité, Gif-sur-Yvette,France, in 1988, and the Ph.D. degree in physics fromUniversity of Paris-Sud, Orsay, France, in 1991. In1991, he joined the Centre National de la RechercheScientifique.He is currently a Senior Researcher with the In-

stitut de Recherche en Communications et Cyberné-tique, Nantes, France. His major scientific interestsare in probabilistic approaches to inverse problems

for signal and image processing.Dr. Idier is currently an elected member of the French National Committee

for Scientific Research.