15
Journal of Statistical Physics manuscript No. (will be inserted by the editor) A path-integral approach to Bayesian inference for inverse problems using the semiclassical approximation Joshua C. Chang · Van M. Savage · Tom Chou July 23, 2014 Abstract We demonstrate how path integrals often used in problems of theoretical physics can be adapted to provide a machinery for performing Bayesian inference in function spaces. Such inference comes about naturally in the study of inverse problems of recovering continuous (innite dimensional) coecient functions from ordinary or partial dierential equations (ODE, PDE), a problem which is typically ill-posed. Regularization of these problems using L 2 function spaces (Tikhonov regularization) is equivalent to Bayesian probabilistic inference, using a Gaussian prior. The Bayesian interpretation of inverse problem regularization is useful since it allows one to quantify and characterize error and degree of precision in the solution of inverse problems, as well as examine assumptions made in solving the problem – namely whether the subjective choice of regularization is compatible with prior knowledge. Using path-integral formalism, Bayesian inference can be explored through various perturbative techniques, such as the semiclassical approximation, which we use in this manuscript. Perturbative path-integral approaches, while oering alternatives to computational approaches like Markov-Chain-Monte-Carlo (MCMC), also provide natural starting points for MCMC methods that can be used to rene approximations. In this manuscript, we illustrate a path-integral formulation for inverse problems and demonstrate it on an inverse problem in membrane biophysics as well as inverse problems in potential theories involving the Poisson equation. Keywords Inverse problems · Bayesian inference · Field theory · Path integral · Potential theory · Semiclassical approximation 1 Introduction One of the main conceptual challenges in solving inverse problems results from the fact that most interesting inverse problems are not well-posed. One often chooses a solution that is “useful," or that optimizes some regularity criteria. Such a task is commonly known as regularization, of which there are many variants. One of the most commonly used methods is Tikhonov Regularization, or L 2 -penalized regularization [8, 9, 20, 32, 44]. Here we rst demonstrate the concept behind Tikhonov regularization using one of the simplest inverse problems, the interpolation problem. Tikhonov regularization, when applied to interpolation, solves the inverse problem of J.C. Chang Mathematical Biosciences Institute,The Ohio State University Jennings Hall, 3rd Floor, 1735 Neil Avenue, Columbus, Ohio 43210 E-mail: [email protected] T. Chou UCLA Departments of Biomathematics and Mathematics BOX 951766, Room 5309 Life Sciences, Los Angeles, CA 90095-1766 E-mail: [email protected] V.M. Savage Santa Fe Institute, Santa Fe, NM 87501 UCLA Department of Biomathematics and Department of Ecology and Evolutionary Biology BOX 951766, Room 5205 Life Sciences, Los Angeles, CA 90095-1766 E-mail: [email protected] arXiv:1312.2974v4 [physics.data-an] 22 Jul 2014

Journal of Statistical Physics manuscript No

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Journal of Statistical Physics manuscript No

Journal of Statistical Physics manuscript No.(will be inserted by the editor)

A path-integral approach to Bayesian inference for inverse problems using thesemiclassical approximation

Joshua C. Chang · Van M. Savage · Tom Chou

July 23, 2014

Abstract We demonstrate how path integrals often used in problems of theoretical physics can be adapted to providea machinery for performing Bayesian inference in function spaces. Such inference comes about naturally in the studyof inverse problems of recovering continuous (infinite dimensional) coefficient functions from ordinary or partialdifferential equations (ODE, PDE), a problem which is typically ill-posed. Regularization of these problems usingL2 function spaces (Tikhonov regularization) is equivalent to Bayesian probabilistic inference, using a Gaussianprior. The Bayesian interpretation of inverse problem regularization is useful since it allows one to quantify andcharacterize error and degree of precision in the solution of inverse problems, as well as examine assumptions madein solving the problem – namely whether the subjective choice of regularization is compatible with prior knowledge.Using path-integral formalism, Bayesian inference can be explored through various perturbative techniques, suchas the semiclassical approximation, which we use in this manuscript. Perturbative path-integral approaches, whileoffering alternatives to computational approaches like Markov-Chain-Monte-Carlo (MCMC), also provide naturalstarting points for MCMC methods that can be used to refine approximations. In this manuscript, we illustrate apath-integral formulation for inverse problems and demonstrate it on an inverse problem in membrane biophysicsas well as inverse problems in potential theories involving the Poisson equation.

Keywords Inverse problems · Bayesian inference · Field theory · Path integral · Potential theory · Semiclassicalapproximation

1 Introduction

One of the main conceptual challenges in solving inverse problems results from the fact that most interesting inverseproblems are not well-posed. One often chooses a solution that is “useful," or that optimizes some regularity criteria.Such a task is commonly known as regularization, of which there are many variants. One of the most commonly usedmethods is Tikhonov Regularization, or L2-penalized regularization [8, 9, 20, 32, 44].

Here we first demonstrate the concept behind Tikhonov regularization using one of the simplest inverse problems,the interpolation problem. Tikhonov regularization, when applied to interpolation, solves the inverse problem of

J.C. ChangMathematical Biosciences Institute,The Ohio State UniversityJennings Hall, 3rd Floor, 1735 Neil Avenue, Columbus, Ohio 43210E-mail: [email protected]

T. ChouUCLA Departments of Biomathematics and MathematicsBOX 951766, Room 5309 Life Sciences, Los Angeles, CA 90095-1766E-mail: [email protected]

V.M. SavageSanta Fe Institute, Santa Fe, NM 87501UCLA Department of Biomathematics and Department of Ecology and Evolutionary BiologyBOX 951766, Room 5205 Life Sciences, Los Angeles, CA 90095-1766E-mail: [email protected]

arX

iv:1

312.

2974

v4 [

phys

ics.

data

-an]

22

Jul 2

014

Page 2: Journal of Statistical Physics manuscript No

constructing a continuous function ϕ : Rd → R from point-wise measurements ϕobs at positions xm by seekingminima with respect to a cost functional of the form

H[ϕ] =1

2

M∑m=1

1

s2m(ϕ(xm)− ϕobs(xm))2︸ ︷︷ ︸Hobs[ϕ]

+1

2

∑α

γα

∫|Dαϕ|2 dx︸ ︷︷ ︸

Hreg[ϕ]

, (1)

where the constants 1/s2m, γα > 0 are weighting parameters, and Dα =∏dj=1(−i∂xj )

αj is a differential operatorof order α = (α1, . . . , αd).

Assuming Dα is isotropic and integer-ordered, it is possible to invoke integration-by-parts to write H[ϕ] in thequadratic form

H[ϕ] =1

2

M∑m=1

1

s2m(ϕ(xm)− ϕobs(xm))2 +

1

2

∫ϕ(x)P (−∆)ϕ(x) dx, (2)

where P (·) is a polynomial of possibly infinite order, ∆ is the Laplacian operator, and we have assumed thatboundary terms vanish. In the remainder of this work, we will focus on energy functionals of this form. Thisexpression is known in previous literature as the Information Hamiltonian [11].

Using this form of regularization serves two primary purposes. First, it selects smooth solutions to the inverseproblem, with the amount of smoothness controlled by Hreg. For example, if only Hobs is used, the solution can beany function that connects the observations ϕobs at the measured points xj , such as a piecewise affine solution. Yet,such solutions may be physically unreasonable (not smooth). Second, it transforms the original inverse problem into aconvex optimization problem that possesses an unique solution [3, 10]. If all of the coefficients of P are non-negative,then the pseudo-differential-operator P (−∆) is positive-definite [22], guaranteeing uniqueness. These features ofTikhonov regularization make it attractive; however, one needs to make certain choices. In practical settings, onewill need to chose both the degree of the differential operator and value of the parameters γα. These two choicesadjust the trade-off between data agreement and regularity.

1.1 Bayesian inverse problems

The problem of parameter selection for regularization is well-addressed in the context of Bayesian inference, whereregularization parameters can be viewed probabilistically as prior-knowledge of the solution. Bayesian inference overcontinuous function spaces has been applied to inverse problems in several contexts. One of the first applicationsof Bayesian inference to inverse problems was in the study of quantum inverse problems [28], where it was notedthat Gaussian priors could be used to formulate field theories. Subsequently, variants of this methodology havebeen used for model reduction [29] and applied to many interpolation problems and inverse problems in fluidmechanics [6, 19, 43], geology [13, 31, 38], cosmology [11, 34], and biology [18].

There is a wealth of literature concerning the computational aspects of Bayesian inverse problems. Many ofthese works on inverse problems are viewed through the framework and language of data assimilation throughMarkov Chain Monte Carlo approaches [4, 4, 37, 39, 40]. Approximation methods based on sparsity have also beendeveloped [42]. Finally, there is a large body of work on the theoretical aspects of maximum aposteriori inference forBayesian inverse problems including questions of existence of solutions and convergence to solutions [7, 25–27, 43]

2 Field-theoretic formulation

Bayesian inference on ϕ entails the construction of a probability density π known as the posterior distribution π(ϕ)which obeys Bayes’ rule,

π(ϕ) =

likelihood︷ ︸︸ ︷Pr(ϕobs|ϕ)

prior︷ ︸︸ ︷Pr(ϕ)

Z[0](3)

where Z[0] is the partition function or normalization factor. The posterior density π is a density in a space offunctions. The inverse problem is then investigated by computing the statistics of the posterior probability densityπ(φ) through the evaluation of Z[0]. The solution of the inverse problem corresponds to the specific ϕ that maximizesπ(φ), subject to prior knowledge encoded in the prior probability density Pr(ϕ). This solution is known as the meanfield solution. The variance, or error, of the mean field solution is found by computing the variance of the posteriordistribution about the mean field solution.

2

Page 3: Journal of Statistical Physics manuscript No

This view of inverse problems also leads naturally to the use of functional integration and perturbation methodscommon in theoretical physics [24, 45]. Use of the probabilistic viewpoint allows for exploration of inverse problemsbeyond mean field, with the chief advantage of providing a method for uncertainty quantification.

As shown in [13, 28], Tikhonov regularization has the probabilistic interpretation of Bayesian inference witha Gaussian prior distribution. That is, the regularization term in Eq 2 combines with the data term to specify aposterior distribution of the form

π(ϕ|ϕobs) =1

Z[0]e−H[ϕ]

=1

Z[0]exp

M∑m=1

1

s2m(ϕ(xm)− ϕobs(xm))2

︸ ︷︷ ︸

likelihood (exp−Hobs)

exp

−1

2

∫ϕ(x)P (−∆)ϕ(x) dx

︸ ︷︷ ︸

prior (exp−Hreg)

(4)

where the partition function

Z[0] =

∫Dϕe−H[ϕ] =

∫Dϕe−Hreg[ϕ]︸ ︷︷ ︸

dW [ϕ]

e−Hobs[ϕ] (5)

is a sum over the contributions of all functions in the separable Hilbert space ϕ : Hreg[ϕ] < ∞. This sum isexpressed as a path integral, which is an integral over a function space. The formalism for this type of integral cameabout first from high-energy theoretical physics [14], and then found application in nearly all areas of physics as wellas in the representation of both Markovian [5, 15, 35], and non-Markovian [16, 36] stochastic processes. In the caseof Eq. 5, where the field theory is real-valued and the operator P (−∆) is self-adjoint, a type of functional integralbased on abstract Wiener measure may be used [23]. The abstract Wiener measure dW [ϕ] used for Eq. 5 subsumesthe prior term Hreg, and it is helpful to think of it as a Gaussian measure over lattice points taken to the continuumlimit.

When the functional integral of the exponentiated energy functional can be written in the form

Z[0] =

∫Dϕ exp

−1

2

∫∫ϕ(x)A(x,x′)ϕ(x′) dx dx′ +

∫b(x)ϕ(x) dx

, (6)

then the probability density is Gaussian in function-space and the functional integral of Eq. 6 has the solution [45]

Z[0] = exp

1

2

∫∫b(x)A−1(x,x′)b(x′) dx dx′ − 1

2log detA

. (7)

The operators A(x,x′) and A−1(x,x′) are related through the relationship∫A(x,x′)A−1(x′,x′′) dx′ = δ(x− x′′). (8)

Upon neglecting Hobs, the functional integral of Eq. 5 can be expressed in the form of Eq. 6 with A(x,x′) =P (−∆)δ(x− x′). The pseudo-differential-operator P (−∆) acts as an infinite-dimensional version of the inverse ofa covariance matrix. It encodes the a-priori spatial correlation, implying that values of the function ϕ are spatiallycorrelated according to a correlation function (Green’s function) A−1(x,y) = G(x,y) : Rd × Rd → R through the

relationship implied by Eq. 8, P (−∆)G(x,y) = δ(x−y) so that G(x,y) =(

12π

)d ∫Rd e−ik·(y−x) 1

P (|k|2) dk where

P (|k|2) is the symbol of the pseudo-differential-operator P (−∆). It is evident that when performing Tikhonov reg-ularization, one should chose regularization that is reflective of prior knowledge of correlations, whenever available.

2.1 Mean field inverse problems

We turn now to the more-general problem, where one seeks recovery of a scalar function ξ given measurements of acoupled scalar function ϕ over interior points xi, and the relationship between the measured and desired functionsis given by a partial differential equation

F (ϕ(x), ξ(x)) = 0 x ∈ Ω \ ∂Ω. (9)

3

Page 4: Journal of Statistical Physics manuscript No

As before, we regularize ξ using knowledge of its spatial correlation, and write a posterior probability density

π[ϕ, ξ|ϕobs] =

δ (F (ϕ, ξ))

Z[0]exp

−1

2

∫ M∑m=1

δ(x− xm)(ϕ(x)− ϕobs(x))

2

s2mdx− 1

2

∫ξ(x)P (−∆)ξ(x) dx

,

where we have used the Dirac-delta function δ to specify that our observations are taken with noise s2m at certain po-sitions xm, and an infinite-dimensional delta functional δ to specify that F (ϕ, ξ) = 0 everywhere. Using the inverseFourier-transformation, one can represent δ in path-integral form as δ (F (ϕ, ξ)) =

∫Dλe−i

∫λ(x)F (ϕ(x),ξ(x)) dx,

where λ(x), is a Fourier wavevector. The reason for this notation will soon be clear. We now have a posteriorprobability distribution of three functions ϕ, ξ, λ of the form

π[ϕ, ξ, λ(x)|ϕobs] =1

Z[0]exp −H[ϕ, ξ, λ] , (10)

where the partition functional is

Z[0] =

∫∫∫DϕDξDλ exp −H[ϕ, ξ, λ] , (11)

and the Hamiltonian

H[ϕ, ξ, λ;ϕobs] =1

2

∫ M∑m=1

δ(x− xm)(ϕ(x)− ϕobs(x))

2

s2mdx

+1

2

∫ξ(x)P (−∆)ξ(x) dx + i

∫λ(x)F (ϕ, ξ) dx, (12)

is a functional of ϕ, ξ, and the Fourier wave vector λ(x). Similar Hamiltonians, providing a probabilistic model fordata in the context of inverse problems, have appeared in previous literature [11, 28, 43], where they have beenreferred to as Information Hamiltonians.

Maximization of the posterior probability distribution, also known as Bayesian maximum a posteriori estimation(MAP) inference, is performed by minimization of the corresponding energy functional (Eq. 12) with respect to thefunctions ϕ, ξ, λ. One may perform this inference by solving the associated Euler-Lagrange equations

P (−∆)ξ +δ

δξ(x)

∫λ(x)F (ϕ, ξ) dx = 0, (13)

M∑n=1

δ(x− xn)(ϕ(x)− ϕobs(x)) +δ

δϕ(x)

∫λ(x)F (ϕ, ξ) dx = 0 (14)

F (ϕ, ξ) = 0, (15)

where λ(x) here serves the role of a Lagrange multiplier. Solving this system of partial differential equations simul-taneously allows one to arrive at the solution to the original Tikhonov-regularized inverse problem. Now, supposeone is interested in estimating the precision of the given solution. The field-theoretic formulation of inverse problemsprovides a way of doing so.

2.2 Beyond mean-field – semiclassical approximation

The functions ϕ, ξ, λ : Rd → R each constitute scalar fields1. Field theory is the study of statistical properties ofsuch fields through evaluation of an associated path integral (functional integral). Field theory applied to Bayesianinference has appeared in prior literature under the names Bayesian Field theory [13, 28, 43], and Information FieldTheory [11].

In general, field theory deals with functional integrals of the form

Z[J ] =

∫Dϕ exp

−[1

2

∫∫ϕ(x)A(x,x′)ϕ(x′) dx dx′ +

∫V [ϕ(x)] dx

]︸ ︷︷ ︸

H[ϕ]

+

∫J(x)ϕ(x) dx

, (16)

1 We will use Greek letters to denote fields

4

Page 5: Journal of Statistical Physics manuscript No

where the Hamiltonian of interest is recovered when the source J = 0, and the potential function V is nonlinearin ϕ. Assuming that after non-dimensionalization, V [ϕ] is relatively small in comparison to the other terms, one isthen able to expand the last term in formal Taylor series so that after completing the Gaussian part of the integralas in Eq. 7,

Z[J ] =

∫Dϕ

exp

[−1

2

∫∫ϕ(x)A(x,x′)ϕ(x′) dx dx′ +

∫J(x)ϕ(x) dx

]︸ ︷︷ ︸

Gaussian

×(1−

∫V [ϕ] dx + . . .

)

∝ exp

[−V

δJ

)]exp

(1

2

∫∫J(x)A−1(x,x′)J(x′) dx dx′

). (17)

In this way, Z[J ] can be expressed in series form as moments of a Gaussian distribution. The integral is ofinterest because one can use it to recover moments of the desired field through functional differentiation,⟨∏

k

ϕ(xk)

⟩=

1

Z[0]

∏k

δ

δJ(xk)Z[J ]

∣∣∣∣∣J=0

. (18)

This approach is known as the weak-coupling approach [45]. For this expansion to hold, however, the external potentialV must be small in size compared to the quadratic term. This assumption is not generally valid during Tikhonovregularization, as common rules of thumb dictate that the data fidelity and the regularization term should be ofsimilar order of magnitude [2, 41]. Another perturbative approach – the one that we will take in this manuscript – isto expand the Hamiltonian in a functional Taylor series

H[ϕ] = H[ϕ?] +1

2

∫∫δ2H[ϕ?]

δϕ(x)ϕ(x′)(ϕ(x)− ϕ?(x))(ϕ(x′)− ϕ?(x′)) dx dx′ + . . . (19)

about its extremal point ϕ?. To the second order (as shown), the expansion is known as the semiclassical approxima-tion [17] which provides an approximate Gaussian density for the field ϕ. Corrections to the semiclassical expansioncan be evaluated by continuing this expansion to higher orders, where evaluation of the functional integral can beaided by the use of Feynman diagrams [14].

2.3 Monte-Carlo for refinement of approximations

The Gaussian approximation is useful because Gaussian densities are easy to sample. One may sample a random fieldϕ(x) from a Gaussian distribution with inverse-covariance A(x,x′) by solving the stochastic differential equation

1

2

∫A(x,x′)ϕ(x′) dx′ = η(x), (20)

where η is the unit white noise process which has mean 〈η(x)〉 = 0, and spatial correlation⟨η(x)η(x′)

⟩= δ(x−x′).

With the ability to sample from the approximating Gaussian distribution of Eq. 19, one may use Monte-Carlosimulation to sample from the true distribution by weighting the samples obtained from the Gaussian distribution.Such an approach is known as importance sampling [30], where samples ϕi are given importance weights wi accordingto the ratio wi = exp (−Happrox +Htrue) /

∑j wj . Statistics of ϕ may then be calculated using the weighted samples;

for instance expectations can be approximated as 〈g(ϕ(x))〉 ≈∑i wig(ϕi(x)). Using this method, one can refine

the original estimates of the statistics of ϕ.

3 Examples

3.1 Interpolation of the height of a rigid membrane or plate

We first demonstrate the field theory for inverse problems on an interpolation problem where one is able to determinethe regularizing differential operator based on prior knowledge. This example corresponds to the interpolation ex-ample mentioned in the Introduction. Consider the problem where one is attempting to identify in three-dimensions

5

Page 6: Journal of Statistical Physics manuscript No

the position of a membrane. For simplicity, we assume that one is interested in obtaining the position of the mem-brane only over a restricted spatial domain, where one can use the Monge parameterization to reduce the problemto two-dimensions and define the height of the membrane ϕ : R2 → R.

Suppose one is able to measure the membrane in certain spatial locations xm, but one seeks to also interpolatethe membrane in regions that are not observable. Physically, models for fluctuations in membranes are well known,for instance the Helfrich free-energy [12] suggests that one should use a regularizing differential operator

P (−∆) = β(κ∆2 − σ∆) β, σ, κ > 0, (21)

where σ and κ are the membrane tension and bending rigidity, respectively. The Hamiltonian associated with theHelfrich operator is

H[ϕ;ϕobs] =1

2

∫ M∑m=1

δ(x− xm)

s2m(ϕ(x)− ϕobs(x))

2 dx +1

2

∫ϕ(x)P (−∆)ϕ(x) dx, (22)

and the mean-field solution for ϕ corresponds to the extremal point of the Hamiltonian, which is the solution of thecorresponding Euler-Lagrange equation

δH

δϕ=

M∑m=1

δ(x− xm)

s2m(ϕ(x)− ϕobs(x)) + P (−∆)ϕ(x) = 0. (23)

To go beyond mean-field, one may compute statistics of the probability distribution Pr(ϕ) ∝ e−H[ϕ], using thegenerating functional which is expressed as a functional integral

Z[J ] ∝∫Dϕ exp

− 1

2

∫∫ϕ(x)

[δ(x− x′)

M∑m=1

δ(x′ − xm)

s2m+ P (−∆)δ(x− x′)

]︸ ︷︷ ︸

A(x,x′)

ϕ(x′) dx dx′

+

∫ [ M∑m=1

ϕobs(x)δ(x− xm)

s2m+ J(x)

]ϕ(x) dx

, (24)

where we have completed the square. According to Eq. 7, Eq. 24 has the solution

Z[J ] ∝ exp

1

2

∫∫J(x)A−1(x,x′)J(x′) dx′ dx +

∫J(x)

M∑m=1

ϕobs(xm)A−1(x,xm)

s2mdx

. (25)

Through functional differentiation of Eq. 25, Eq. 18 implies that the mean-field solution is

〈ϕ(x)〉 =M∑m=1

ϕobs(xm)A−1(x,xm)

s2m, (26)

and variance in the solution is ⟨ϕ(x)− 〈ϕ(x)〉 , ϕ(x′)−

⟨ϕ(x′)

⟩ ⟩= A−1(x,x′). (27)

To solve for these quantities, we compute the operator A−1, which according to Eq. 8, satisfies the partialdifferential equation

M∑m=1

δ(xm − x)

s2mA−1(x,x′′) + P (−∆)A−1(x,x′′) = δ(x− x′′). (28)

Using the Green’s function for P (−∆),

G(x,x′) =−1

2πβσ

[log(|x− x′|

)+K0

(√σ

κ|x− x′|

)], (29)

we find

6

Page 7: Journal of Statistical Physics manuscript No

A−1(x,x′′) =

known︷ ︸︸ ︷G(x,x′′)−

M∑m=1

known︷ ︸︸ ︷G(x,xm)

unknown︷ ︸︸ ︷A−1(xm,x

′′)

s2m. (30)

To calculate A−1(x,x′), we need A−1(xm,x′), for m ∈ 1, . . . ,M. Solving for each of these simultaneously

yields the equation

A−1(x,x′) = G(x,x′)−Gs(x) (I +Λ)−1 G(x′), (31)

where Gs(x) ≡[G(x,x1)s21

, G(x,x2)s22

, . . . , G(x,xM )s2M

], G(x) ≡ [G(x,x1), G(x,x2), . . . , G(x,xM )], and Λij ≡

G(xi,xj)/s2i .

0.0 0.5 1.0 1.5 2.0 2.5 0.00.5

1.01.5

2.02.5−6−5−4−3−2−10123

(a) ϕ(x)

0.0 0.5 1.0 1.5 2.0 2.5 0.00.5

1.01.5

2.02.5−6−5−4−3−2−1012

(b) 〈ϕ(x)〉

0.0 0.5 1.0 1.5 2.0 2.5 0.00.5

1.01.5

2.02.50.00

0.05

0.10

0.15

0.20

(c)√〈(ϕ(x)− 〈ϕ(x)〉)2〉

Fig. 1: Interpolation of a membrane. (a) A simulated membrane undergoing thermal fluctuations is the object ofreconstruction. (b) Mean-field reconstruction of the membrane using 100 randomly-placed measurements with noise.(c) Pointwise standard error in the reconstruction of the membrane. Parameters used: σ = 10−2, β = 103, κ =10−4, sm = 10−2.

Fig. 1 shows an example of the use of the Helfrich free energy for interpolation. A sample of a membrane under-going thermal fluctuations was taken as the object of recovery. Uniformly, 100 randomly-placed, noisy observations ofthe height of the membrane were taken. The mean-field solution for the position of the membrane and the standarderror in the solution are presented. The standard error is not uniform and dips to approximately the measurementerror at locations where measurements were taken.

3.2 Source recovery for the Poisson equation

Now consider an example where the function to be recovered is not directly measured. This type of inverse problemoften arises when considering the Poisson equation in isotropic medium:

∆ϕ(x) = ρ(x). (32)

Measurements of ϕ are taken at points xm and the objective is to recover the source function ρ(x). Previousresearchers have explored the use of Tikhonov regularization to solve this problem [1, 21]; here we quantify theprecision of such solutions.

Making the assumption that ρ is correlated according to the Green’s function of the pseudo-differential-operatorP (−∆), we write the Hamiltonian

H[ϕ, ρ, λ;ϕobs] =1

2

∫ M∑m=1

δ(x− xm)

s2m(ϕ(x)− ϕobs(x))

2 dx +1

2

∫ρ(x)P (−∆)ρ(x) dx

+ i

∫λ(x) (∆ϕ(x)− ρ(x)) dx. (33)

7

Page 8: Journal of Statistical Physics manuscript No

The extremum of H[ϕ, ρ, λ;ϕobs] occurs at (ϕ?, ρ?), which are found through the corresponding Euler-Lagrange

equations(δHδϕ = 0, δHδρ = 0, δHiδλ = 0

),

M∑m=1

δ(x− xm)

s2m(ϕ?(x)− ϕobs(x)) + P (−∆)∆2ϕ?(x) = 0,

ρ? = ∆ϕ?. (34)

In addition to the extremal solution, we can also evaluate how precisely the source function has been recovered byconsidering the probability distribution given by the exponentiated Hamiltonian,

π(ρ(x)|ϕobs(xi)) =1

Z[0]

× exp

−1

2

∫ M∑m=1

δ(x− xm)

s2m(ϕ(x)− ϕobs(x))

2 dx− 1

2

∫∆ϕ(x)P (−∆)∆ϕ(x) dx

, (35)

where we have integrated out the λ and ρ variables by making the substitution ρ = ∆ϕ. To compute the statistics ofϕ, we first compute Z[J ], the generating functional which by Eq. 7 has the solution

Z[J ] ∝ exp

1

2

∫∫∆J(x)A−1(x,x′)∆x′J(x′) dx′ dx

+

∫J(x)∆

M∑m=1

ϕobs(xm)A−1(x,xm)

s2mdx

, (36)

where

A(x,x′) = ∆2P (−∆)δ(x− x′) + δ(x− x′)M∑m=1

δ(x− xm)

s2m(37)

and A−1 is defined as in Eq. 8. The first two moments have the explicit solution given by the generating functional,

δZ[J ]

δJ(x)

∣∣∣∣J=0

=

(M∑m=1

ϕobs(xm)∆A−1(x,xm)

s2m

)Z[0]

δ2Z[J ]

δJ(x)δJ(x′)

∣∣∣∣J=0

= Z[0]

[∆∆x′A−1(x,x′)

+

(M∑m=1

ϕobs(xm)∆A−1(x,xm)

s2m

)(M∑k=1

ϕobs(xk)∆x′A−1(x′,xk)

s2k

)].

These formulae imply that our mean-field source has the solution

〈ρ(x)〉 =M∑m=1

ϕobs(xm)∆A−1(x,xm)

s2m, (38)

subject to the weighted unbiasedness condition∑m ϕ(xm)/s2m =

∑m ϕobs(xm)/s2m, and the variance in the

source has the solution ⟨ρ(x)− 〈ρ(x)〉 , ρ(x′)−

⟨ρ(x′)

⟩ ⟩= ∆∆x′A−1(x,x′). (39)

The inverse operator A−1 is solved in the same way as in the previous section, yielding for the fundamental solutionG satisfying P (−∆)∆2G(x) = δ(x),

A−1(x,x′) = G(x,x′)−Gs(x) (I +Λ)−1 G(x′), (40)

where G,Gs and Λ are defined as they are in Eq. 31.

8

Page 9: Journal of Statistical Physics manuscript No

(a) ρ(x)

0.0 0.2 0.4 0.6 0.8 1.00.0

0.2

0.4

0.6

0.8

1.0

0.0

1.5

3.0

4.5

6.0

7.5

9.0

(b) ϕ(x)

0.0 0.2 0.4 0.6 0.8 1.00.0

0.2

0.4

0.6

0.8

1.0

−0.225

−0.200

−0.175

−0.150

−0.125

−0.100

−0.075

−0.050

(c) ϕobs(x)

0.0 0.2 0.4 0.6 0.8 1.00.0

0.2

0.4

0.6

0.8

1.0

−0.225

−0.200

−0.175

−0.150

−0.125

−0.100

−0.075

−0.050

0.0 0.2 0.4 0.6 0.8 1.00.0

0.2

0.4

0.6

0.8

1.0

0.0

1.5

3.0

4.5

6.0

7.5

9.0

(d) 〈ρ(x)〉0.0 0.2 0.4 0.6 0.8 1.0

0.0

0.2

0.4

0.6

0.8

1.0

−0.225

−0.200

−0.175

−0.150

−0.125

−0.100

−0.075

−0.050

(e) 〈ϕ(x)〉

0.0 0.2 0.4 0.6 0.8 1.00.0

0.2

0.4

0.6

0.8

1.0

0.90

1.05

1.20

1.35

1.50

1.65

1.80

1.95

(f)√〈(ρ(x)− 〈ρ(x)〉)2〉

Fig. 2: Source inversion for Poisson’s equation in isotropic medium in R2. (a) Synthetic source function thatis the object of recovery. (b) A solution to Poisson’s equation on the infinite domain corresponding to this source.(c) 125 randomly-placed observations of ϕ taken with noise. (d) Reconstruction of the source. (e) Reconstruction ofthe potential. (f) Pointwise standard error in the reconstruction of the source. Parameters used: β = 10−4, sm =10−3, γ = 102.

As an example, we recover the source function in R2 shown in Fig. 2a. This source was used along with auniform unit dielectric coefficient to find the solution for the Poisson equation that is given in Fig. 2b. Noisy samplesof the potential field were taken at 125 randomly-placed locations (depicted in Fig. 2c). For regularization, wesought solutions for ρ in the Sobolev space H2(R2). Such spaces are associated with the Bessel potential operatorP (−∆) = β(γ −∆)2. Using 125 randomly placed observations, reconstructions of both ϕ and ρ were performed.The standard error of the reconstruction is also given.

3.3 Recovery of a spatially-varying dielectric coefficient field

Finally, consider the recovery of a spatially varying dielectric coefficient ε(x) by inverting the Poisson equation

∇ · (ε∇ϕ)− ρ = 0, (41)

where ρ is now known, and ϕ is measured. This problem is more difficult than the problems in the previous sections.While Eq. 41 is bilinear in ε and ϕ, the associated inverse problem of the recovery of ε given measurements of ϕis nonlinear, since ε does not relate linearly to data in ϕ. This situation is also exacerbated by the fact that noclosed-form solution for ε as a function of ϕ exists.

Assuming that the gradient of the dielectric coefficient is spatially correlated according to the Gaussian processgiven by P (−∆), we work with the Hamiltonian

H[ϕ, ε, λ; ρ, ϕobs] =1

2

M∑m=1

∫δ(x− xm)

s2m|ϕ(x)− ϕobs(x)|2 dx− 1

2

∫ε(x)∆P (−∆)ε(x) dx

+ i

∫λ(x) [∇ · (ε∇ϕ)− ρ] dx, (42)

which yields the Euler-Lagrange equations

9

Page 10: Journal of Statistical Physics manuscript No

∇ · (ε∇ϕ)− ρ = 0, (43)

−∆P (−∆)ε−∇λ · ∇ϕ = 0, (44)M∑j=1

δ(x− xj)

s2j(ϕ(x)− ϕobs(x)) +∇ · (ε∇λ) = 0. (45)

We have assumed that ε is sufficiently regular such that∫∇ε(x)·P (−∆)∇ε(x)dx <∞, thereby imposing vanishing

boundary-conditions at |x| → ∞. The Lagrange multiplier λ satisfies the Neumann boundary conditions ∇λ = 0outside of the convex hull of the observed points. In order to recover the optimal ε, one must solve these three PDEssimultaneously. A general iterative strategy for solving this system of partial differential equations is to use Eq. 43to solve for ϕ, use Eq 44 to solve for ε, and use Eq. 45 to solve for λ. Given λ and ϕ, the left-hand-side of Eq 44provides the gradient of the Hamiltonian with respect to ε which can be used for gradient descent. Eqs. 43 and 45are simply the Poisson equation.

For quantifying error in the mean-field recovery, we seek a formulation of the problem of recovering ε using thepath integral method. We are interested in the generating functional Z[J ] =

∫∫∫DϕDεDλ exp (−H[ϕ, ε, λ] +

∫Jε dx) .

Integrating in λ and ϕ, yields the marginalized generating functional

Z[J ] =

∫Dε exp

−H[ε; ρ, ϕobs] +

∫J(x)ε(x) dx

=

∫Dε exp

1

2

M∑m=1

∫δ(x− xm)

s2m[ϕ(ε(x))− ϕobs(x)]

2 dx

−1

2

∫ε(x)(−∆)P (−∆)ε(x) dx+

∫J(x)ε(x) dx

. (46)

To approximate this integral, one needs an expression for the ϕ as a function of ε. To find such an expression,one can use the product rule to write Poisson’s equation as ε∆ϕ + ∇ε · ∇ϕ = ρ. Assuming that ∇ε is small, onemay solve Poisson’s equation in expansion of powers of ∇ε by using the Green’s function L(x,x′) of the Laplacian

operator to write ϕ(x) =∫L(x,x′)ρ(x

′)ε(x′) dx′−

∫L(x,x′)∇x′ log ε(x′) ·∇x′ϕ(x′)dx′, which is a Fredholm integral

equation of the second kind. The function ϕ then has the Liouville-Neumann series solution

ϕ(x) =∞∑n=0

ϕn(x) (47)

ϕn(x) =

∫K(x,y)ϕn−1(y) dy n ≥ 1 (48)

ϕ0(x) =

∫L(x,y)

ρ(y)

ε(y)dy (49)

K(x,y) = ∇y ·[L(x,y)∇y log ε(y)

], (50)

where ∇ε is assumed to vanish at the boundary of reconstruction. Taken to two terms in the expansion of ϕ(ε) givenin Eqs. 47-50, the second-order term in the Taylor expansion of Eq. 46 is of the form (see Appendix A)

δ2H

δε(x)δε(x′)∼ −∆P (−∆)δ(x− x′) +

M∑m=1

am(x,x′).

This expression, evaluated at the solution of the Euler-Lagrange equations ε?, ϕ?, provides an an approximation ofthe original probability density from which the posterior variance

⟨ε(x)− ε?(x), ε(x′)− ε?(x′)

⟩= A−1(x,x′) can

be estimated. To find this inverse operator, we discretize spatially and compute the matrix A−1ij = A−1(xi,xj),

A−1 = (I + GA−1m )−1G,

where I is the identity matrix, G is a matrix of values [(−∆)P (−∆)δ(x,x′)]−1, A−1m =

[δx∑m am(x,x′)

]−1,

and δx is the volume of a lattice coordinate.As an example, we present the recovery of a dielectric coefficient in R1 over the compact interval x ∈ [0, 1] of

a dielectric coefficient shown in Fig. 3a given a known source function (10 × 1x∈[0,1]). A solution to the Poissonequation given Eq. 41 is shown in Fig. 3b. For regularization, we use the operator P (−∆) = β(γ −∆), and assume

10

Page 11: Journal of Statistical Physics manuscript No

(a) ε(x), ε?(x)

0.0 0.2 0.4 0.6 0.8 1.00.92

0.93

0.94

0.95

0.96

0.97

0.98

0.99

1.00

1.01

(b) ϕ(x), ϕobs(x), ϕ?(x)

0.0 0.2 0.4 0.6 0.8 1.0−1

0

1

2

3

4

5

6

(c)√〈(ε(x)− ε?(x))2〉approx

0.0 0.2 0.4 0.6 0.8 1.00.00

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

0.0 0.2 0.4 0.6 0.8 1.00.0

0.2

0.4

0.6

0.8

1.0

0.0000

0.0008

0.0016

0.0024

0.0032

0.0040

0.0048

0.0056

(d) 〈(δε(x), δε(x′)〉approx

0.0 0.2 0.4 0.6 0.8 1.00.0

0.2

0.4

0.6

0.8

1.0

0.00000

0.00002

0.00004

0.00006

0.00008

0.00010

0.00012

0.00014

(e) 〈(δε(x), δε(x′)〉MC

0.0 0.2 0.4 0.6 0.8 1.00.000

0.002

0.004

0.006

0.008

0.010

0.012

0.014

(f)√〈(ε(x)− ε?(x))2〉MC

Fig. 3: Dielectric inversion for Poisson’s equation in R1. (a) (blue) Spatially-varying dielectric coefficient ε that isthe object of recovery, and the mean field recovery ε? (green). (b) A solution ϕ (blue) to Poisson’s equation given aknown step-source supported on [0, 1] and the spatially-varying dielectric coefficient, 50 randomly placed samplesof the solution taken with error, and the mean field recovery of the potential function ϕ? (green). (c) Standard errorin the mean-field recovery of the dielectric field. (d) Approximate posterior variance in the recovery of the dielectricfield. (δε = ε − ε?). (e) Monte-Carlo corrected covariance field estimate (f) Monte-Carlo corrected point-wise errorestimate. Parameters used: sm = 0.2, β = 2.5, γ = 100.

that ∇ε → 0 at the boundaries of the recovery, which are outside of the locations where measurements are taken.For this reason, we take the Green’s function G of the differential operator − d2

dx2P (− d2

dx2 ) = − d2

dx2 β(γ − d2

dx2 ) tovanish along with its first two derivatives at the boundary of recovery.

The point-wise standard error and the posterior covariance are shown in Figs. 3c and 3d, respectively. Monte-Carlo corrected estimates are also shown. Note that approximate point-wise errors are much larger than the Monte-Carlo point-wise errors. This fact is due in-part to inaccuracy in using the series solution for the Poisson equationgiven in Eq 47, which relies on ∇ε to be small. While the approximate errors were inaccurate, the approximationwas still useful in providing a sampling density for use in importance sampling.

4 Discussion

In this paper we have presented a general method for regularizing ill-posed inverse problems based on the Bayesianinterpretation of Tikhonov regularization, which we investigated through the use of field-theoretic approaches. Wedemonstrated the approach by considering two linear problems – interpolation (Sec. 3.1) and source inversion(Sec. 3.2), and a non-linear problem – dielectric inversion (Sec. 3.3). For linear problems Tikhonov regularizationyields Gaussian functional integrals, where the moments are available in closed-form. For non-linear problems, wedemonstrated a perturbative technique based on functional Taylor series expansions, for approximate local densityestimation near the maximum a-posteriori solution of the inverse problem. We also discussed how such approxima-tions can be improved based on Monte-Carlo sampling (Sec. 2.3).

Our first example problem was that of membrane or plate interpolation. In this problem the regularizationterm is known based on a priori knowledge of the physics of membranes with bending rigidity. The Helfrich freeenergy describes the thermal fluctuations that are expected of rigid membranes, and provided us with the differentialoperator to use for Tikhonov regularization. Using the path integral, we were able to calculate an analytical expressionfor the error in the reconstruction of the membrane surface. It is apparent that the error in the recovery depends onboth the error of the measurements and the distance to the nearest measurements. Surprisingly, the reconstructionerror did not explicitly depend on the misfit error.

11

Page 12: Journal of Statistical Physics manuscript No

The second example problem was the reconstruction of the source term in the Poisson equation given measure-ments of the field. In this problem, the regularization is not known from physical constraints and we demonstratedthe use of a regularizer chosen from a general family of regularizers. This type of regularization is equivalent to thenotion of weak solutions in Sobolev spaces. Since the source inversion problem is linear, we were able to analyticallycalculate the solution as well as the error of the solution. Again, the reconstruction error did not explicitly dependon the misfit error.

The last example problem was the inversion of the dielectric coefficient of Poisson’s equation from potential mea-surements. This problem was nonlinear, yielding non-Gaussian path-integrals. We used this problem to demonstratethe technique of semiclassical approximation for use in Bayesian inverse problems.

The reliability of the semiclassical approximation depends on how rapidly the posterior distribution falls off fromthe extremum or mean field solution. Applying the semiclassical approximation to the Information Hamiltonian(Eq 12), one sees that the regularization only contributes to terms up to second order. Higher-order terms in theexpansion rely only on the likelihood term in the Hamiltonian. Since the data error is assumed to be normallydistributed with variance s2m, one expects each squared residual (ϕ(xm) − ϕobs(xm))2 to be O(s2m). For thisreason, each observation contributes a term of O(1) to the Hamiltonian. As a result, there is an implicit largeprefactor of O(M) in the Hamiltonian, where M is defined as before as the number of observations. The first ordercorrection to the semiclassical method is then expected to be O(1/M).

4.1 Future directions

By putting inverse problems into a Bayesian framework, one gains access to a large toolbox of methodology thatcan be used to construct and verify models. In particular, Bayesian model comparison [33] methods can be used foridentifying the regularization terms to be used when one does not have prior information available about the solution.Such methods can also be used when one has some knowledge of the object of recovery, modulo the knowledge ofsome parameters. For instance, one may seek to recover the height of a plate or membrane but not know the surfacetension or elasticity. Then, Bayesian methods can be used to recover probability distributions for the regularizationparameters along with the object of recovery.

Finally, Tikhonov regularization works naturally in the path integral framework because it involves quadratic pe-nalization terms which yield Gaussian path integrals. It would be interesting to examine other forms of regularizationover function spaces within the path integral formulation, such as L1 regularization.

5 Acknowledgements

This material is based upon work supported by the National Science Foundation under Agreement No. 0635561. JCand TC also acknowledge support from the National Science Foundation through grant DMS-1021818, and from theArmy Research Office through grant 58386MA. VS acknowledges support from UCLA startup funds.

A Functional Taylor approximations for the dielectric field problem

We wish to expand the Hamiltonian

H[ε; ρ, ϕobs] =1

2

M∑m=1

∫δ(x− xm)

s2m

[ ∞∑n=0

ϕn(ε(x))− ϕobs(x)

]2dx+

1

2

∫ε(x)(−∆)P (−∆)ε(x) dx (51)

about its extrema ε∗. We take variations with respect to ε(x) to calculate its first functional derivative,∫∂H

∂ε(x)φ(x) dx =

∫(−∆)P (−∆)ε(x)φ(x) dx+ lim

h→0

d

dh

1

2

M∑m=1

∫δ(x− xm)

s2m

[ ∞∑n=0

ϕn(ε(x) + hφ(x))− ϕobs(x)

]2dx

=

∫(−∆)P (−∆)εφ dx+ lim

h→0

M∑m=1

∫δ(x− xm)

s2m(ϕ(x)− ϕobs(x))

d

dhϕ0(ε(x) + hφ(x)) dx

+ limh→0

M∑m=1

∫δ(x− xm)

s2m(ϕ(x)− ϕobs(x))

d

dhϕ1(ε(x) + hφ(x)) dx

+ limh→0

M∑m=1

∫δ(x− xm)

s2m(ϕ(x)− ϕobs(x))

∞∑n=2

d

dhϕn(ε(x) + hφ(x)) dx︸ ︷︷ ︸

I1

(52)

12

Page 13: Journal of Statistical Physics manuscript No

Let us define the quantities

K(y, z) = ∇z ·[L(y, z)∇z

(φ(z)

ε(z)

)]ϕ0(x) = −

∫L(x,y)

ρ(y)φ(y)

ε2(y)dy

Ψ(x) =

M∑m=1

δ(x− xm)

s2m(ϕ(x)− ϕobs(x)) .

Through direct differentiation we find that

I1 =

∞∑n=2

∫Ψ(x)K(x,yn)

n−1∏j=1

K(yj+1,yj)

ϕ0(y1) dxn∏k=1

dyk

+

∞∑n=2

∫Ψ(x)K(x,yn)

n−1∏j=1

K(yj+1,yj)

ϕ0(y1) dxn∏k=1

dyk

+

∞∑n=2

∫Ψ(x)K(x,yn)

n−1∑k=0

K(yk+1,yk)

n−1∏j=1j 6=k

K(yj+1,yj)

ϕ0(y1) dxn∏k=1

dyk.

Integrating in x:

I1 =∞∑n=2

M∑m=1

ϕ(xm)− ϕobs(xm)

s2m

∫K(xm,yn)

n−1∏j=1

K(yj+1,yj)

ϕ0(y1)n∏k=1

dyk

+

∞∑n=2

M∑m=1

ϕ(xm)− ϕobs(xm)

s2m

∫K(xm,yn)

n−1∏j=1

K(yj+1,yj)

ϕ0(y1)n∏k=1

dyk

+∞∑n=2

M∑m=1

ϕ(xm)− ϕobs(xm)

s2m

∫K(xm,yn)

n−1∑k=1

K(yk+1,yk)

n−1∏j=1j 6=k

K(yj+1,yj)

ϕ0(y1)n∏k=1

dyk.

We shift φ(·)→ φ(x), and integrate-by-parts to find

I1 = −∞∑n=2

M∑m=1

ϕ(xm)− ϕobs(xm)

s2m

∫K(xm,yn)

n−1∏j=1

K(yj+1,yj)

L(y1,x)ρ(x)

ε2(x)φ(x) dx

n∏k=1

dyk

+∞∑n=2

M∑m=1

ϕ(xm)− ϕobs(xm)

s2m

∫φ(x)

ε(x)∇ · [L(xm,x)∇ (K(x,yn−1))]

n−2∏j=1

K(yj+1,yj)

ϕ0(y1) dxn∏k=1

dyk

+

∞∑n=2

M∑m=1

ϕ(xm)− ϕobs(xm)

s2m

∫K(xm,yn)

n−1∑k=1

φ(x)

ε(x)∇ · [L(yk+1,x)∇K(x,yk−1)]

n−2∏j=1j 6=k

K(yj+1,yj)

ϕ0(y1) dxn∏k=1

dyk.

Note that all boundary terms disappear since we can take φ to disappear on the boundary. With I1 computed, we find

δH

δε(x)= (−∆)P (−∆)ε(x)−

M∑m=1

ϕ(xm)− ϕobs(xm)

s2m

[L(xm,x)

ρ(x)

ε2(x)

]

+M∑m=1

ϕ(xm)− ϕobs(xm)

s2mε(x)∇ · [L(xm,x)∇ϕ0(x)]−

M∑m=1

ϕ(xm)− ϕobs(xm)

s2m

(ρ(x)

ε2(x)

)∫K(xm,y1)L(x,y1) dy1

−∞∑n=2

M∑m=1

ϕ(xm)− ϕobs(xm)

s2m

∫K(xm,yn)

n−1∏j=1

K(yj+1,yj)

L(y1,x)ρ(x)

ε2(x)

n∏k=1

dyk

+

∞∑n=2

M∑m=1

ϕ(xm)− ϕobs(xm)

s2mε(x)

∫∇ · [L(xm,x)∇ (K(x,yn−1))]

n−2∏j=1

K(yj+1,yj)

ϕ0(y1)

n∏k=1

dyk

+

∞∑n=2

M∑m=1

ϕ(xm)− ϕobs(xm)

s2mε(x)

∫K(xm,yn)

n−1∑k=1

∇ · [L(yk+1,x)∇K(x,yk−1)]

n−2∏j=1j 6=k

K(yj+1,yj)

ϕ0(y1)

n∏k=1

dyk.

(53)

13

Page 14: Journal of Statistical Physics manuscript No

Taken to two terms in the series expansion for ϕ, the first variation is

δH

δε(x)∼ (−∆)P (−∆)ε(x) +

M∑m=1

ϕ(xm)− ϕobs(xm)

s2mε(x)

[∇L(x,xm) · ∇ϕ0(x)−

ρ(x)

ε(x)

∫K(xm,y1)L(x,y1) dy1

]. (54)

To calculate the second-order term in the Taylor-expansion, we take another variation. Truncated at two terms in the expansion forϕ:

δ2H

δε(x)δε(x′)= (−∆)P (−∆)δ(x− x′) +

M∑m=1

am(x,x′), (55)

where after canceling like terms,

am(x,x′) = δ(x− x′)ϕ(xm)− ϕobs(xm)

s2mε2(x′)

[2ρ(x′)

ε(x′)

∫K(xm,y1)L(x

′,y1) dy1 −∇x′L(x′,xm) · ∇x′ϕ0(x′)− L(xm,x′)

ρ(x′)

ε(x′)

]−ϕ(xm)− ϕobs(xm)

s2m

∇L(x,xm) · ∇x′L(x,x′)

ρ(x′)

ε(x)ε2(x′)+

ρ(x)

ε2(x)ε(x′)∇L(x,x′) · ∇x′L(xm,x

′)

+

[∇L(x,xm) · ∇ϕ0(x)−

ρ(x)

ε(x)

∫K(xm,y1)L(x,y1) dy1

1

s2mε(x)ε(x′)

[∇x′L(x′,xm) · ∇x′ϕ0(x

′)−ρ(x′)

ε(x′)

∫K(xm,y1)L(x

′,y1) dy1

].

It is using this expression that we can construct an approximate probability density for our field ε.

References

1. Alves C, Colaço M, Leitão V, Martins N, Orlande H, Roberty N (2008) Recovering the source term in a linear diffusion problem bythe method of fundamental solutions. Inverse Problems in Science and Engineering 16(8):1005–1021

2. Anzengruber SW, Ramlau R (2010) Morozov’s discrepancy principle for tikhonov-type functionals with nonlinear operators. InverseProblems 26(2):025,001

3. Bertero M, De Mol C, Viano G (1980) The stability of inverse problems. In: Inverse scattering problems in optics, Springer, pp 161–2144. Bui-Thanh T, Ghattas O, Martin J, Stadler G (2013) A computational framework for infinite-dimensional bayesian inverse problems

part i: The linearized case, with application to global seismic inversion. SIAM Journal on Scientific Computing 35(6):A2494–A25235. Chow CC, Buice MA (2010) Path integral methods for stochastic differential equations. arXiv preprint arXiv:100959666. Cotter S, Dashti M, Robinson J, Stuart A (2009) Bayesian inverse problems for functions and applications to fluid mechanics. Inverse

Problems 25:115,0087. Dashti M, Law KJ, Stuart AM, Voss J (2013) Map estimators and their consistency in bayesian nonparametric inverse problems. Inverse

Problems 29(9):095,0178. Engl H, Kunisch K, Neubauer A (1999) Convergence rates for Tikhonov regularisation of non-linear ill-posed problems. Inverse

problems 5(4):5239. Engl H, Flamm C, Kügler P, Lu J, Müller S, Schuster P (2009) Inverse problems in systems biology. Inverse Problems 25(12):123,01410. Engl HW, Kunisch K, Neubauer A (1989) Convergence rates for tikhonov regularisation of non-linear ill-posed problems. Inverse

problems 5(4):52311. Enßlin TA, Frommert M, Kitaura FS (2009) Information field theory for cosmological perturbation reconstruction and nonlinear signal

analysis. Physical Review D 80(10):105,00512. Evans A, Turner M, Sens P (2003) Interactions between proteins bound to biomembranes. Physical Review E 67(4):041,90713. Farmer C (2007) Bayesian field theory applied to scattered data interpolation and inverse problems. Algorithms for Approximation pp

147–16614. Feynman RP, Hibbs AR (2012) Quantum mechanics and path integrals: Emended edition. DoverPublications. com15. Graham R (1977) Path integral formulation of general diffusion processes. Zeitschrift für Physik B Condensed Matter 26(3):281–29016. Hänggi P (1989) Path integral solutions for non-markovian processes. Zeitschrift für Physik B Condensed Matter 75(2):275–28117. Heller EJ (1981) Frozen gaussians: A very simple semiclassical approximation. The Journal of Chemical Physics 75:292318. Heuett WJ, Miller III BV, Racette SB, Holloszy JO, Chow CC, Periwal V (2012) Bayesian functional integral method for inferring

continuous data from discrete measurements. Biophysical journal 102(3):399–40619. Hoang VH, Law KJ, Stuart AM (2013) Determining white noise forcing from eulerian observations in the navier stokes equation. arXiv

preprint arXiv:1303467720. Hohage T, Pricop M (2008) Nonlinear Tikhonov regularization in Hilbert scales for inverse boundary value problems with random

noise. Inverse Problems and Imaging 2:271–29021. Hon Y, Li M, Melnikov Y (2010) Inverse source identification by Green’s function. Engineering Analysis with Boundary Elements

34(4):352–35822. Hörmander L (2007) The analysis of linear partial differential operators III: pseudo-differential operators, vol 274. Springer23. Itô K (1961) Wiener integral and feynman integral. In: Proceedings of the 4th Berke ley Symposium on Mathematical Statistics and

Probability, vol 2, pp 227–23824. Kardar M (2007) Statistical physics of fields. Cambridge University Press25. Lasanen S (2007) Measurements and infinite-dimensional statistical inverse theory. PAMM 7(1):1080,101–1080,10226. Lasanen S (2012) Non-gaussian statistical inverse problems. part i: Posterior distributions. Inverse Problems & Imaging 6(2)27. Lasanen S (2012) Non-gaussian statistical inverse problems. part ii: Posterior convergence for approximated unknowns. Inverse Prob-

lems & Imaging 6(2)

14

Page 15: Journal of Statistical Physics manuscript No

28. Lemm JC (1999) Bayesian field theory: Nonparametric approaches to density estimation, regression, classification, and inverse quantumproblems. arXiv preprint physics/9912005

29. Lieberman C, Willcox K, Ghattas O (2010) Parameter and state model reduction for large-scale statistical inverse problems. SIAMJournal on Scientific Computing 32(5):2523–2542

30. Liu JS (2008) Monte Carlo strategies in scientific computing. springer31. Martin J, Wilcox LC, Burstedde C, Ghattas O (2012) A stochastic newton mcmc method for large-scale statistical inverse problems

with application to seismic inversion. SIAM Journal on Scientific Computing 34(3):A1460–A148732. Neubauer A (1999) Tikhonov regularisation for non-linear ill-posed problems: optimal convergence rates and finite-dimensional

approximation. Inverse problems 5(4):54133. O’Hagan A, Forster J, Kendall MG (2004) Bayesian inference. Arnold London34. Oppermann N, Robbers G, Enßlin TA (2011) Reconstructing signals from noisy data with unknown signal and noise covariance.

Physical Review E 84(4):041,11835. Peliti L (1985) Path integral approach to birth-death processes on a lattice. Journal de Physique 46(9):1469–148336. Pesquera L, Rodriguez M, Santos E (1983) Path integrals for non-markovian processes. Physics Letters A 94(6):287–28937. Petra N, Martin J, Stadler G, Ghattas O (2013) A computational framework for infinite-dimensional bayesian inverse problems: Part ii.

stochastic newton mcmc with application to ice sheet flow inverse problems. arXiv preprint arXiv:1308622138. Potsepaev R, Farmer C (2010) Application of stochastic partial differential equations to reservoir property modelling. In: 12th European

Conference on the Mathematics of Oil Recovery39. Quinn JC, Abarbanel HD (2010) State and parameter estimation using monte carlo evaluation of path integrals. Quarterly Journal of

the Royal Meteorological Society 136(652):1855–186740. Quinn JC, Abarbanel HD (2011) Data assimilation using a gpu accelerated path integral monte carlo approach. Journal of Computa-

tional Physics 230(22):8168–817841. Scherzer O (1993) The use of morozov’s discrepancy principle for tikhonov regularization for solving nonlinear ill-posed problems.

Computing 51(1):45–6042. Schwab C, Stuart AM (2012) Sparse deterministic approximation of bayesian inverse problems. Inverse Problems 28(4):045,00343. Stuart A (2010) Inverse problems: a Bayesian perspective. Acta Numerica 19(1):451–55944. Tikhonov AN (1943) On the stability of inverse problems. In: Dokl. Akad. Nauk SSSR, vol 39, pp 195–19845. Zee A (2005) Quantum field theory in a nutshell. Universities Press

15