22
1 Quantitative Analysis for Image-Based Rendering with Depth Information: Part II–2D Occluded Scenes and 3D Scenes Ha T. Nguyen and Minh N. Do Abstract We quantitatively analyze the rendering quality of image-based rendering (IBR) algorithms using per-pixel depth. Assuming the ideal pinhole camera model, we extend the error aggregation framework proposed in a companion paper to 2D occluded scenes and 3D unoccluded scenes. For 2D occluded scenes, we measure the effects of jumps in sample intervals around the discontinuities of the virtual image, resulting in additional terms. For 3D scenes, we derive an error bound for triangulation-based linear interpolation and exploit properties of Poisson Delaunay triangles. We show that the mean absolute errors (MAE) of the virtual images can be bounded based on the characteristics of the scene and the camera configuration. An intriguing finding is that triangulation-based linear interpolation for 3D scenes results in a decay order O(λ -1 ) of the MAE in smooth regions, where λ is the local density of actual samples, compared to O(λ -2 ) for 2D scenes. Index Terms Image-Based Rendering, Nonuniform Interpolation, Poisson Delaunay Triangulation, Jitter, Error Bound. I. I NTRODUCTION Although many algorithms and systems have been proposed for image-based rendering (IBR) appli- cations [15], [16], little research has been addressing the fundamental issue of analyzing the effects Ha T. Nguyen is with the Department of Electrical and Computer Engineering and the Coordinated Science Laboratory, University of Illinois, Urbana IL 61801 (email: [email protected]). Minh N. Do is with the Department of Electrical and Computer Engineering, the Coordinated Science Laboratory, and the Beckman Institute, University of Illinois, Urbana IL 61801 (email: [email protected]). This work was supported by the National Science Foundation under Grant ITR-0312432. September 28, 2007 DRAFT

1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

1

Quantitative Analysis for Image-Based

Rendering with Depth Information: Part II–2D

Occluded Scenes and 3D ScenesHa T. Nguyen and Minh N. Do

Abstract

We quantitatively analyze the rendering quality of image-based rendering (IBR) algorithms using

per-pixel depth. Assuming the ideal pinhole camera model, we extend theerror aggregation framework

proposed in a companion paper to 2D occluded scenes and 3D unoccluded scenes. For 2D occluded

scenes, we measure the effects of jumps in sample intervals around the discontinuities of the virtual

image, resulting in additional terms. For 3D scenes, we derive an error bound for triangulation-based

linear interpolation and exploit properties of Poisson Delaunay triangles. We show that the mean absolute

errors (MAE) of the virtual images can be bounded based on thecharacteristics of the scene and the

camera configuration. An intriguing finding is that triangulation-based linear interpolation for 3D scenes

results in a decay orderO(λ−1) of the MAE in smooth regions, whereλ is the local density of actual

samples, compared toO(λ−2) for 2D scenes.

Index Terms

Image-Based Rendering, Nonuniform Interpolation, Poisson Delaunay Triangulation, Jitter, Error

Bound.

I. I NTRODUCTION

Although many algorithms and systems have been proposed forimage-based rendering(IBR) appli-

cations [15], [16], little research has been addressing thefundamental issue of analyzing the effects

Ha T. Nguyen is with the Department of Electrical and Computer Engineering and the Coordinated Science Laboratory,

University of Illinois, Urbana IL 61801 (email: [email protected]).

Minh N. Do is with the Department of Electrical and Computer Engineering, the Coordinated Science Laboratory, and the

Beckman Institute, University of Illinois, Urbana IL 61801 (email: [email protected]).

This work was supported by the National Science Foundation under Grant ITR-0312432.

September 28, 2007 DRAFT

Page 2: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

2

of the scene and the camera setting on the rendering quality.Understanding these effects is crucial to

controlling the rendering quality and the cost of IBR systems. Many IBR algorithms in practice have to

rely on oversampling to counter undesirable aliasing effects.

In a companion paper [10], we quantitatively analyzed, for 2D unoccluded scenes, the quality of IBR

texture mapping algorithms using explicit depth maps. We proposed anerror aggregation frameworkto

bound rendering errors based on the sample values, sample positions, and their errors, whether the virtual

image is interpolated in image-space or object-space. The union of sample positions is approximated as a

generalized Poisson process, while the sample jitters are bounded based on the relative position between

the virtual camera and the scene. We derived error bounds forseveral IBR algorithms using per-pixel

depth. The derived error bounds show the effects on the rendering quality of various factors including

depth and intensity estimate errors, the scene geometry andtexture, the number of actual cameras,

their positions and resolutions. Implications of the proposed analysis include camera placement, budget

allocation, and bit allocation.

In this paper, we extend the analysis in [10] to 2D occluded scenes and 3D unoccluded scenes. The

main contribution of the paper is a methodology armed with a set of techniques to analyze the rendering

quality of IBR algorithms, assuming per-pixel depth as inputs, using image-space interpolation. To analyze

2D occluded scenes, we measure, in Proposition 1, the effectsof jumps in sample intervals around the

discontinuities of the virtual image, resulting in additional terms in the error bound. We extend the analysis

to 3D unoccluded scenes by proposing novel machineries, including an error bound for triangulation-

based linear interpolation and the use of Poisson Delaunay triangles’ properties–classical results from

stochastic geometry [11]. We find that, in smooth regions, triangulation-based linear interpolation for 3D

scenes results in a decay orderO(λ−1) of the mean absolute error (MAE), whereλ is the local density

of actual samples, compared toO(λ−2) for 2D scenes. This intriguing finding implies that for 3D scenes,

building IBR systems that can simplify to 2D, such as adopting image rectifications and planar camera

configurations, besides decreasing the complexity, also increases the decay order of the rendering errors

in smooth regions.

This paper is organized as follows. The problem setup is presented in Section II. We present analysis

of 2D occluded scenes in Section III. Generalization to 3D is given in Section IV. Finally, we offer

concluding remarks in Section V.

DRAFT September 28, 2007

Page 3: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

3

II. PROBLEM SETUP

We start with a description of the scene model in Section II-A.The camera model is presented in

Section II-B. We describe our models for the 3D case that are parallel with 2D models considered in the

companion paper [10]. This description also introduces the notation used in the paper. Finally, we state

the problem in Section II-C.

A. The scene model

The surface of the scene is modeled as a 3D parameterized surfaceS(u, v) : Ω → R3, for some region

Ω ⊂ R2. The texture mapT (u, v) : Ω → R is an intensity function “painted” on the surfaceS(u, v). We

assume that the surface is Lambertian [2], that is, images of the same surface point at different cameras

have the same intensity. Furthermore, we assume that the surfaceS(u, v) and the textureT (u, v) have

derivative of second order at all points and all directions,except at the discontinuities.

B. The camera model

X

Y

Z

∆x∆y x

y

p = HΠ(u, v)

S(u, v)

(u, v) ∈ Ω

Fig. 1. The 3D calibrated scene-camera model. The scene surface is modeled as a 3D parameterized surfaceS(u, v) for

(u, v) ∈ Ω ⊂ R2. The textureT (u, v) is “painted” on the surface. We assume pinhole camera model with calibrated positional

matrix Π ∈ R3×4. The camera resolution is characterized by the pixel intervals∆x, ∆y in horizontal and vertical direction on

the image plane.

A 3D pinhole camera (Fig. 1) is characterized by the positional matrix Π = [π1, π2, π3]T ∈ R

3×4. For

each surface pointS = [X, Y, Z]T in the scene, let its homogeneous coordinate [2] beS = [X, Y, Z, 1]T .

The projection equation is

d · [x, y, 1]T = Π · [X, Y, Z, 1]T , (1)

September 28, 2007 DRAFT

Page 4: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

4

whered = πT3 · S is the depth ofS relative toΠ. We derive ascene-to-image mappingHΠ(u, v) from

surface pointsS(u, v), for (u, v) ∈ Ω, to their image points(x, y) as x

y

def

= HΠ(u, v) =

Hx(u, v)

Hy(u, v)

, (2)

where

Hx(u, v) =πT

1 · S(u, v)

πT3 · S(u, v)

, Hy(u, v) =πT

2 · S(u, v)

πT3 · S(u, v)

. (3)

The Jacobian matrix ofHΠ is

∂HΠ(u, v)

∂(u, v)=

∂Hx/∂u ∂Hx/∂v

∂Hy/∂u ∂Hy/∂v

. (4)

At the image plane of a cameraΠ, the image light fieldfΠ(x, y) at image point(x, y) characterizes

the “brightness”T (u, v) of surface pointS(u, v) having image at(x, y). In other words, the image light

field fΠ(x, y) is perspectively corrected from the texture mapT (u, v) as

fΠ(x, y) = T(H−1

Π(x, y)

). (5)

Let ∆x, ∆y be the sample intervals in horizontal and vertical directions of the discrete grid on which

the actual images are sampled from the image light field. We refer the product∆x∆y to the resolution of

the camera. Ifϕ(x, y) is the sampling kernel of the cameraΠ, the pixel intensityIΠ[m, n] is the value

of the convolution offΠ(x, y) andϕ(x, y), evaluated at(xm, yn) = (m∆x, n∆y), as follows:

IΠ[m, n]= (fΠ ∗ ϕ)(xm, yn) (6)

=

HΠ(Ω)fΠ(x, y) · ϕ(xm − x, yn − y)dxdy. (7)

In this paper, we assume the ideal pinhole camera model with the Dirac delta function as the sampling

kernelϕ(x, y), i.e., ϕ(x, y) = δ(x, y). In other words,

IΠ[m, n] = fΠ(m∆x, n∆y). (8)

Depth and intensity estimate error. In practice, the depth and the intensity at actual pixels are

subjected to errorsεD = [Xe − X, Ye − Y, Ze − Z]T and εT = Te(u, v) − T (u, v), respectively. We

suppose thatεD andεT are bounded byED andET , that is,

‖εD‖2 ≤ ED, |εT | ≤ ET . (9)

BoundsET , ED can be related to the signal to noise ratio of the actual images and depth maps.

DRAFT September 28, 2007

Page 5: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

5

C. Problem statement

IBR algorithms. Many IBR algorithms have been proposed [5], [15], [16]. This paper is concerned

with IBR algorithms using per-pixel depth and image-space interpolation [6], [7], [8], [14]. We present

our analysis for the Propagation Algorithm [8], although theproposed techniques are applicable to other

algorithms.

We assume that piecewise linear interpolation is used for the 2D case and Delaunay triangulation-based

linear interpolation is used for the 3D case. Both methods are widely used in practice thanks to their

simplicity and decent interpolation qualities. Furthermore, we hope to help IBR practitioners find the

paper directly useful. We note that the proposed analysis also applies for interpolation techniques using

higher order splines [1], [13].

Problem statement.Suppose the virtual image at virtual cameraΠv is rendered using images and

depth maps ofN actual camerasΠiNi=1. We want to quantify the effects on the rendering quality of

projection matricesΠiNi=1 andΠv, the resolution∆x∆y, the depth and intensity estimate error bound

ED, ET , the texture mapT (u, v), and the surface geometryS(u, v).

III. A NALYSIS FOR 2D SCENES

In this section, we extend the analysis proposed in a companion paper [10] to 2D occluded scenes.

We present the new methodology in Section III-A and revisit relevant results of [10] in Section III-B. In

Section III-C, we analyze the rendering quality for 2D occluded scenes.

A. Methodology

We extend the methodology proposed in [10] to consider discontinuous functions. The linear interpo-

lation f(x) of f(x) in an interval[x1, x2] is defined as

f(x) =x2 − x

x2 − x1· f(x1) +

x − x1

x2 − x1· f(x2). (10)

In the presence of sample errorsε1, ε2 and sample jittersµ1, µ2 (see Fig. 2), the sample valuesf(x1), f(x2)

in (10) are replaced byf(x1 + µ1) + ε1, f(x2 + µ2) + ε2, respectively. TheL∞ norm of a functiong(x)

is defined as

‖g‖∞ = supx

g(x). (11)

In the following, we bound the linear interpolation error for functions with only one discontinuity.

Note that general analysis is possible, although less elegant, and provides similar conclusions. First, for

September 28, 2007 DRAFT

Page 6: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

6

actual samplesmeasured samples

εa

εb

f(x−d )

f(x+d )

xdx1 x2

µaµb

f(x)

f(x)

Fig. 2. Linear interpolation of a discontinuous function.

simplicity we introduced notations:

∆1 = xd − x1, ∆2 = x2 − xd, ∆ = x2 − x1. (12)

Proposition 1: Consider a functionf(x) that is twice continuously differentiable except atthediscon-

tinuity xd. The aggregated error over[x1, x2] of the linear interpolation given in (10), defined below, can

be bounded by∫ x2

x1

|f(x) − f(x)| ≤1

8∆3 · ‖f ′′‖∞ +

1

2∆1∆2 · |J1| +

3

2∆ · |J0|

+∆(

maxi=1,2

|εi| + maxi=1,2

|µi| · ‖f′‖∞

), (13)

whereJ0, J1 are the jumps off(x) and its derivative at the discontinuityxd:

J0 = f(x+d ) − f(x−

d ), J1 = f ′(x+d ) − f ′(x−

d ). (14)

Proof: See Appendix A.

Remark 1:The bound in Proposition 1 is proposed for the aggregated errorover the interval[x1, x2].

This is different from the pointwise bound given in [10, Proposition 1]. Proposition 1 can be considered

as a local analysis, providing a bound for the interpolationerror in individual intervals. Because of the

discontinuity atxd, the aggregated error increases by an amount of

1

2∆1∆2 · |J1| +

3

2∆ · |J0|.

If J0 = J1 = 0, the bound of Proposition 1 simplifies to the case wheref(x) is twice continuously

differentiable in[x1, x2] (see [10, Proposition 1]). The bound is characterized by sample intervals, sample

DRAFT September 28, 2007

Page 7: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

7

errors, and jitters, in addition to intrinsic properties off(x). Similar remarks can be drawn for interpolation

using splines of higher orders [1].

The bound in (13) suggests that we need to investigate the sample intervals, especially observed sample

intervals around the discontinuities, and sample jitters in the context of IBR.

B. Part I revisited – analysis for 2D unoccluded scenes

We state in this section key results of our companion paper [10] for 2D unoccluded scenes. The

presentation helps to understand previous results and the development of this paper. In Proposition 2, we

present the property of sample intervals. We give a bound forsample jitters in Proposition 3. We provide

an error bound, in Theorem 1, for the virtual images rendered using the Propagation Algorithm [8].

Properties of sample intervals.On the image plane of the virtual cameraΠv, let Y be the set of

pointspropagatedfrom actual pixels [8].

Proposition 2: The point processY can be approximated as a generalized Poisson process with density

function

λx(x) =1

∆xH ′v(u)

·N∑

i=1

H ′i(u), (15)

whereu = H−1v (x). The sum of powers of the sample intervals can be computed as

NY−1∑

n=1

(ym+1 − ym)k ≈ k! Yk ∆k−1x , (16)

where

Yk =

∫ b

a

(N∑

i=1

H ′i(u)

)1−k(H ′

v(u))k

du. (17)

Bounds for sample jitters. Let S be a surface point andSe be an erroneous estimate ofS. We

suppose thaty and y are images ofS andSe at the virtual cameraΠv.

Proposition 3: The jitter µ = y − y, at virtual cameraΠv with camera centerCv, caused by depth

estimate errors is bounded by

|µ| ≤ EDBv. (18)

In the above inequality,Bv is determined as

Bv = supu∈[a,b]

‖Cv − S(u)‖2

d(u)2

. (19)

September 28, 2007 DRAFT

Page 8: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

8

Bound for rendering errors. Apply the methodology of Proposition 1, using the results of Proposi-

tion 2 and 3, we can derive an error bound for the rendered image of the Propagation Algorithm.

Theorem 1:The mean absolute errorMAEPA of the virtual image using the Propagation Algorithm

is bounded by

MAEPA ≤ 3Y3

4Y1∆2

x‖f ′′v ‖∞ + ET + EDBv‖f ′

v‖∞, (20)

wherefv(x) is the virtual image,Yk is defined as in (17),Bv is as in (19), andED, ET are as in (9).

Remark 2: In the first term of (20),Y1 = Hv(b)−Hv(a) is independent of the number of actual cameras

N . The value ofY3, called image-space multiple-view term of third order, encodes the geometrical position

between the actual cameras and the scene. Note that the first term has decay orderO(λ−2), whereλ

is the local density of actual samples. The second term is the intensity estimate error boundET of the

actual cameras. The third term relates to the depth estimate error boundED and the geometrical position

between the scene and the virtual camera (viaBv).

C. Analysis for 2D occluded scenes

In this section, we consider 2D occluded scenes by introducing two adjustments compared to the

analysis in [10]. First, the presence of occlusions requiresmodification of the sample densityλx(x).

Second, intervals containing discontinuities of the virtual image, either caused by the intensity or depth

discontinuities, need to be analyzed using the new methodology of Proposition 1. For simplicity, we

assume that all the occluded samples are successfully removed, and the set of remaining samples are

dense enough so that there exists at most one discontinuity in each sample interval. General analysis is

possible, though less elegant, and produces similar conclusions.

Modification of the sample density.Consider a 2D occluded scene (see Fig. 3). For a cameraΠ, we

define the visibility function

VΠ(u) =

1, if S(u) is visible atΠ

0, if S(u) is not visible atΠ.(21)

Proposition 2 is modified as

λx(x) =1

∆xH ′v(u)

·N∑

i=1

Vi(u)H ′i(u), (22)

whereu is the parameter of the surface pointS(u) having image atx:

u = arg minu

d(u) : HΠ(u) = x. (23)

DRAFT September 28, 2007

Page 9: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

9

X

Y

C

xd,n xd,n+1

xt,m

u−d,n

u+d,n

u+d,n+1

u−d,n+1

VΠ(u) = 0

VΠ(u) = 1

intensity discontinuitydepth discontinuity

ut,m

Fig. 3. A 2D occluded scene. We differentiate two kinds of discontinuities: those due to occlusions (such asxd,n with

parametersu+

d,n andu−

d,n) and those due to the textureT (u) (such asxt,n with parameterut,m).

For this modification,Yk will also be changed to

Yk =

∫ b

a

(N∑

i=1

Vi(u)H ′i(u)

)1−k(Vv(u)H ′

v(u))k

du. (24)

Intuitively, the modification in (24) signifies that, if a surface point S(u) is occluded at an actual

cameraΠi, or equivalentlyVi(u) = 0, this cameraΠi contributes no information to the rendering

of virtual pixel x = Hv(u). Similarly, if S(u) is occluded at the virtual cameraΠv, or equivalently

Vv(u) = 0, no information from actual cameras is necessary.

Incorporation of jumps. We differentiate two categories of discontinuities at the virtual imagefv(x),

namely, the depth discontinuities and the texture discontinuities (see Fig. 3). Thedepthdiscontinuities

are at image object boundaries (backgrounds and foregrounds). LetXd be the set of depth discontinuities.

For each pointxd,n ∈ Xd, denote

u+d,n = lim

x→x+

d,n

H−1v (x), (25)

u−d,n = lim

x→x−

d,n

H−1v (x). (26)

The above equations are well defined sinceH−1v (x) is a one-to-one mapping everywhere except at

discontinuities offv(x). Intuitively, u+d,n is the parameter on the background andu−

d,n is the parameter

on the foreground, or vice-versa.

The texture discontinuities are discontinuities of the textureT (u). We denote the set of texture

discontinuitiesXt. For consistency, we also use notationu+t,n andu−

t,n, as in (25) and (26), forxt,n ∈ Xt,

though they are in fact equal.

September 28, 2007 DRAFT

Page 10: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

10

xnymnymn+1

λx(x) λx(x)

Fig. 4. The observed interval[ymn , ymn+1] around a discontinuityxn of the virtual imagefv(x). Note that the sample density

function λx(x) may or may not, depending on whetherxn ∈ Xd or xn ∈ Xt, be discontinuous atxn.

For each discontinuity

xn ∈ X = Xt

⋃Xd, (27)

the interval[ymn, ymn+1] containingxn is called anobserved interval(or sampled interval–see Fig. 4).

The following lemma is a classical result of Poisson processes.

Lemma 1: [3], [12] Let (ymn+1 − ymn) be the observed interval around each discontinuityxn. The

length of intervals∆2,n = ymn+1 − xn and ∆1,n = xn − ymnare independent and follow exponential

distributions of parameterλ(Hv(u+n )) andλ(Hv(u

−n )), respectively.

Corollary 1: The following equations hold:

E[ymn+1 − ymn] =

1

λ(Hv(u+n ))

+1

λ(Hv(u−n ))

(28)

E[∆1,n∆2,n] =1

λ(Hv(u+n ))

· 1

λ(Hv(u−n ))

. (29)

We define operatorsJ0(f) andJ1(f) as

J0(f) = supx

∣∣ limy→x+

f(y) − limy→x−

f(y)∣∣

(30)

J1(f) = supx

∣∣ limy→x+

f ′(y) − limy→x−

f ′(y)∣∣

. (31)

Theorem 2:The mean absolute errorMAEPA of the virtual image using the Propagation Algorithm

is bounded by

MAEPA ≤ 3Y3

4Y1∆2

x‖f ′′v ‖∞ + ET + EDBv‖f ′

v‖∞

+3

2D0∆xJ0(fv) +

1

2D1∆

2xJ1(fv), (32)

wherefv(x) is the virtual image,Bv is as in (19),Yk is defined in (24),J0(f) andJ1(f) are defined

DRAFT September 28, 2007

Page 11: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

11

in (30) and (31), andD0, D1 are

D0 =∑

xd∈X

1

λx(x+d )

+1

λx(x−d )

(33)

D1 =∑

xd∈X

1

λx(x+d )

· 1

λx(x−d )

. (34)

Proof: The proof is similar to the proof of [10, Theorem 1]; we need to consider in addition the

aggregated error in observed intervals[ymn, ymn+1] around jumpsxn ∈ X. Hence, the error bound

needs to increase by an amount

3

2|ymn+1 − ymn

| · J0(f) +1

2∆1,n∆2,n · J1(f). (35)

The summation these terms, for allxn, in fact results in the additional fourth and fifth terms.

Remark 3:Compared to Theorem 1, the bound in (32) has additional fourthand fifth terms to incor-

porate the discontinuities of the virtual imagefv(x). Overall, the fourth term decays asO(λ−1) and the

fifth term decays asO(λ−2), whereλ is the local density of actual samples.

IV. A NALYSIS FOR 3D SCENES

In this section, we extend the analysis into the 3D case. A natural generalization of piecewise linear

interpolation into 2D is the Delaunay triangulation-basedlinear interpolation. We present the 3D method-

ology for individual triangles in Section IV-A. Then, we show properties of Poisson Delaunay triangles

in Section IV-B and a bound for sample jitters in Section IV-C. Finally, an error analysis for 3D scenes

without occlusions is given in Section IV-D.

A. Methodology

In this section, we investigate the interpolation error foran individual triangle. We define theL∞ norm

of the gradient∇f(x, y) and the Hessian matrix∇2f(x, y) as follows:

‖∇f(x, y)‖∞ = sup(x,y)

‖∇f(x, y)‖2 (36)

‖∇2f(x, y)‖∞ = sup(x,y)

σmax

[∇2f(x, y)

], (37)

whereσmax[M ] denotes the maximum singular value [4] of a matrixM . The linearly interpolated value

at a 2D pointX inside a triangleABC is defined as

f(X) =SA

Sf(A) +

SB

Sf(B) +

SC

Sf(C), (38)

September 28, 2007 DRAFT

Page 12: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

12

X

A

B C

f(X)

f(A)

f(B)

f(C)

O R

Fig. 5. Triangulation-based linear interpolation is often used with the Delaunay triangulation. In each triangle, the interpolation

error can be bounded, as in Proposition 4, using the triangle’s circumcircle radiusR, the sample errorsε, and the sample jitters

µ.

whereSA, SB, SC, andS denote the area of trianglesXBC, AXC, ABX, andABC, respectively.

In other words,f(X) is a bivariate linear function that is equal tof(X) at locationsA, B, andC (see

Fig. 5). In the presence of sample errors and jitters, sample valuesf(A), f(B), andf(C) in (38) are

replaced byf(A + µA) + εA, f(B + µB) + εB, andf(C + µC) + εC, respectively.

Proposition 4: We consider a functionf(x, y) that is twice continuously differentiable. The linear

interpolation on a triangle given in (38) has the error bounded by

|f(x, y) − f(x, y)| ≤ 1

2R2 · ‖∇2f‖∞ + max|ε|

+ max‖µ‖2 · ‖∇f‖∞, (39)

whereR is the circumcircle radius of the triangleABC.

Proof: We show the proof forε = 0 and µ = 0. In this case, the error bound in the right-hand

side of (39) reduces into the first term. The techniques to incorporate the sample error (second term) and

jitter (third term) are similar to the proof of [10, Proposition 1].

Let O be the center of the circumcircle of triangleABC. Using vector manipulations, it can be shown

that

R2 − ‖X − O‖22 =

SA

S‖∆A‖2

2 +SB

S‖∆B‖2

2 +SC

S‖∆C‖2

2, (40)

where∆A = A−X, ∆B = B −X, and∆C = C −X. Using the 2D Taylor expansion we can obtain

f(A) = f(X) + ∇f(X)T · ∆A +1

2∆T

A · ∇2f(Xa) · ∆A

DRAFT September 28, 2007

Page 13: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

13

for some pointXa. Similar equations can be obtained forB andC as well. Hence,

|f(X) − f(X)| =1

2·∣∣∣SA

S∆T

A · ∇2f(Xa) · ∆A +

SB

S∆T

B · ∇2f(Xb) · ∆B +SC

S∆T

C · ∇2f(Xc) · ∆C

∣∣∣

≤ 1

2‖∇2f‖∞

(SA

S‖∆A‖2

2 +SB

S‖∆B‖2

2 +SC

S‖∆C‖2

2

)

≤ 1

2‖∇2f‖∞R2.

The bound in (39) suggests that we need to investigate the properties of Delaunay triangles and the

sample jitters. The next two sections will present these properties.

B. Properties of Poisson Delaunay triangles

We assume in this section that the scene is unoccluded. We start by proposing an equivalence of [10,

Lemma 1] for the 2D case. A 2D processp is calledidentically distributed scattering[11] if the density

of points ofp over an arbitrary regionω follows a fixed probability mass distribution independent ofω.

Intuitively, there is a profound similarity between the 1D and 2D cases, since they are both related to

the probability mass function (pmf) of the number of points falling inside an arbitrary region. Hence, in

the following, we assume that the Hypothesis 1 below holds.

Hypothesis 1:The superposition of 2D point processes with identically distributed scattering property

can be approximated as a 2D Poisson process.

On the image plane of the virtual cameraΠv, let Y be the set of pointspropagatedfrom actual

pixels [8].

Proposition 5: The point processY can be approximated as a 2D generalized Poisson process with

density function

λ(x, y) =1

∆x∆y

N∑

i=1

det(∂H i/∂(u, v)

)

det(∂Hv/∂(u, v)

) , (41)

where(u, v) = H−1v (x, y).

Proof: Since we assume that Hypothesis 1 holds, in each infinitesimal region, the point processYcan be considered as a 2D Poisson process. Hence, overall,Y can be considered as a generalized Poisson

process. The densityλ(x, y) can be computed, similarly to [10, Section III.B], as the average number of

points falling on an unit area. This indeed results in (41).

September 28, 2007 DRAFT

Page 14: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

14

Once we approximate the set of propagated pointsY as a 2D Poisson process, the next step is to

investigate properties of Poisson Delaunay triangles. In the following, we exploit results from stochastic

geometry.

Lemma 2: [11, Chapter 5] The circumradiusR and the areaS of Delaunay triangles of a 2D Poisson

process of densityλ are independent. The circumradiusR has the probability density function (pdf)

2(πλ)2r3e−πλr2

, r > 0. (42)

The momentsE[Sk] can be computed using explicit formula. In particular

E[R2] =2

πλ, E[S] =

1

2λ, E[S2] =

35

8π2λ2. (43)

C. Bound for sample jitters

Let S = [X, Y, Z]T be a surface point, andp be the image ofS at the virtual cameraΠv. We denote

Se = [Xe, Ye, Ze]T a noisy estimate ofS with reconstruction errorεD = Se − S, and p the image of

Se at Πv.

Proposition 6: The jitter µ = p − p, at virtual cameraΠv with camera centerCv, caused by depth

estimate errors can be bounded by

‖µ‖2 ≤√

2EDBv. (44)

In the above inequality,Bv is computed as

Bv = sup(u,v)∈Ω

‖Cv − S(u, v)‖2

d(u, v)2

. (45)

Proof: The jitter µ is a two-coordinate vectorµ = [µx, µy]T . It can be shown [10, Section III.C]

that the norm of bothµx andµy is bounded byEDBv. Hence,

‖µ‖2 =√

µ2x + µ2

y ≤√

2EDBv.

The boundEDBv depends on the depth estimate errorED and the relative position between the virtual

camera and the scene defined byBv.

DRAFT September 28, 2007

Page 15: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

15

D. Analysis for 3D unoccluded scenes

Consider the intensity functionfv(x, y) = T(H−1

v (x, y))

at virtual cameraΠv. Let e(x, y) = fv(x, y)−fv(x, y) be the interpolation error andNΩ be the set of virtual pixels(m, n) being images of surface

points S(u, v) for (u, v) ∈ Ω. Denote#NΩ the number of pixels inNΩ. The mean absolute error

MAEPA is defined as

MAEPA =1

#NΩ

(m,n)∈NΩ

|e(m∆x, n∆y)|. (46)

Theorem 3:The mean absolute errorMAEPA of the virtual image using the Propagation Algorithm

is bounded by

MAEPA ≤ X2

πX1∆x∆y‖∇2fv‖∞ + ET +

√2EDBv‖∇fv‖∞, (47)

wherefv is the the virtual image,Bv is as in (45),ED, ET are as in (9), and

Xk =

Ω

(N∑

i=1

det

(∂H i(u, v)

∂(u, v)

))1−k (det

(∂Hv(u, v)

∂(u, v)

))k

dudv. (48)

Proof: Let D be the set of Delaunay triangles, andΩxy = Hv(Ω) be the image region of the

surfaceS(u, v) at the virtual camera. TheMAEPA can be approximated as

MAEPA ≈ 1

SΩxy

Ωxy

|e(x, y)|dxdy (49)

=1

SΩxy

∆i∈D

∆i

|e(x, y)|dxdy

≤ 1

SΩxy

∆i∈D

S∆i

(1

2R2

i ‖∇2fv‖∞ + max|ε|

+ max‖µ‖2 · ‖∇fv‖∞). (50)

In each infinitesimal patchdω around(x, y) ∈ Ωxy, we can approximateR2 ≈ 2/(πλ(x, y)) (see

Lemma 2). Hence∑

∆i∈D

S∆iR2

i ≈∫

Ωxy

2

πλ(x, y)dω =

2X2

π· ∆x∆y. (51)

By changing the variables from(x, y) to (u, v), and substituting (51) into inequality (50), we indeed

get (47).

Remark 4:The first term of (47),X1 = SΩxy, is the area of the scene’s image on the virtual image

plane and does not depend on the actual camera configuration. The value ofX2, called 3D multiple-view

term of second order, encodes the geometrical information of the actual cameras and the scene. We note

that X2 decays with orderN−1 whenN tends to infinity. The first term also depends on the resolution

September 28, 2007 DRAFT

Page 16: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

16

TABLE I

EXPERIMENTAL VALUES OF E[R2], E[S], AND E[S2] IN THE CASE WHEREN = 10 ACTUAL CAMERAS ARE USED,

COMPARED TO THEORETICAL VALUES OFPOISSONDELAUNAY TRIANGLES .

E[R2] E[S] E[S2]

Experiments 0.0568 0.05 0.003845

Poisson Delaunay triangles0.0637 0.05 0.004432

Relative error 11% 0% 13%

∆x∆y. Overall, in smooth regions, the first term has decay orderO(λ−1), whereλ is the local density

of actual samples. The second term is the intensity estimate error boundET . The third term relates to

the depth estimate error boundED (linearly) and the geometrical position between the scene and the

virtual camera (viaBv).

Remark 5:A notable difference between the 3D case and the 2D case resides in the decay order of

the first term. In (20), the first term has decay orderO(λ−2), while in (47) the decay order isO(λ−1).

To see this difference, note that the first term in inequality (39) containsR2 having the same dimension

with the sample densityλ, whereas in (13), the first term contains(x2−x1)2 of the same dimension with

λ2. This intriguing finding supports a common practice of conducting image rectifications to simplify the

rendering process into 2D. Rectifying images using bilinear interpolation offers a decay ofO(λ−2) in

smooth regions, andO(λ−1) around the discontinuities. Hence, the image rectification not only reduces

the complexity, but also increases the decay rate in smooth regions fromO(λ−1) to O(λ−2). Obviously,

one needs to take into account that image rectifications causeadditional errors elsewhere.

E. Numerical experiments

Support for a positive answer of Hypothesis 1 is shown in TableI. Experimental values forR2, S and

S2 of Delaunay triangles, whereN = 10 actual cameras are used, are computed and compared to the

theoretical values of Poisson Delaunay triangles proposed in Lemma 2. Observe that the approximations

are relatively accurate.

Next, we validate the error bound (47) of Theorem 3 for a 3D synthetic scene consisting of a flat

surface with constant depthz = 10 and the texture mapT (u, v) = sin(u) + sin(v). The Propagation

Algorithm [8] is used for a planar camera configuration. All the actual and virtual cameras are placed in

DRAFT September 28, 2007

Page 17: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

17

101

102

103

104

10−3

10−2

10−1

MAEPA

Theoretical bound

MA

E

number of actual pixels

Fig. 6. The rendering errors plotted against the total number of actual pixels. We note the errors indeed have decayO(λ−1),

whereλ is the local sample density, as stated in Theorem 3.

the xy-plane and focus to the direction of thez-axis. Specifically,N = 10 actual cameras are randomly

placed in the square of dimensions2 × 2 centered around the virtual camera position at[5, 5, 0]T .

To validate the first term, we setED = ET = 0 and plot in Fig. 6 the mean absolute errors MAE

(solid) and the error bound (dashed) against the total number of actual pixels (equivalent to the local

density of actual samplesλ). The variation ofλ is obtained by changing the resolution∆x∆y. Observe

that the MAE indeed decays asO(λ−1), conforming to Theorem 3.

To validate the second term, we fix∆x = ∆y = 0.02 andED = 0, and varyET . For each value of

ET , the intensity estimate errors are chosen randomly in the interval [−ET , ET ] following the uniform

distribution. In Fig. 7, we show the mean absolute error MAE (solid) and the theoretical bound (dashed)

plotted against the intensity estimate error boundET . Observe that the actual MAE fluctuates around

one half of the error bound. The reason is that the error bound of Theorem 3 is derived for the worst

case, whereas the actual MAE tends to follow the average errors.

Finally, we validate the last term of (47) by fixing∆x = ∆y = 0.02, ET = 0, and varyingED. For

each value ofED, the depth estimate errors are chosen randomly in the interval [−ED, ED] following

the uniform distribution. In Fig. 8, we show the mean absoluteerror MAE (solid) and the theoretical

bound (dashed) plotted against the depth estimate error bound ED. Observe that the MAE indeed appears

below the error bound and approximately linear toED.

September 28, 2007 DRAFT

Page 18: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

18

1 2 3 4 5 6 7 8 9 100

0.002

0.004

0.006

0.008

0.01

0.012MAE

PA

Theoretical bound

MA

E

ET

Fig. 7. The mean absolute error (MAE) (solid) and the theoretical bound(dashed) plotted against the intensity estimate error

boundET .

1 2 3 4 5 6 7 8 9 10

x 10−3

0

0.005

0.01

0.015

0.02

0.025

0.03MAE

PA

Theoretical bound

ED

MA

E

Fig. 8. The mean absolute error (MAE) (solid) and the theoretical bound(dashed) plotted against the depth estimate error

boundED.

V. CONCLUSION

We presented a quantitative analysis for IBR algorithms to 2D occluded scenes and 3D unoccluded

scenes, extending the error aggregation framework proposed in a companion paper [10]. To analyze 2D

occluded scenes, we modified the sample density function and measured the effects of jumps in observed

sample intervals around the discontinuities. For 3D unoccluded scenes, we proposed an error bound for

the technique of triangulation-based linear interpolation and exploited properties of Poisson Delaunay

triangles. We derived an error bound for the mean absolute error (MAE) of the virtual images. The error

DRAFT September 28, 2007

Page 19: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

19

bound successfully captures the effects of the scene and thecamera configuration to the rendering quality,

as validated by numerical experiments. In particular, the proposed analysis suggests that the decay order

of the MAE is O(λ−1) for 3D scenes andO(λ−2) for 2D scenes. An implication is that building IBR

systems that can simplify to 2D, besides reducing the complexity, also increases the decay rate of the

rendering errors fromO(λ−1) to O(λ−2) in smooth regions.

Limitations. The proposed analysis approximates summations as integralsat (49) and (51), and actual

samples as a generalized Poisson process. These approximations can be further analyzed, though it might

not lead to further understandings in the context of IBR.

Future work. We would like to prove Hypothesis 1, extend the analysis to 3Doccluded scenes, and

analyze the mean and variance of the rendering errors.

APPENDIX

A. Proof of Proposition 1

Denote the following functions as linear interpolations ofcorresponding samples:

f12(x) =x − x1

x2 − x1f(x2) +

x2 − x

x2 − x1f(x1)

f1d(x) =x − x1

xd − x1f(x−

d ) +xd − x

xd − x1f(x1)

fd2(x) =x − xd

x2 − xd

f(x2) +x2 − x

x2 − xd

f(x+d ).

Let e12(x), e1d(x), anded2(x) be the corresponding interpolation errors. The aggregated interpolation

error is defined as

E12 =

∫ x2

x1

|e12(x)|dx (52)

for f12(x) over the interval[x1, x2]. The aggregated errorsE1d andEd2 are defined similarly.

Lemma 3:The equality

f12(xd) =∆1

∆f(x+

d ) +∆2

∆f(x−

d ) +∆1∆2

∆J1 + B (53)

holds for someB such that

|B| ≤ 1

2∆1∆2 · ‖f ′′‖∞. (54)

Proof: Using the Taylor expansion we write

f(x1) = f(x−d ) − ∆1f

′(x−d ) +

1

2∆2

1f′′(ξ1) (55)

September 28, 2007 DRAFT

Page 20: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

20

f(x−d )

f(x+d )

xdx1 x2

f(x)

f12(x)

Fig. 9. Linear interpolation error.

for someξ1 ∈ [x1, xd]. A similar equation can be also derived forx2. Hence (53) holds for

B =∆2

1∆2

2∆f ′′(ξ1) +

∆1∆22

2∆f ′′(ξ2). (56)

For B defined above, it is easy verify (53).

Next, we propose a bound, for the caseµi = εi = 0 for i = 1, 2, that is in fact tighter than the one

proposed in Proposition 1.

Lemma 4:The aggregated errorE12, when there are no sample errors and jitters, can be bounded by

E12 ≤ 1

12∆3 · ‖f ′′‖∞ +

∆21 + ∆2

2

2∆· |J0| +

1

2∆1∆2 · |J1|. (57)

Proof: We can boundE12 by the summation ofE1d, Ed2, and the area of the quadrangle formed

by [x1, f(x1)]T , [x2, f(x2)]

T , [xd, f(x+d )]T , and [xd, f(x−

d )]T (the shaded region in Fig. 9). Hence

E12 ≤ E1d + Ed2 +∆1

2|f12(xd) − f(x−

d )| + ∆2

2|f12(xd) − f(x+

d )|. (58)

Next, inequalities similar to [10, Equation (11)] can be derived for E1d, Ed2. Integrating both sides of

these inequalities we obtain

E1d ≤ 1

12∆3

1 · ‖f ′′‖∞, Ed2 ≤ 1

12∆3

2 · ‖f ′′‖∞. (59)

Substitutingf12(xd) as in (53) into (58), together with inequalities (59), we will indeed prove (57).

Finally, to extend Lemma 4 in the presence of sample errors and jitters, it is sufficient to prove the

following lemma.

Lemma 5:The following inequality holds fori = 1, 2:

|f(xi + µi) + εi − f(xi)| ≤ |εi| + |µi| · ‖f ′‖∞ + |J0|. (60)

DRAFT September 28, 2007

Page 21: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

21

Proof: For arbitraryx, y ∈ [x1, x2], with x ≤ xd ≤ y:

|f(y) − f(x)| ≤ |f(y) − f(x+d )| + |J0| + |f(x−

d ) − f(x)|

≤ |y − xd| · |f ′(θ1)| + |xd − x| · |f ′(θ2)| + |J0|

≤ |y − x| · ‖f ′‖∞ + |J0|.

The last inequality easily implies (60).

B. Geometrical interpretation ofHΠ(u, v)

We present a property of the scene-to-image mappingHΠ(u, v) in this appendix–a generalization of

the 2D case shown in [10]. In the following, we useS instead ofS(u, v). We denote

Su(u, v) = S(u, v) +∂S(u, v)

∂u,

Sv(u, v) = S(u, v) +∂S(u, v)

∂v.

Lemma 6:The Jacobian∂HΠ(u, v)/∂(u, v) of the scene-to-image mapping has the determinant

det

(∂HΠ(u, v)

∂(u, v)

)=

det(A)

d(u, v)3, (61)

where

A = Π ·[Su, Sv, S

]. (62)

Let e4 = [0, 0, 0, 1]T ∈ R4. Taking the determinant of the following equality

Π

eT4

·

[Su, Sv, S, C

]=

A 0

1 1

,

we obtain:

det(A) = 6VSuSvSC, (63)

whereVSuSvSC is the volume of the tetrahedronSuSvSC. We summarize the result in Proposition 7.

Proposition 7: The Jacobian∂HΠ(u, v)/∂(u, v) of the scene-to-image mapping the has determinant

det

(∂HΠ(u, v)

∂(u, v)

)=

6VSuSvSC

d(u, v)3. (64)

September 28, 2007 DRAFT

Page 22: 1 Quantitative Analysis for Image-Based Rendering with ...minhdo/publications/ibr_qa_p2.pdf · depth and intensity estimate errors, the scene geometry and texture, the number of actual

22

REFERENCES

[1] C. de Boor,A Practical Guide to Splines. New York, NY: Springer-Verlag, 1978.

[2] D. A. Forsyth and J. Ponce,Computer Vision: A Modern Approach. New York, NY: Prentice-Hall, 2002.

[3] R. G. Gallager,Discrete Stochastic Processes. Norwell, MA: Kluwer Academic Publishers, 1996.

[4] G. H. Golub and C. F. V. Loan,Matrix Computations (3rd Ed.). Baltimore, MD: Johns Hopkins University Press, 1996.

[5] S. B. Kang, Y. Li, X. Tong, and H.-Y. Shum,Foundations and Trends in Computer Graphics and Vision: Image-based

Rendering, B. Curless, L. V. Gool, and R. Szeliski, Eds. Hanover, MA: Now Publishers, 2006, vol. 2, no. 3.

[6] W. Mark, L. McMillan, and G. Bishop, “Post-rendering 3D warping,” in Proc. I3D Graphics Symp., 1997, pp. 7–16.

[7] L. McMillan, “An image-based approach to three-dimensional computer graphics,” Ph.D. dissertation, University of North

Carolina at Chapel Hill, 1997.

[8] H. T. Nguyen and M. N. Do, “Image-based rendering with depth information using the propagation algorithm,” inProc.

IEEE Int. Conf. Acoust., Speech, and Signal Proc., vol. 2, Philadelphia, March 2005, pp. 589–592.

[9] ——, “Error analysis for image-based rendering with depth information,” in Proc. IEEE Int. Conf. on Image Proc., October

2006, pp. 381–384.

[10] ——, “Quantitative analysis for image-based rendering with depth information: 2D unoccluded scenes,”IEEE Trans. Image

Proc., 2007, submitted for publication.

[11] A. Okabe, B. Boots, K. Sugihara, and S. N. Chiu,Spatial Tessellations: Concepts and Applications of Voronoi Diagrams,

2nd ed. New York, NY: Wiley, 2000.

[12] A. Papoulis,Probability, Random Variables and Stochastic Processes, 2nd ed. New York, NY: McGraw-Hill, 1984.

[13] P. M. Prenter,Splines and Variational Methods. New York, NY: Wiley, 1975.

[14] J. Shade, S. Gortler, L. He, and R. Szeliski, “Layered depth images,” in Proc. of SIGGRAPH, 1998, pp. 231–242.

[15] H.-Y. Shum, S.-C. Chan, and S. B. Kang,Image-Based Rendering. New York, NY: Springer, 2007.

[16] C. Zhang and T. Chen, “A survey on image-based rendering - representation, sampling and compression,”EURASIP Signal

Processing: Image Communication, pp. 1–28, January 2004.

DRAFT September 28, 2007