8
Supplementary Material for Layout-Guided Novel View Synthesis from a Single Indoor Panorama Jiale Xu 1 Jia Zheng 2 Yanyu Xu 3 Rui Tang 2 Shenghua Gao 1,4 * 1 ShanghaiTech University 2 KooLab, Manycore 3 Institute of High Performance Computing, A*STAR 4 Shanghai Engineering Research Center of Intelligent Vision and Imaging {xujl1, gaoshh}@shanghaitech.edu.cn {jiajia, ati}@qunhemail.com xu [email protected] In this supplementary material, we first present the de- tails of the spherical geometric transformation process. Then, we explain the relationship between pixel missing rate and the camera translation distance. Finally, we show more qualitative results and some failure cases. 1. Spherical Geometric Transformation A description of the panorama. In this work, we assume that panoramas are generated with equirectangular projec- tion, which means the horizontal and vertical pixel coor- dinates relate linearly to the longitude and latitude of the spherical coordinates, respectively. A panorama I covers 360° field-of-view (FoV) horizontally and 180° FoV verti- cally, and the center (W/2, H/2) aligns to the 0-longitude and 0-latitude. Different from perspective images, each pixel of the panorama corresponds to a direction in the spherical coor- dinate system. Its depth value denotes the 3D Euclidean distance from the camera center to the nearest point along this direction. Therefore, we can know the 3D location of each source-view pixel once the depth map D s is estimated, then the feature map F s can be seen as a point cloud P and projected into the novel views. Mapping from source-view to target-view. The mapping from a source-view pixel (u s ,v s ) ∈P s to a target-view pixel (u t ,v t ) ∈P s is determined by a series of coordinate transformation f from P s to P t : f = f St7→Pt f Ct7→St f Cs7→Ct f Ss7→Cs f Ps7→Ss . (1) Given a source-view pixel (u s ,v s ) ∈P s , we first trans- form it to spherical coordinates (φ s s ) ∈S s as follows: f Ps7→Ss : φ s = u s · 2π/W - π θ s = -v s · π/H + π/2 (2) * Corresponding author. The spherical coordinates (φ s s ) can only determine the viewing direction of the pixel, we still need its distance relative to the camera p s to locate its 3D position. Fortu- nately, the distance d and can be obtained from the esti- mated depth map D s . Thus, the 3D position (x s ,y s ,z s ) of the pixel relative to camera p s can be calculated by: f Ss7→Cs : x s = d · sin(φ s ) · cos(θ s ) y s = d · cos(φ s ) · cos(θ s ) z s = d · sin(θ s ) (3) Now we have a 3D point (x s ,y s ,z s ) ∈C s , it is easy to transform it to the target view since there is only a transla- tion t =[t x ,t y ,t z ] T = p t - p s between the two views: f Cs7→Ct : x t = x s - t x y t = y s - t y z t = z s - t z (4) Finally, we conduct the inverse mapping of Eq. (3) and Eq. (2) to project the 3D point (x t ,y t ,z t ) ∈C t to a target- view panorama pixel (u t ,v t ) ∈P t : f Ct7→St : φ t = arctan(x t /y t ) θ t = arctan(z t / p x 2 t + y 2 t ) (5) f St7→Pt : u t =(π + φ t ) · W/(2π) v t =(π/2 - θ t ) · H/π (6) Details of layout transformation. The estimated 2D room layout L s R N×2 is an ordered list of the pixel coordinates of room corners. Two adjacent rows in L s represent the upper and lower corners on the same wall-wall boundary, respectively. The transformation from the source-view 2D layout L s to the target-view 2D layout L t is similar to the feature map transformation. For each 2D corner (u s ,v s ) ∈P s , we 1

Supplementary Material for Layout-Guided Novel View

  • Upload
    others

  • View
    8

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Supplementary Material for Layout-Guided Novel View

Supplementary Material forLayout-Guided Novel View Synthesis from a Single Indoor Panorama

Jiale Xu1 Jia Zheng2 Yanyu Xu3 Rui Tang2 Shenghua Gao1,4*

1ShanghaiTech University 2KooLab, Manycore3Institute of High Performance Computing, A*STAR

4Shanghai Engineering Research Center of Intelligent Vision and Imaging{xujl1, gaoshh}@shanghaitech.edu.cn {jiajia, ati}@qunhemail.com

xu [email protected]

In this supplementary material, we first present the de-tails of the spherical geometric transformation process.Then, we explain the relationship between pixel missingrate and the camera translation distance. Finally, we showmore qualitative results and some failure cases.

1. Spherical Geometric TransformationA description of the panorama. In this work, we assumethat panoramas are generated with equirectangular projec-tion, which means the horizontal and vertical pixel coor-dinates relate linearly to the longitude and latitude of thespherical coordinates, respectively. A panorama I covers360° field-of-view (FoV) horizontally and 180° FoV verti-cally, and the center (W/2, H/2) aligns to the 0-longitudeand 0-latitude.

Different from perspective images, each pixel of thepanorama corresponds to a direction in the spherical coor-dinate system. Its depth value denotes the 3D Euclideandistance from the camera center to the nearest point alongthis direction. Therefore, we can know the 3D location ofeach source-view pixel once the depth map Ds is estimated,then the feature map Fs can be seen as a point cloud P andprojected into the novel views.Mapping from source-view to target-view. The mappingfrom a source-view pixel (us, vs) ∈ Ps to a target-viewpixel (ut, vt) ∈ Ps is determined by a series of coordinatetransformation f from Ps to Pt:

f = fSt 7→Pt◦ fCt 7→St ◦ fCs 7→Ct ◦ fSs 7→Cs ◦ fPs 7→Ss . (1)

Given a source-view pixel (us, vs) ∈ Ps, we first trans-form it to spherical coordinates (φs, θs) ∈ Ss as follows:

fPs 7→Ss :

{φs = us · 2π/W − πθs = −vs · π/H + π/2

(2)

*Corresponding author.

The spherical coordinates (φs, θs) can only determinethe viewing direction of the pixel, we still need its distancerelative to the camera ps to locate its 3D position. Fortu-nately, the distance d and can be obtained from the esti-mated depth map Ds. Thus, the 3D position (xs, ys, zs) ofthe pixel relative to camera ps can be calculated by:

fSs 7→Cs :

xs = d · sin(φs) · cos(θs)ys = d · cos(φs) · cos(θs)zs = d · sin(θs)

(3)

Now we have a 3D point (xs, ys, zs) ∈ Cs, it is easy totransform it to the target view since there is only a transla-tion t = [tx, ty, tz]

T = pt − ps between the two views:

fCs 7→Ct :

xt = xs − txyt = ys − tyzt = zs − tz

(4)

Finally, we conduct the inverse mapping of Eq. (3) andEq. (2) to project the 3D point (xt, yt, zt) ∈ Ct to a target-view panorama pixel (ut, vt) ∈ Pt:

fCt 7→St :

{φt = arctan(xt/yt)

θt = arctan(zt/√x2t + y2t )

(5)

fSt 7→Pt :

{ut = (π + φt) ·W/(2π)vt = (π/2− θt) ·H/π

(6)

Details of layout transformation. The estimated 2D roomlayout Ls ∈ RN×2 is an ordered list of the pixel coordinatesof room corners. Two adjacent rows in Ls represent theupper and lower corners on the same wall-wall boundary,respectively.

The transformation from the source-view 2D layout Ls

to the target-view 2D layout Lt is similar to the featuremap transformation. For each 2D corner (us, vs) ∈ Ps, we

1

Page 2: Supplementary Material for Layout-Guided Novel View

0.0 0.5 1.0 1.5 2.0 2.5 3.0

Camera translation distance (m)

0

20

40

60

80

100

Ave

rage

pix

elm

issi

ng

rate

(%)

Figure 1. The relationship between the average pixel missing rateand the camera translation distance. The red dotted lines and thegreen dotted lines show the range of our generated easy set andhard set, respectively.

need to map it to (ut, vt) ∈ Pt. The transformation from(us, vs) ∈ Ps to (φs, θs) ∈ Ss is the same as Eq. (2). How-ever, the corner depth d is not available from Ds since thecorner may be occluded by foreground objects. Therefore,Eq. (3) cannot be applied directly.

Instead, we estimate the corner depth with the cam-era height h. Given the source camera position ps =[px, py, pz], the camera height relative to the ground ish = pz . Let (φs,u, θs,u) and (φs,l, θs,l) denote the sphericalcoordinates of a pair of upper and lower corners, which be-long to the same wall-wall boundary. We can estimate thedepths of both corners as:{

du = h/| tan(θs,l) · cos(θs,u)|dl = h/| sin(θs,l)|

(7)

With the estimated depths of all layout corners, we applyEq. (3), Eq. (4), Eq. (5), and Eq. (6) to obtain their target-view pixel coordinates and form the target 2D layout Lt.

2. More Details of Our DatasetTo clarify the difficulties of our dataset settings, we vi-

sualize the relationship between the pixel missing rate afterthe forward splatting operation and the camera translationdistance. First, we randomly sample 100 panoramas fromour dataset. Then we set the camera translation from 0mto 3m with a step size of 0.1m in 5 random directions, andrecord the pixel missing rates after the forward splatting op-eration. As shown in Figure 1, the average pixel missingrate increases as the camera translation distance gets larger,which means that the difficulty of target-view inpaintingrises. For the easy set, the camera translation ranges from0.2m to 0.3m, and the average pixel missing rate ranges

from 4.04% to 6.03%. For the hard set, the camera trans-lation ranges from 1.0m to 2.0m, and the average pixelmissing rate ranges from 34.1% to 75.8%, which is a verychallenging setting that has rarely been considered before.

3. More Qualitative ResultsMore results on our synthetic dataset. Additional qualita-tive results on our synthetic easy set and hard set are shownin Figure 2 and Figure 3, respectively.More results on real-scene datasets. Additional qualita-tive results on real-scene datasets, i.e., 2D-3D-S dataset [1]and PanoContext dataset [2], are shown in Figure 4 and Fig-ure 5, respectively.Qualitative comparisons of the ablation study. Addi-tional qualitative comparisons for the ablation study of thelayout guidance are shown in Figure 6. We make the follow-ing observations: (i) both models generate plausible resultswithin the seen regions. (ii) the model with layout guid-ance can preserve the room layout better. In some cases,the network can even generate unseen room structures fromscratch with the help of the layout prior. (iii) the modelwithout layout guidance produces blurry textures and indis-tinguishable layout boundaries in the unseen regions. Theresults demonstrate the effectiveness of the layout guidance.

4. Failure CasesWe show some typical failure cases in Figure 7. In Fig-

ure 7a and Figure 7b, the textures within the mirror andglass regions are incorrect. Since the depth estimation mod-ule does not explicitly model the reflection and refraction,the depth of the objects reflected by the mirror or behind theglass cannot be represented by the estimated depth map. InFigure 7c, some furniture in the scene has complex and thinstructures, which are hard to be recovered from the coarseestimated depth map and result in conspicuous pixel dis-tortion in the synthesized image. In Figure 7d, when theviewpoint change is too violent, e.g., from one side to an-other side of an object, the network could not synthesizethe unseen surface without any contextual information ofthe object.

References[1] Iro Armeni, Sasha Sax, Amir R Zamir, and Silvio Savarese.

Joint 2d-3d-semantic data for indoor scene understanding.CoRR, abs/1702.01105, 2017. 2

[2] Yinda Zhang, Shuran Song, Ping Tan, and Jianxiong Xiao.Panocontext: A whole-room 3d context model for panoramicscene understanding. In ECCV, pages 668–686, 2014. 2

2

Page 3: Supplementary Material for Layout-Guided Novel View

Target View (Ground Truth)Source View Synsin (sup. by GT depth) MPI (128 layers) Ours (sup. by GT depth)

Figure 2. Additional qualitative results on our synthetic easy set.

3

Page 4: Supplementary Material for Layout-Guided Novel View

Target View (Ground Truth)Source View Synsin (sup. by GT depth) MPI (128 layers) Ours (sup. by GT depth)

Figure 3. Additional qualitative results on our synthetic hard set.

4

Page 5: Supplementary Material for Layout-Guided Novel View

Source View Camera Translation = 0.5 m Camera Translation = 1.0 m Camera Translation = 1.5 m

Figure 4. Additional qualitative results on 2D-3D-S dataset.

5

Page 6: Supplementary Material for Layout-Guided Novel View

Source View Camera Translation = 0.5 m Camera Translation = 1.0 m Camera Translation = 1.5 m

Figure 5. Additional qualitative results on PanoContext dataset.

6

Page 7: Supplementary Material for Layout-Guided Novel View

Source View Target View (Ground Truth)Ours (without layout) Ours (with layout)

Figure 6. Additional qualitative comparisons for the ablation study of the room layout guidance.

7

Page 8: Supplementary Material for Layout-Guided Novel View

(a) Mirrors.

(b) Glass.

(c) Fine structures.

(d) Lack of context.Figure 7. Failure cases on our synthetic dataset. From left to right: source-view image, estimated depth map, synthesized target-viewimage, and ground-truth target-view image.

8