7
Supplementary Material for Video Propagation Networks Varun Jampani 1 , Raghudeep Gadde 1,2 and Peter V. Gehler 1,2 1 Max Planck Institute for Intelligent Systems, T ¨ ubingen, Germany 2 Bernstein Center for Computational Neuroscience, T ¨ ubingen, Germany {varun.jampani,raghudeep.gadde,peter.gehler}@tuebingen.mpg.de 1. Parameters and Additional Results In this supplementary, we present experiment protocols and additional qualitative results for experiments on video object segmentation, semantic video segmentation and video color propagation. Table 1 shows the feature scales and other parame- ters used in different experiments. Figures 1, 2 show some qualitative results on video object segmentation with some failure cases in Fig. 3. Figure 4 shows some qualitative results on semantic video segmentation and Fig. 5 shows results on video color propagation. Experiment Feature Type Feature Scale-1, Λa Feature Scale-2, Λ b α Input Frames Loss Type Video Object Segmentation (x, y, Y, Cb, Cr, t) (0.02,0.02,0.07,0.4,0.4,0.01) (0.03,0.03,0.09,0.5,0.5,0.2) 0.5 9 Logistic Semantic Video Segmentation with CNN1 [5]-NoFlow (x, y, R, G, B, t) (0.08,0.08,0.2,0.2,0.2,0.04) (0.11,0.11,0.2,0.2,0.2,0.04) 0.5 3 Logistic with CNN1 [5]-Flow (x+ux,y+uy, R, G, B, t) (0.11,0.11,0.14,0.14,0.14,0.03) (0.08,0.08,0.12,0.12,0.12,0.01) 0.65 3 Logistic with CNN2 [3]-Flow (x+ux,y+uy, R, G, B, t) (0.08,0.08,0.2,0.2,0.2,0.04) (0.09,0.09,0.25,0.25,0.25,0.03) 0.5 4 Logistic Video Color Propagation (x,y,I,t) (0.04,0.04,0.2,0.04) No second kernel 1 4 MSE Table 1. Experiment Protocols. Experiment protocols for the different experiments presented in this work. Feature Types: Feature spaces used for the bilateral convolutions, with position (x, y) and color (R, G, B or Y,Cb,Cr) features [0, 255]. ux, uy denotes optical flow with respect to the present frame and I denotes grayscale intensity. Feature Scales (Λa, Λ b ): Validated scales for the features used. α: Exponential time decay for the input frames. Input Frames: Number of input frames for VPN. Loss Type: Type of loss used for back-propagation. “MSE” corresponds to Euclidean mean squared error loss and “Logistic” corresponds to multinomial logistic loss. 1

Supplementary Material for Video Propagation Networksopenaccess.thecvf.com/content_cvpr_2017/... · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Supplementary Material for Video Propagation Networksopenaccess.thecvf.com/content_cvpr_2017/... · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep

Supplementary Material for Video Propagation Networks

Varun Jampani1, Raghudeep Gadde1,2 and Peter V. Gehler1,21Max Planck Institute for Intelligent Systems, Tubingen, Germany

2Bernstein Center for Computational Neuroscience, Tubingen, Germany{varun.jampani,raghudeep.gadde,peter.gehler}@tuebingen.mpg.de

1. Parameters and Additional ResultsIn this supplementary, we present experiment protocols and additional qualitative results for experiments on video object

segmentation, semantic video segmentation and video color propagation. Table 1 shows the feature scales and other parame-ters used in different experiments. Figures 1, 2 show some qualitative results on video object segmentation with some failurecases in Fig. 3. Figure 4 shows some qualitative results on semantic video segmentation and Fig. 5 shows results on videocolor propagation.

Experiment Feature Type Feature Scale-1, Λa Feature Scale-2, Λb αInput

Frames Loss Type

Video Object Segmentation (x, y, Y, Cb, Cr, t) (0.02,0.02,0.07,0.4,0.4,0.01) (0.03,0.03,0.09,0.5,0.5,0.2) 0.5 9 Logistic

Semantic Video Segmentationwith CNN1 [5]-NoFlow (x, y, R,G,B, t) (0.08,0.08,0.2,0.2,0.2,0.04) (0.11,0.11,0.2,0.2,0.2,0.04) 0.5 3 Logistic

with CNN1 [5]-Flow (x+ux, y+uy, R,G,B, t) (0.11,0.11,0.14,0.14,0.14,0.03) (0.08,0.08,0.12,0.12,0.12,0.01) 0.65 3 Logistic

with CNN2 [3]-Flow (x+ux, y+uy, R,G,B, t) (0.08,0.08,0.2,0.2,0.2,0.04) (0.09,0.09,0.25,0.25,0.25,0.03) 0.5 4 Logistic

Video Color Propagation (x, y, I, t) (0.04,0.04,0.2,0.04) No second kernel 1 4 MSE

Table 1. Experiment Protocols. Experiment protocols for the different experiments presented in this work. Feature Types: Feature spacesused for the bilateral convolutions, with position (x, y) and color (R,G,B or Y,Cb, Cr) features ∈ [0, 255]. ux, uy denotes opticalflow with respect to the present frame and I denotes grayscale intensity. Feature Scales (Λa,Λb): Validated scales for the features used.α: Exponential time decay for the input frames. Input Frames: Number of input frames for VPN. Loss Type: Type of loss used forback-propagation. “MSE” corresponds to Euclidean mean squared error loss and “Logistic” corresponds to multinomial logistic loss.

1

Page 2: Supplementary Material for Video Propagation Networksopenaccess.thecvf.com/content_cvpr_2017/... · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep

Frame 5 Frame 10 Frame 15 Frame 20 Frame 25 Frame 30

Input

Vid

eo

GT

BV

SO

FL

VP

NV

PN

-DLab

Figure 1. Video Object Segmentation. Shown are the different frames in example videos with the corresponding ground truth (GT) masks,predictions from BVS [2], OFL [4], VPN (VPN-Stage2) and VPN-DLab (VPN-DeepLab) models.

Page 3: Supplementary Material for Video Propagation Networksopenaccess.thecvf.com/content_cvpr_2017/... · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep

Frame 5 Frame 15 Frame 30 Frame 50

Input

Vid

eo

GT

BV

SO

FL

VP

NV

PN

-DLab

Frame 5 Frame 10 Frame 20 Frame 30

Input

Vid

eo

GT

BV

SO

FL

VP

NV

PN

-DLab

Figure 2. Video Object Segmentation. Shown are the different frames in example videos with the corresponding ground truth (GT) masks,predictions from BVS [2], OFL [4], VPN (VPN-Stage2) and VPN-DLab (VPN-DeepLab) models.

Page 4: Supplementary Material for Video Propagation Networksopenaccess.thecvf.com/content_cvpr_2017/... · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep

Frame 4 Frame 14 Frame 24 Frame 36

Input

Vid

eo

GT

BV

SO

FL

VP

NV

PN

-DLab

Frame 5 Frame 15 Frame 30 Frame 50

Input

Vid

eo

GT

BV

SO

FL

VP

NV

PN

-DLab

Figure 3. Failure Cases for Video Object Segmentation. Shown are the different frames in example videos with the corresponding groundtruth (GT) masks, predictions from BVS [2], OFL [4], VPN (VPN-Stage2) and VPN-DLab (VPN-DeepLab) models.

Page 5: Supplementary Material for Video Propagation Networksopenaccess.thecvf.com/content_cvpr_2017/... · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep

Input GT CNN +VPN(Ours)

Figure 4. Semantic Video Segmentation. Input video frames and the corresponding ground truth (GT) segmentation together with thepredictions of CNN [5] and with VPN-Flow.

Page 6: Supplementary Material for Video Propagation Networksopenaccess.thecvf.com/content_cvpr_2017/... · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep

Frame 2 Frame 7 Frame 13 Frame 19

Input

Vid

eo

GT

-Colo

rLevin

et

al.

VP

N(O

urs)

Frame 2 Frame 7 Frame 13 Frame 19

Input

Vid

eo

GT

-Colo

rLevin

et

al.

VP

N(O

urs)

Figure 5. Video Color Propagation. Input grayscale video frames and corresponding ground-truth (GT) color images together with colorpredictions of Levin et al. [1] and VPN-Stage1 models.

Page 7: Supplementary Material for Video Propagation Networksopenaccess.thecvf.com/content_cvpr_2017/... · Supplementary Material for Video Propagation Networks Varun Jampani 1, Raghudeep

References[1] A. Levin, D. Lischinski, and Y. Weiss. Colorization using optimization. ACM Transactions on Graphics (ToG), 23(3):689–694, 2004.

6[2] N. Marki, F. Perazzi, O. Wang, and A. Sorkine-Hornung. Bilateral space video segmentation. In Computer Vision and Pattern

Recognition, IEEE Conference on, pages 743–751, 2016. 2, 3, 4[3] S. R. Richter, V. Vineet, S. Roth, and V. Koltun. Playing for data: Ground truth from computer games. In European Conference on

Computer Vision, pages 102–118. Springer, 2016. 1[4] Y.-H. Tsai, M.-H. Yang, and M. J. Black. Video segmentation via object flow. In Computer Vision and Pattern Recognition, IEEE

Conference on, 2016. 2, 3, 4[5] F. Yu and V. Koltun. Multi-scale context aggregation by dilated convolutions. International Conference on Learning Representations,

2016. 1, 5