13
0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information. This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 1 Fine-grained Human-centric Tracklet Segmentation with Single Frame Supervision Si Liu, Guanghui Ren, Yao Sun, Jinqiao Wang, Changhu Wang, Bo Li, Shuicheng Yan Abstract—In this paper, we target at the Fine-grAined human-Centric Tracklet Segmentation (FACTS) problem, where 12 human parts, e.g., face, pants, left-leg, are segmented. To reduce the heavy and tedious labeling efforts, FACTS requires only one labeled frame per video during training. The small size of human parts and the labeling scarcity makes FACTS very challenging. Considering adjacent frames of videos are continuous and human usually do not change clothes in a short time, we explicitly consider the pixel-level and frame-level context in the proposed Temporal Context segmentation Network (TCNet). On the one hand, optical flow is on-line calculated to propagate the pixel-level segmentation results to neighboring frames. On the other hand, frame-level classification likelihood vectors are also propagated to nearby frames. By fully exploiting the pixel-level and frame- level context, TCNet indirectly uses the large amount of unlabeled frames during training and produces smooth segmentation results during inference. Experimental results on four video datasets show the superiority of TCNet over the state-of-the-arts. The newly annotated datasets can be downloaded via http://liusi-group.com/projects/FACTS for the further studies. Index Terms—Video object segmentation, human-centric, fine-grained, optical flow estimation. 1 I NTRODUCTION During recent years, increasing efforts are devoted to the video object segmentation task. The famous benchmarks include DAVIS 2016 [32], 2017 [35] and 2018 challenges [3], shown as Figure 1 (a). These challenges focus on the semi-supervised video object segmentation task, that is, the algorithm is given a video sequence and the mask of the objects in the first frame, and the output should be the masks of those objects in the rest of the frames. Recently, Xu et al. collect an even larger video object segmentation dataset [43] named as YouTube-VOS, shown as Figure 1 (b). However, in some real applications, object-level segmentation is insufficient. For example, to tackle the person re-identification [47], [10], person attribute pre- diction [23], [45], people search [16] and fashion syn- thesis [50] problems, human instance segmentation is too coarse. The more detailed, fine-grained human parts segmentation, e.g., face, upper clothes and pants, is required. In this paper, we tackle a rarely studied problem, namely Fine-grAined human-Centric Tracklet Seg- mentation (FACTS). As a starting point, we take the human parts segmentation as a concrete example of FACTS. As shown in Figure 1 (c), each person is labeled with 12 human parts, including face, left/right arm, bag, left/right shoes etc. Comparing with the Si Liu and Bo Li are with School of Computer Science and Engineer- ing, Beihang University, Email: {liusi,boli}@buaa.edu.cn. Guanghui Ren and Yao Sun are with Institute of Information Engineering, Chinese Academy of Science, E-mail: {renguanghui, sunyao}@iie.ac.cn. Jinqiao wang is with Institute of Automation, Chinese Academy of Sciences, E-mail:[email protected]. Changhu Wang is with Toutiao AI Lab, Email:[email protected]. Shuicheng Yan is with Qihoo 360 AI Institute, Beijing, China, Email:[email protected] (a) DAVIS (b) YouTube (c) Outd 2017 e-VOS door Fig. 1. Examples of the training data from DAVIS 2017 [35], Youtube-VOS [43] and our newly collected Outdoor dataset. Best viewed in color. labeling in Figure 1 (a) and (b) that both only seg- ment the human out of the background, annotating multiple human parts is much more time-consuming. Therefore, an affordable and applicable choice is to label one single frame per video. To sum up, FACTS is different with the video object segmentation tasks in two aspects. First, FACTS solves a more fine-grained, human-centric problem while DAVIS and YouTube- VOS only segment the coarser object masks. Second, FACTS needs only single frame supervision per video during training. Therefore, the heavy job of manual labeling is much alleviated. The FACTS is difficult in two aspects. Firstly, human parts are very small in the video frame. For example, in the Indoor dataset, the average width and height of the face are only 13 and 16 in the 241 × 121 frame. The human parts are often occluded, deformed and easily affected by illumination, which makes the FACTS even more difficult. Secondly, the labeling is

Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 1

Fine-grained Human-centric TrackletSegmentation with Single Frame Supervision

Si Liu, Guanghui Ren, Yao Sun, Jinqiao Wang, Changhu Wang, Bo Li, Shuicheng Yan

Abstract—In this paper, we target at the Fine-grAined human-Centric Tracklet Segmentation (FACTS) problem, where 12 humanparts, e.g., face, pants, left-leg, are segmented. To reduce the heavy and tedious labeling efforts, FACTS requires only onelabeled frame per video during training. The small size of human parts and the labeling scarcity makes FACTS very challenging.Considering adjacent frames of videos are continuous and human usually do not change clothes in a short time, we explicitlyconsider the pixel-level and frame-level context in the proposed Temporal Context segmentation Network (TCNet). On the onehand, optical flow is on-line calculated to propagate the pixel-level segmentation results to neighboring frames. On the other hand,frame-level classification likelihood vectors are also propagated to nearby frames. By fully exploiting the pixel-level and frame-level context, TCNet indirectly uses the large amount of unlabeled frames during training and produces smooth segmentationresults during inference. Experimental results on four video datasets show the superiority of TCNet over the state-of-the-arts.The newly annotated datasets can be downloaded via http://liusi-group.com/projects/FACTS for the further studies.

Index Terms—Video object segmentation, human-centric, fine-grained, optical flow estimation.

F

1 INTRODUCTION

During recent years, increasing efforts are devotedto the video object segmentation task. The famousbenchmarks include DAVIS 2016 [32], 2017 [35] and2018 challenges [3], shown as Figure 1 (a). Thesechallenges focus on the semi-supervised video objectsegmentation task, that is, the algorithm is given avideo sequence and the mask of the objects in thefirst frame, and the output should be the masks ofthose objects in the rest of the frames. Recently, Xuet al. collect an even larger video object segmentationdataset [43] named as YouTube-VOS, shown as Figure1 (b). However, in some real applications, object-levelsegmentation is insufficient. For example, to tackle theperson re-identification [47], [10], person attribute pre-diction [23], [45], people search [16] and fashion syn-thesis [50] problems, human instance segmentation istoo coarse. The more detailed, fine-grained human partssegmentation, e.g., face, upper clothes and pants, isrequired.

In this paper, we tackle a rarely studied problem,namely Fine-grAined human-Centric Tracklet Seg-mentation (FACTS). As a starting point, we take thehuman parts segmentation as a concrete example ofFACTS. As shown in Figure 1 (c), each person islabeled with 12 human parts, including face, left/rightarm, bag, left/right shoes etc. Comparing with the

Si Liu and Bo Li are with School of Computer Science and Engineer-ing, Beihang University, Email: {liusi,boli}@buaa.edu.cn. Guanghui Renand Yao Sun are with Institute of Information Engineering, ChineseAcademy of Science, E-mail: {renguanghui, sunyao}@iie.ac.cn. Jinqiaowang is with Institute of Automation, Chinese Academy of Sciences,E-mail:[email protected]. Changhu Wang is with Toutiao AI Lab,Email:[email protected]. Shuicheng Yan is with Qihoo 360 AIInstitute, Beijing, China, Email:[email protected]

(a) DAVIS

(b) YouTube

(c) Outd

2017

e-VOS

door

Fig. 1. Examples of the training data from DAVIS2017 [35], Youtube-VOS [43] and our newly collectedOutdoor dataset. Best viewed in color.

labeling in Figure 1 (a) and (b) that both only seg-ment the human out of the background, annotatingmultiple human parts is much more time-consuming.Therefore, an affordable and applicable choice is tolabel one single frame per video. To sum up, FACTSis different with the video object segmentation tasksin two aspects. First, FACTS solves a more fine-grained,human-centric problem while DAVIS and YouTube-VOS only segment the coarser object masks. Second,FACTS needs only single frame supervision per videoduring training. Therefore, the heavy job of manuallabeling is much alleviated.

The FACTS is difficult in two aspects. Firstly, humanparts are very small in the video frame. For example,in the Indoor dataset, the average width and heightof the face are only 13 and 16 in the 241 × 121frame. The human parts are often occluded, deformedand easily affected by illumination, which makes theFACTS even more difficult. Secondly, the labeling is

Page 2: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 2

(i) Pixel-level Context

(ii)Frame-level Context

Final Result

bag

No dress

𝑂𝑡,𝑡−𝑙 𝑂𝑡,𝑡−𝑠

𝑃𝑡−𝑙𝑏𝑎𝑔

𝑃𝑡−𝑠𝑏𝑎𝑔 𝑃𝑡

𝑏𝑎𝑔

𝑃𝑡ba𝑔

𝐼𝑡−𝑙 𝐼𝑡−𝑠 𝐼𝑡

… … … …

𝑓𝑡−𝑙 𝑓𝑡−𝑠 𝑓𝑡 𝐹𝑡

Fig. 2. During testing, a segmentation window is slided along the video. The parsing result of frame It isdetermined by current frame It and two historical frames It−l and It−s.

insufficient. For example, Indoor dataset contains 400training tracklets. In each tracklet, only 1 frame islabeled. Therefore, we have only 400 labeled framesto learn the segmentation model for 13 fine-grainedcategories (12 human parts and “background”). Howto effectively leverage the large amount of unlabeledvideo frames is the key.

To tackle these difficulties, we propose a TemporalContext segmentation Network (TCNet), shown inFigure 2. For each tracklet, the human-centric trackletis produced by applying Faster R-CNN [36] based hu-man detection to the first frame, and Kernelized Cor-relation Filters [11] based human tracking to the restframes. The detected/tracked human-centric bound-ing boxes are then enlarged to 241× 121 to feed intothe network. The preprocessing partially addressesthe challenge of parts being small. To further solvethe two challenges, TCNet fully explores the rich tem-poral context among the video. To segment the frameIt, both It and two historical frames It−l and It−sare considered, where l > s. The triplet {It, It−l, It−s}is fed into the network to collaboratively producethe segmentation result. More specifically, two kindsof context are explicitly considered. 1) Pixel-levelcontext: adjacent frames are similar, therefore theneighboring frames can help segment one particularframe. Here {P bag

t−l , Pbagt−s , P

bagt } are the segmentation

confidence of “bag”. The bag is not obvious in P bagt

due to the dark lighting. However, both P bagt−l and

P bagt−s have strong responses for “bag”, which serves as

the pixel-wise context. Therefore, “bag” can be easilyidentified in the refined segmentation result P bag

t

thus the missed detection is reduced. 2) Frame-level

context: one won’t change the clothes and accessoriesin a short time. Therefore, we regularize the categoriesto be basically unchanged for all frames of the tracklet.Technically, for each frame, we estimate the likelihoodvector whose dimension is the number of categories.The vectors {Ft−l, Ft−s, Ft} together vote for the finalconfidence vector Ft. Because of the interference of thedark lighting and particular human pose, the “pants”is misclassified as “dress” in Ft. The likelihood of“dress” is denoted by the green solid bin. By con-sidering the frame-level context, the correspondinglikelihood is suppressed thus false alarm is reduced.Extensive experiments on four datasets show that,although the human part is small and labeling isscare, TCNet achieves state-of-the-art performance bymaking full use of both pixel-level and frame-levelcontext.

The contributions of our work are listed as follows:

1) To the best of our knowledge, it is the firstattempt to explore FACTS task by labeling onesingle frame per video. It has extensive applica-tion prospects.

2) The proposed TCNet fully explores the pixel-level and frame-level context in the video to useboth the unlabeled and labeled frames. The pro-posed framework is end-to-end and thus veryapplicable for practical usage.

3) We have released the 4 newly collected datasets,including Indoor, Outdoor, iLIDS-Parsing andDaily to the public via http://liusi-group.com/projects/FACTS.

An earlier version of this manuscript appeared inCVPR 2017 [24]. In this version we have refined the

Page 3: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 3

pixel-level context by the unsupervised optical flowrefinement, added the frame-level context propaga-tion module, conducted extensive experiments on twomore datasets and presented more in-depth discus-sion about the experimental results.

2 RELATED WORK

FACTS is relevant with several tasks, including videoobject segmentation, video semantic segmentation,fine-grained segmentation as well as video analysiswith sparse annotations.

Video Object Segmentation: DAVIS 2016 [32] is asingle object segmentation challenge. Khoreva et al.[33] propose the MaskTrack which combines an offlinemask refinement model and an online model. Caelleset al. [2] propose OSVOS to use the generic semanticinformation learned on ImageNet. MaskTrack and OS-VOS are the top-performing algorithms in the DAVIS2016 challenge. Both DAVIS 2017 [35] and 2018 [3] aremultiple objects segmentation challenges. Khoreva etal. [15] proposes Lucid Tracker which generates in-domain training data using lucid dream. It reachescompetitive results without pre-training. Luiten et al.[26] present the PReMVOS which first generates somemask proposals for each video frame and then selectsand merges these proposals into accurate and tempo-rally consistent object tracks. Li et al. propose VSReID[18] and DyeNet [17] which combines both temporalpropagation and re-identification functionalities into asingle framework. Lucid Tracker, VSReID, PReMVOSand DyeNet are among the best performing algo-rithms in the DAVIS 2017/2018 challenges.

Although great breakthroughs have been made inthe video object segmentation field, their results aretoo coarse in some real applications. Moreover, theyrequire dense pixel-wise annotation. On the contrary,the results of FACTS are more detailed and fine-grained. In addition, much manual labeling efforts aresaved.

Video Semantic Segmentation: Existing video seg-mentation methods can be mainly classified into twokinds. The first kind of methods target at high seg-mentation accuracies. Representative methods includespatio-temporal FCN (STFCN) [7], Spatio-TemporalTransformer Gated Recurrent Unit (STGRU) [30],Parsing with prEdictive feAtuRe Learning (PEARL)[14], NetWarp [9] and Adaptive Temporal EncodingNetwork (ATEN) [48]. The second kind of methodsfocus on speed acceleration, including clockwork FCN[37], Deep Feature Flow [52], budget-aware deepsemantic video segmentation [29] and Low-LatencyVideo Semantic Segmentation [19].

These video semantic segmentation methods do nottrack objects across the video while FACTS focuses onsegmenting the region inside the tracklet and set thepixels outside the tracking box as background.

Fine-grained Segmentation: Compared with thecoarse object/instance segmentation, object part seg-mentation is much more fine-grained. Representa-tive works are conducted on the Horse-Cow parsingdataset [39] and the PASCAL Quadrupeds dataset [5].Wang et al. explore how to jointly achieve object andpart segmentation using deep learned potentials [40].In recent years, human part segmentation [44], [27]has received increasing attention. Most human pars-ing methods target at parsing human-centric images,such as the fashion photos [20] or the daily photos[21], [42], [13].

These part segmentation works are not speciallydesigned for the FACTS task, and cannot address thechallenge of training data insufficiency.

Video Analysis With Sparse Annotations: Someother video analysis tasks, such as video pose esti-mation and video semantic segmentation also sufferfrom labeling difficulty. Pfister et al. [34] investigatea video pose estimation architecture that is able tobenefit from temporal context by combining informa-tion across the multiple frames using optical flow. Zhuet al. [51] propose flow-guided feature aggregation(FGFA) to leverage temporal coherence on featurelevel to improve the video object detection accuracy.

Although these methods utilize optical flow to re-fine the results, they do not mine the frame-levelcontext, which is proven to greatly increase the seg-mentation accuracy. In addition, these methods ad-dress the video pose estimation or video segmentationproblems, instead of the FACTS problem.

3 APPROACH

3.1 FrameworkAs mentioned, one of the main challenges of theFACTS task lies in the insufficient training data. Con-sidering the video frames are continuous and redun-dant, we resort to the temporal context to take advan-tages of the large amount of unlabeled video frames.Therefore, FACTS is tackled in a semi-supervisedlearning manner. The key is how to build the inter-frame correlations (context) between labeled and unla-beled frames. Basically, we mainly exploit two kindsof context. On the one hand, the pixel-level context(Section 3.2) builds the pixel-wise correspondencesbetween labeled and unlabeled frames. On the otherhand, the frame-level context (Section 3.3) guaranteesthe smoothness of the classification results of adjacentframes. The two kinds of context are complementarywith each other. The pixel-wise context is local whilethe frame-level context is global.

The framework of TCNet is shown in Figure 3 ,which consists of pixel-level context and frame-levelcontext. The input is the triplet {It−l, It−s, It}, amongwhich only It is labeled during training. The long-range contextual frame It−l and short-range contex-tual frame It−s are l and s frames temporally ahead

Page 4: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 4

Optical FlowEstimation

𝐶𝑡,𝑡−𝑙

𝐼𝑡−𝑙 𝐼𝑡𝐼𝑡−𝑠

𝑂𝑡,𝑡−𝑙 𝑂𝑡,𝑡−𝑠

Warp

𝐶𝑡,𝑡−𝑠

OutputInputdress baghair

. . . 0.92 0.10 0.32

. . . 0.95 0.06 0.56

. . . 0.93 0.25 0.16

Pixel-level Fusion

Fusion

Frame Parsing

FrameClassification

FeatureExtraction

𝑓𝑡−𝑙

𝑃𝑡−𝑠 𝑃𝑡

Frame-level Fusion

ConfidenceEstimation

𝑃𝑡−𝑙

𝑓𝑡−𝑠

𝑓𝑡

𝐹𝑡

𝑃(𝑡−𝑙)→𝑡 𝑃(𝑡−𝑠)→𝑡𝑃𝑡

෨𝑃𝑡

Fig. 3. The architecture of the proposed Temporal Context segmentation Network.

of It. The values of l and s are set empirically. Duringtesting phase, no manual annotation is needed. Theoutput of TCNet is the parsing result of It. First,the triplet images {It−l, It−s, It} are fed into featureextraction module. Then three feature maps are fedinto several parallel modules. (i) The frame parsingmodule estimates three rough labelmaps for the triplet,namely {Pt−l,Pt−s,Pt}. (ii) The frame classificationmodule produces the frame-level likelihood vectors{ft−l, ft−s, ft}. (iii) The optical flow estimation moduleestimates the optical flow Ot,t−l and Ot,t−s. More-over, the confidence of optical flow Ct,t−l and Ct,t−sare simultaneously estimated. Afterwards, the frameparsing results, the estimated optical flows as wellas the confidences are fed into the pixel-level fusionmodule to produce a better segmentation result Pt.In addition, the estimated frame classification resultsvote for a refined Ft via the frame-level fusion module.Finally, Pt and Ft are pixel-wisely added to producethe final parsing result Pt.

The optical flow estimation and frame parsingmodules are able to share the feature maps for thefollowing reasons. Firstly, the two modules are se-mantically relevant. Only pixels of the same labels canbe matched by optical flow. Secondly, both networksmake pixelwise predictions. Frame parsing estimatespixelwise category while optical flow is the pixelwiseoffset/shift. It provides a prerequisite for feature shar-ing. Another advantage of sharing is much compu-tation can be saved. Any architecture for semanticsegmentation, e.g, [46], [4] can be used for frameparsing module. In our experiments, DeepLab v2 [4]based on the VGG [38] architecture is selected.

3.2 Pixel-level Context PropagationSegmenting human parts frame by frame is difficultdue to small size, the illumination, human pose orcamera viewpoint issues. Because videos are continu-ous, the segmentation result of one frame is expectedto be refined by its nearby frames. In order to build thepixel-wise correspondences between frames, opticalflow Oa,b : R2 → R2 where R is the rational fieldis calculated. The flow field Op

a,b = (qx − px, qy − py)computes the relative offsets from each point p inimage Ia to a corresponding point q in image Ib. TheOt,t−l is the optical flow between It, and It−l. Ot,t−sis estimated similarly.

One feasible approach is to off-line calculate theoptical flow via any off-the-shelf method [1] and loadthem into the network during optimization. It makestraining and testing be a multi-stage pipeline, andthus very expensive in space and time. To this end, wecompute the optical flow in an on-line manner. Afterthe shared feature extraction module, a “correlationlayer” which is used in the Flownet and its successors[8], [28], performs multiplicative patch comparisonsbetween two feature maps. After that, several “upcon-volutional” layers are used to obtain the optical flowwith the same resolution as the input image pairs.Since our videos have no groundtruth optical flow,we use the Flying Chairs dataset [8] for training.

Unsupervised Optical Flow Refinement: Becausethe Flying Chairs dataset is synthetic and quite dif-ferent from the real-world videos, we propose torefine the optical flow obtained using Flying Chairsby minimizing an appearance reconstruction loss [49].

Given the image pair {It−l, It}, the ideal optical

Page 5: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 5

𝐶𝑡,𝑡−𝑙

𝑃𝑡

𝐶𝑡,𝑡−𝑠

conv1x1 𝑃𝑡

𝐹𝑡

෨𝑃𝑡

weighted sum reshape

concatconcat

𝑃(𝑡−𝑙)→𝑡

𝑃(𝑡−𝑠)→𝑡

conv1x1

𝑓𝑡−𝑙

𝑓𝑡

𝑓𝑡−𝑠

መ𝑓𝑡

𝑤𝑡−𝑙

𝑤𝑡−𝑠

𝑤𝑡

Pixel-level Fusion

Frame-level Fusion

.

.

Fig. 4. Details of pixel-level fusion and frame-level fusion. �,+ respectively denote element-wise product andelement-wise add.

flow Ot,t−l should perfectly match one image to theother. Mathematically, the appearance reconstructionloss L(It−l, It, Ot,t−l) should be minimized:

L(It−l, It, Ot,t−l) =∥∥∥Iit − Iit

∥∥∥1=∥∥Iit − wi (It−l, Ot,t−l)

∥∥1,

(1)where ‖·‖1 denotes the L1 norm. Iit is the wrappedcounterpart of Iit , and w (a, b) is the operation ofapplying the estimated optical flow b to warp im-age a. The coordinate of pixel i in It is

(xi, yi

),

while the mapped coordinate in It−l is(xi

′, yi

′)

=(xi, yi

)+ Oi

t,t−l. When(xi

′, yi

′)

falls into sub-pixel

coordinate, we rewrite the Iit of Equation 1 via bilinearinterpolation:

Iit = wi (It−l, Ot,t−l)

=∑

q∈{neighbors of (xi′ ,yi′ )}Iqt−l(1−

∣∣∣xi′ − xq∣∣∣)(1− ∣∣∣yi′ − yq

∣∣∣),(2)

where q denotes the 4-pixel neighbors (top-left, top-right, bottom-left, bottom-right) of

(xi

′, yi

′)

.Note that optimizing Equation 1 does not require

labeling pixel-wise correspondence which is indis-pensable in traditional optical flow estimation models[1]. Therefore, it is very applicable in real applications.

Optical Flow Confidence Estimation: The opticalflow estimated after the unsupervised refinement isstill not perfect. Therefore, we estimate the confidenceof the estimated optical flow Ot,t−l. The distance ofwarping Di

t,t−l is defined as:

Dit,t−l =

∥∥Iit − wi (It−l, Ot,t−l)∥∥1, (3)

the flow Ot,t−s can be handled in similar manners.The confidence defined in Equation 3 is the distance

between the original image and its warped counter-

part. The normalized confidence is calculated via:

Cit,t−l = exp(−Di

t,t−l/(2σ2)), (4)

where σ is the standard deviation of Dt−l,t. Highervalue indicates more confident optical flow estima-tion. As shown in Figure 3, the estimated parsingresults P(t−l)→t and P(t−s)→t are obtained by warpingPt−l and Pt−s according to the optical flow Ot,t−l andOt,t−s via:

P(t−l)→t = w(Pt−l, Ot,t−l),P(t−s)→t = w(Pt−s, Ot,t−s).

(5)

Pixel-level Fusion: As shown in the upper panel ofFigure 4, P(t−l)→t and P(t−s)→t are further weightedby the confidence map of Equation 4 to reduce the in-fluence of inaccurate optical flow by: P(t−l)→t�Ct,t−land P(t−s)→t � Ct,t−s, where ‘�’ denotes element-wise product. They are concatenated with Pt andthen fused with several 1 × 1 convolutional layersto produce Pt, which is the final result of pixel-levelcontext propagation.

3.3 Frame-level Context PropagationHuman may not change clothes and accessories ina very short time, therefore, we use the historicalframe-level classification likelihood vectors to refinethe categories in one particular frame. For example, ifthe human is predicted as wearing “pant” in the pastseveral frames but predicted as wearing “dress” in anew frame, the “dress” is very likely to be wrong.

As shown in Figure 3, the frame classification mod-ule builds upon the feature extraction modules. Itgenerates the likelihood vectors f , whose element in-dicates the existence of one particular label, e.g., bag.The likelihood vectors {ft−l, ft−s, ft} are weighted

Page 6: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 6

by [wt−l, wt−s, wt] to produce the smoother ft. ft isreshaped, duplicated to produce a M ×N ×K tensorFt whose values in each channel are identical. Then,Ft is added pixel-wisely with Pt to produce the finalsegmentation results Pt. More details about the frame-level context propagation is illustrated in the lowerpanel of Figure 4.

3.4 Loss Functions & OptimizationIn this section, we will first introduce our overallloss functions and then elaborate the training andinference strategies.

Two kinds of losses, i.e., segmentation and clas-sification losses, are applied on the labeled frameIt. The unlabeled frames are not supervised. Thesegmentation loss Lseg is the sum of cross-entropyterms for each spatial position in Pt. All positions andlabels are equally weighted in the segmentation lossterm. Mathematically, it is defined as:

Lseg(Pt, Gt) =∑i∈It

cross entropy(P it , G

it) (6)

where Gt is the groundtruth labelmap of the frameIt, i is the coordinate index of the image It. Moreover,we add a classification loss Lcls of frame classificationthat is binary cross-entropy loss for each category.Mathematically, it is defined as:

Lcls =∑c∈C

cross entropy(f ct , g

ct

)(7)

where ft and gct are the estimated and groundtruthvector of category c for the frame It. gct is obtainedby summarizing Gt. When any pixel of It is labeledas class c, gct is set as 1. Otherwise, gct is 0. C is thecategory set. Thus, the overall loss function is:

L = Lseg + Lcls (8)

TCNet is optimized by 2-step alternating trainingstrategy. (i) We train both the feature extraction andoptical flow estimation modules via the strategiesin Section 3.2. The optical flow estimation moduleis initialized via Gaussian distribution, and trainedfirstly by Flying chairs dataset [8], and then fine-tunedby the proposed unsupervised optical flow refinementmethod. The feature extraction module is initializedwith VGG model [38] trained on Pascal VOC dataset.(ii) We train all components together in an end-to-end manner. Both pixel-level and frame-level fusioncomponent are initialized via standard Gaussian dis-tribution (with zero mean and unit variance).

During inference, we slide a segmentation windowalong the video to specifically consider the temporalcontext. The parsing result of It is jointly determinedby {It−l, It−s, It}. Note that It−l and It−s are earlierframes whose feature maps and parsing results arestored in the memory buffer. Therefore, the extra

computational cost brought by considering the seg-mentation window is negligible. Note that becausethe first l frames of a video do not have enoughpreceding frames to form a sliding parsing window,only the frame parsing module is applied to producetheir segmentation results.

Datasets Training Validation TestingIndoor [24] 400 500 1000

Outdoor [24] 421 582 1000iLIDS-Parsing [41] 100 285 500

Daily 244 120 670

TABLE 1The number of labeled images of the four datasets.

4 EXPERIMENTS

4.1 Experimental settingDataset Collection & Labeling: Experiments are con-ducted on four datasets. To produce human-centrictracklet, persons are first detected via Faster R-CNN[36] fine-tuned on VOC dataset [6] and then trackedby KCF[11]. The statistical data of these four datasetare summarized in Table 1. Representative videoframes are shown in Figure 5.

1) The Indoor dataset contains 700 videos, amongwhich 400 videos, 100 videos and 200 videosare used as training set, validation set and test-ing set, respectively. This dataset contains 13categories (12 human parts and “background”),namely face, hair, upper-clothes, left-arm, right-arm, pants, left-leg, right-leg, left-shoe, right-shoe, bag, dress, and background. We randomlyselect and pixel-wisely label 1 frame from eachtraining video. For each validation and testingvideo, we randomly label 5 frames.

2) The Outdoor dataset contains 421 trainingvideos, 120 validation videos and 200 testingvideos. The categories and configurations are thesame as the Indoor dataset.

3) We also labeled several frames from iLIDS Videore-IDentification (iLIDS-VID) dataset 1 [41]. Thenew labeled set is called iLIDS-Parsing dataset.Some categories, e.g. left-leg, right-leg and dress,seldom appear in the dataset since this datasetis taken in winter and people all wear pants.

4) The Daily dataset is a newly collected datasetfrom several video sharing website, such asyouku.com. A variety of scenes such as shop,road, are included. The categories and configu-rations are same as the Indoor dataset.

Evaluation Metrics We use the same metrics asPaperDoll [44] to evaluate the performance. Amongall evaluation metrics, the average F-1 is the mostimportant. We train TCNet via Caffe [12] using Nvidia

1. We have obtained the approval of the database publisher.

Page 7: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 7

Indoor dataset Outdoor dataset

iLIDS-Parsing dataset Daily dataset

Fig. 5. Images in the four datasets.

GeForce GTX 1080 Ti. The initial learning rates forframe parsing and optical flow estimation modulesare 1e-8 and 1e-5 respectively. The long range l andshort range s are empirically set as 3 and 1 in theindoor dataset. Because the outdoor dataset has alower frame rate and contains more quick dynamics, land s are set to 2 and 1. The settings in iLIDS-Parsingand Daily dataset are the same as the Indoor dataset.

4.2 Comparison with state-of-the-artWe compare our results with seven state-of-the-artmethods. The 1st is PaperDoll [44]. It is a tradi-tional human parsing method. The 2nd is ATR [20]which formulates the human parsing task as an activetemplate regression problem. The 3rd baseline is M-CNN [22], which is a quasi-parametric human parsingmethod. The 4th is Co-CNN [21] which uses a Con-textualized Convolutional Neural Network to tacklethe problem. The 5th is FCN-8s [25] which achievescompetitive results in several semantic segmentationbenchmark datasets. The 6th baseline is DeepLab V2[4]. The above mentioned six methods are supervisedalgorithms. Therefore, only the labeled frames areused for training. The 7th baseline is EM-Adapt 2 [31]which treats image-level annotation as weak super-vision. For each tracklet, its label can be obtained bysummarizing the pixel-level annotations of the labeledframe into a label vector. Thus all unlabeled framesof the tracklet share the same image-level annotation.The 8th baseline is NetWarp [9] which uses opticalflow of adjacent frames for warping internal networkrepresentations across time. To be fair, the backbone ofNetWarp is kept the same as ours (DeepLab v2 basedon the VGG architecture). We add the NetWarp mod-ule to Conv5 3. Moreover, two versions of NetWarpare implemented by utilizing two kinds of optical flowmethods. One is NetWarp-DIS which uses the DISflow in their original paper [9], the other is NetWarp-Flownet which adopts Flownet [8]. Note that our TC-Net is different from NetWarp in that our optical flow

2. http://liangchiehchen.com/projects/DeepLab-LargeFOV-Semi-EM-Fixed.html.

is online calculated, while NetWarp uses the offlineoptical flow. We also consider different variants ofTCNet. “l” or “s” means using long or short rangenearby frames for pixel-level context propagation,while “c” means the confidence is used in the fusion ofwarped optical flows. The techniques correspondingto “l”, “s”, and “c” have been discussed in [24]. “f”refers the method of using frame-level context, and“uf” means using the unsupervised fine-tuned opticalflow.

Methods Acc. Fg.acc. Avg.pre. Avg.rec. Avg.F-1PaperDoll [44] 46.71 78.69 33.55 45.68 36.41ATR [20] 85.69 71.24 47.39 53.21 49.14M-CNN [22] 85.19 71.31 42.9 50.11 45.68Co-CNN [21] 87.58 72.58 53.54 51.87 51.38FCN-8s [25] 88.33 71.56 55.05 52.15 53.12DeepLab [4] 86.88 77.45 49.88 64.3 54.89EM-Adapt [31] 86.63 80.88 53.01 63.64 56.40NetWarp-DIS [9] 88.52 75.21 55.05 60.18 57.02NetWarp-Flownet [9] 88.66 75.68 55.35 60.82 57.59TCNet l 88.81 74.42 56.28 59.81 57.74TCNet l+s 88.91 77.12 55.9 61.21 58.04TCNet l+c 88.75 77.28 56.07 61.94 58.43TCNet s+c 89.07 77.06 56.86 61.98 58.73TCNet l+s 88.85 78.68 56.77 62.73 59.21TCNet l+s+c 89.88 76.48 61.52 59.38 60.20TCNet l+s+c+f 90.11 79.92 60.84 64.63 62.25TCNet l+s+c+uf 89.99 79.81 59.64 68.14 62.71

TABLE 2Comparison with state-of-the-arts and several variants

of TCNet in Indoor dataset. (%).

Indoor dataset: Table 2 shows the comparisons be-tween TCNet and 7 state-of-the-art methods in theIndoor dataset. Different variants of TCNet are gener-ated by gradually adding more components, whichwill be discussed in the next subsection. It can beseen that “TCNet l+s+c+uf” reaches the average F-1score of 62.71%, which is superior than all baselines,even better than the score 60.20% in [24]. The 1st tothe 6th baselines all use labeled images. Therefore,the improvements show the advantage of utilizingthe unlabeled dataset. EM-Adapt, NetWarp-DIS andNetWarp-Flownet also use unlabeled images, andthus reach a higher F1-score comparing with otherbaselines. Especially, NetWarp-Flownet reaches a F1-score of 57.59%, which is better than other baselines.However, NetWarp-Flownet is still worse than allthe variants of TCNet. It shows that pixel-level con-text is helpful in the FACTS task. The frame-levelcontext and refined optical flow estimation also im-prove the performance. The F1-scores of each categoryare shown in Table 3. We can observe that “TCNetl+s+c+uf” beats all baselines in 11 categories, whichagain shows the big improvements brought by theproposed TCNet.Outdoor dataset: Among all the baselines, EM-Adapt,NetWarp-DIS and NetWarp-Flownet show superiorperformances. Therefore, we only compare with the

Page 8: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 8

Methods bk face hair U-clothes

L-arm

R-arm pants L-

legR-leg Dress L-

shoeR-

shoe bag

PaperDoll [44] 92.62 57.16 58.22 62.52 19.96 14.99 52.47 25.43 20.7 9.92 20.66 24.41 14.32ATR [20] 93.62 59.08 60.79 81.36 32.54 28.65 75.40 29.19 29.60 70.22 11.68 17.75 48.97M-CNN [22] 93.40 53.94 59.12 75.53 24.46 20.51 78.46 36.15 21.92 43.61 14.53 18.79 53.43Co-CNN [21] 94.06 64.64 73.53 81.54 26.82 31.66 77.13 25.47 34.11 76.08 15.42 20.57 46.91FCN-8s [25] 94.80 71.35 74.90 79.53 33.55 32.29 81.89 36.57 33.98 43.53 33.03 31.50 43.66DeepLab [4] 93.64 63.01 69.61 81.54 40.97 40.31 81.12 34.25 33.24 64.60 28.39 26.40 56.50EM-Adapt [31] 93.46 66.54 70.54 77.72 42.95 42.20 82.19 39.42 37.19 63.22 33.18 31.68 53.00NetWarp-DIS [9] 94.64 66.09 71.59 81.80 44.18 43.45 82.91 37.13 39.10 61.93 32.01 30.03 56.33NetWarp-Flownet [9] 94.70 66.46 71.77 81.94 44.84 44.14 83.16 37.58 39.76 63.24 33.29 31.27 56.53TCNet l 94.68 67.28 72.74 82.12 42.96 43.35 81.91 39.26 38.31 67.17 31.47 30.38 58.99TCNet s 94.65 66.27 73.48 83.12 45.17 44.89 82.72 38.62 38.43 66.04 30.93 31.46 58.81TCNet l+c 94.44 67.29 73.76 83.06 43.56 43.56 82.33 41.36 39.46 68.36 31.75 31.73 59.04TCNet s+c 94.64 67.62 74.13 83.48 45.13 45.08 83.21 39.89 40.11 68.17 31.15 32.27 58.75TCNet l+s 94.50 67.08 73.52 83.10 45.51 44.26 82.59 41.82 42.31 69.43 33.71 33.36 58.58TCNet l+s+c 94.89 70.28 76.75 84.18 44.79 43.29 83.59 42.69 40.30 70.76 34.77 35.81 60.43TCNet l+s+c+f 95.22 68.30 77.04 84.44 50.82 50.20 84.29 45.92 45.86 71.33 36.70 38.13 61.06TCNet l+s+c+uf 95.23 67.09 77.64 84.78 51.16 50.85 84.60 46.53 46.98 67.96 39.30 40.55 62.62

TABLE 3Per-Class Comparison of F-1 scores with state-of-the-arts and several architectural variants of TCNet in Indoor

dataset. (%).

Methods Acc. Fg.acc. Avg.pre. Avg.rec. Avg.F-1EM-Adapt [31] 84.37 66.77 48.95 50.29 48.93NetWarp-DIS [9] 84.91 65.26 50.21 49.54 49.10NetWarp-Flownet [9] 84.84 65.67 49.64 50.69 49.49TCNet l+s+c 85.73 67.19 52.64 50.93 50.70TCNet l+s+c+f 85.06 70.01 52.06 52.64 51.43TCNet l+s+c+uf 85.76 68.95 51.94 54.81 52.94

TABLE 4Comparison with state-of-the-arts and several variants

of TCNet in Outdoor dataset. (%).

EM-Adapt, NetWarp-DIS and NetWarp-Flownet inthe Outdoor dataset. Table 4 shows the results. Itcan be seen that our method reaches the average F-1score of 52.94% while NetWarp-Flownet only reaches49.49%. We find that the performances of Table 4 aregenerally lower than that of Table 2. The reason isthat the outdoor dataset set has a relative lower framerate, so the changes in nearby frames are bigger thanthose in Indoor dataset, which makes the optical flowestimation difficult. The F1-scores of each category areshown in Table 5. Again, “TCNet l+s+c+uf” performsthe best in most cases.iLIDS-Parsing dataset: Similarly, we only comparethe EM-Adapt, NetWarp-DIS and NetWarp-Flownetmethod with the variants of TCNet. Table 6 showsthe results. The overall performance in iLIDS-Parsingdataset are poor, because there are only 100 imagesfor training, which is much less than the images usedfor training in Indoor and Outdoor datasets. Even so,the method “TCNet l+s+c+uf” beats the NetWarp-Flownet in the average F-1 score by 1.85%. Table4.2 shows the detailed F1-scores for each category.Note that the scores for “L/R-arm”, “L/R-leg”, and“Dress” are very low. That’s because the images in

iLIDS-Parsing dataset are taken in winter, the trainingsamples are extremely insufficient. Still, the method“TCNet l+s+c+f” and “TCNet l+s+c+uf” win in mostcategories.Daily dataset: The overall comparison is shown inTable 8 and the detailed F1 scores are illustrated inTable 9. The method “TCNet l+s+c+uf” performs bestin 11 categories. Although the number of labeledimages for training is only one half of those in Indoordataset, the average F-1 score is high. We believe thisis because the images in the Daily dataset are muchclearer than those in the previous datasets. This isbetter illustrated in Figure 5.

4.3 Component Analysis

Image A OF [24]Image B Warped B [24]Refined OF Warped B (Refined)

Fig. 6. Comparisons of optical flow with and withoutrefinement. The optical flow warps Image B to ImageA. The warped image using the refined optical flow aremore similar with Image A.Unsupervised optical flow refinement: The opticalflow estimation is improved by using the unsuper-vised optical flow refinement technique introduced

Page 9: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 9

Methods bk face hair U-clothes

L-arm

R-arm pants L-

legR-leg Dress L-

shoeR-

shoe bag

EM-Adapt [31] 93.71 65.05 63.56 77.80 38.16 37.24 75.99 36.95 35.48 18.59 33.72 31.10 28.81NetWarp-DIS [9] 94.22 64.94 64.72 76.45 38.56 39.05 76.64 38.54 37.80 18.32 35.43 25.98 27.66NetWarp-Flownet [9] 94.20 65.50 64.24 76.93 37.44 39.93 76.43 38.65 38.41 17.78 35.22 24.73 33.94TCNet l+s+c 94.24 67.80 66.89 79.50 41.74 43.00 78.03 40.00 36.85 23.24 32.87 25.85 29.07TCNet l+s+c+f 93.78 67.98 68.22 78.78 44.44 44.83 77.27 39.63 39.54 21.75 33.03 29.59 29.82TCNet l+s+c+uf 94.44 68.40 68.85 79.62 43.52 43.64 78.83 41.82 38.74 25.67 35.54 32.41 36.79

TABLE 5Per-Class Comparison of F-1 scores with state-of-the-arts and several architectural variants of TCNet in

Outdoor dataset. (%).

Methods Acc. Fg.acc. Avg.pre. Avg.rec. Avg.F-1EM-Adapt [31] 85.26 68.59 44.11 35.40 36.04NetWarp-DIS [9] 84.97 70.66 44.22 36.67 36.15NetWarp-Flownet [9] 85.29 71.09 44.85 37.01 36.83TCNet l+s+c 84.97 77.53 47.87 38.39 36.66TCNet l+s+c+f 84.48 75.33 45.81 38.87 37.82TCNet l+s+c+uf 85.91 76.91 47.95 39.55 38.68

TABLE 6Comparison with state-of-the-arts and several variants

of TCNet in iLIDS-Parsing dataset. (%).

in Section 3.2. As shown in Figure 6, the last twocolumns are the wrapped images using optical flow[24] and refined optical flow. In the first row, therefined optical flow puts the left and right legs closerin the warped image, which is more similar withImage A. In the second row, the warped result usingrefined optical flow can change the human pose to bemore similar with Image A, e.g., the man’s right legis bending and his right heel is lifting.

Quantitative results in the Indoor, Outdoor, iLIDS-Parsing and Daily dataset are shown in Table 2, Table4, Table 8 and Table 6 respectively. It can be concludedthat “TCNet l+s+c+uf” consistently outperforms “TC-Net l+s+c+f”, which verifies the effectiveness of un-supervised optical flow refinement.Pixel-level fusion weights: As shown in Figure 3,the pixel-level fusion module summarizes the seg-mentation results of It−l, It−s and It. In the Indoordataset, l and s are set to 3 and 1. We visualize thelearned weights of “Face” and “L-arm” in the Indoordataset in Figure 7. The horizontal axis has 3 × Kticks, corresponding to the K fine-grained categoriesfor It−3, It−1 and It sequentially. K is 13 in theIndoor dataset. The vertical axis illustrates the fusionweights. The black dot denotes the maximum weightof each frame. By observing the sub-figure of “Face”,we have the following conclusions. First, the weightsof It−3, It−1 and It share similar distributions. Second,all maximum values for the It−3, It−1 and It, i.e.,the three black dots, are positive. It demonstratesthat they all contribute to the final result. Third, themaximum values correspond to “Face”, which is thecategory under consideration. Fourth, the maximum

Face

Fusi

on

wei

ghts

Fusi

on

wei

ghts

L-arm

Fig. 7. The fusion weights for “Face” and “L-arm” in theIndoor dataset. The 13-dim weights for It−l, It−s andIt are shown sequentially, separated by vertical dottedlines. For each frame, the category with the maximumfusion weight is denoted by the black dot.

value of It is greater than that of It−1, which isgreater than that of It−3. It is because temporallycloser frame contributes more to the frame of interest.Similar phenomenon can be found in the ‘L-arm” case.Long/Short range context: We test the effectivenessof long and short range context. “TCNet l” meansTCNet with long-range context only. To implementthis TCNet variant, th image pair {It, It−l} are fedinto TCNet during both training and testing phases.Similarly, “TCNet s” is TCNet inputed with {It, It−s}.“TCNet l+s” means both long-range and short-rangeframes are considered. Table 2 shows the results inindoor dataset. The Ave.F-1 of “TCNet l” and “TCNets” reach 57.74% and 58.04% respectively, which arelower than “TCNet l+s” 58.43%. It proves the long andshort range context are complementary. The per-classF1 score of “TCNet l”, “TCNet s” and “TCNet l+s”in the Indoor datasets can be found in Table 3. Theyagain show that both long and short range context arenecessary.

Page 10: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 10

Methods bk face hair U-clothes

L-arm

R-arm pants L-

legR-leg Dress L-

shoeR-

shoe bag

EM-Adapt [31] 92.86 51.92 59.72 81.40 9.14 8.90 76.54 0.00 0.00 0.00 28.92 27.76 31.31NetWarp-DIS [9] 92.85 52.81 62.39 80.35 6.03 6.76 76.68 0.00 0.00 0.00 32.93 27.68 31.48NetWarp-Flownet [9] 93.06 54.07 63.27 80.54 8.59 7.02 77.20 0.00 0.00 0.00 33.46 29.59 31.95TCNet l+s+c 92.39 53.83 66.64 82.87 0.19 3.11 76.63 0.00 0.00 0.00 28.45 29.48 42.99TCNet l+s+c+f 91.91 59.06 67.97 83.41 10.75 7.40 73.53 0.00 0.00 1.45 27.51 26.36 42.26TCNet l+s+c+uf 93.01 62.01 68.36 83.45 5.55 10.34 77.70 0.00 0.00 2.90 32.64 24.68 42.17

TABLE 7Per-Class Comparison of F-1 scores with state-of-the-arts and several variants of TCNet in iLIDS-Parsing

dataset. (%).

Methods Acc. Fg.acc. Avg.pre. Avg.rec. Avg.F-1EM-Adapt [31] 85.78 63.45 57.57 60.75 58.92NetWarp-DIS [9] 86.46 63.61 57.91 59.48 59.00NetWarp-Flownet [9] 86.64 64.06 58.51 59.57 59.42TCNet l+s+c 86.27 67.45 58.98 63.62 60.48TCNet l+s+c+f 86.29 67.49 58.09 65.26 60.76TCNet l+s+c+uf 87.11 67.33 61.41 62.63 61.72

TABLE 8Comparison with state-of-the-arts and several variants

of TCNet in Daily dataset. (%).

Frame-level Context: By observing Table 2, Table4, Table 8 and Table 6, we conclude that The “TCNetl+s+c+f” consistently outperforms “TCNet l+s+c” interms of Avg. F-1. It proves that the FACTS taskbenefits from adding the frame-level context.

4.4 Qualitative Results

Figure 8 shows the stepwise results of the pixel-levelcontext fusion process in the Indoor dataset. The 1∼4columns, 5∼8 columns and 9∼12 columns correspondto It−l, It−s and It respectively. In the first row, theleft shoe of the man is missing in Pt indicated by theblack ellipse. Thanks to the warped label P(t−s)→t, theleft shoe is retrieved back in the refined prediction Pt.In the second row, the belt in the left shoulder of Pt

is removed by fusing the warped result P(t−l)→t andP(t−s)→t. For the women in the third row, the wronglypredicted “pants” disappear by pixel-level fusion. Inthe fourth row, the left arm of the person are warpedincorrectly in both P(t−l)→t and P(t−s)→t. Thanks tothe confidences Ct,t−l and Ct,t−s, these errors arenot propagated to the Pt. All the above qualitativeresults verify the effectiveness of pixel-level contextpropagation.

Figure 9 shows the qualitative comparisons of theEM-Adapt, NetWarp-DIS, NetWarp-Flownet and sev-eral variants of the proposed TCNet in four datasets.The segmentation of the “bag” becomes increasinglybetter from the left to the right in the first row ofFigure 9 (a). Similar observations can be found bycomparing the “R-shoe” in the second row of Figure 9(a). The “face” of the woman in the first row of Figure

9 (b) are gradually refined by adding more modules ofTCNet. All the baselines are not able to well estimate“R-shoe” in the second row of Figure 9 (b). Besides,only TCNet can correctly infer the “bag” and “R-shoe”in the first and second row of Figure 9 (c) respectively.The segmentation results of all baselines are worsethan TCNet in Figure 9 (d).

4.5 Time ComplexityIn the inference stage, TCNet does not bring too muchcomputation cost than single frame based segmenta-tion methods, such as Deeplab and EM-Adapt. Onone hand, when parsing frame It, the long-rangeframe It−l and short-range frame It−s do not needto go through the computationally expensive featureextraction and frame parsing modules because theirrough parsing results Pt−l and Pt−s have already beencalculated. On the other hand, the extra computationbrought by the optical flow estimation is small be-cause the deep features are shared. It takes about 31msto segment one 241x121px frame on a modern GPU(Nvidia GeForce GTX 1080Ti).

5 CONCLUSION & FUTURE WORKSIn this work, we present the temporal context segmen-tation network to solve a rarely studied problem, i.e.,the fine-grained human-centric tracklet segmentationproblem. Since labeling the object parts is tedious,FACTS only requires one labeled frame per tracklet.To fully make use of the large amount of unlabeledframes, TCNet exploits both pixel-level and frame-level context by estimating optical flow and classifi-cation likelihood vector online. We demonstrate theeffectiveness of TCNet on four newly collected videoparsing datasets, which have already been releasedon our project page. In future, we plan to developa real-time FACTS system on mobile phones basedon the model compression and pruning techniques.Moreover, we plan to apply TCNet to segment otherkinds of object tracklet, such as cars.

REFERENCES[1] Thomas Brox and Jitendra Malik. Large displacement optical

flow: descriptor matching in variational motion estimation.TPAMI, 2011.

Page 11: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 11

Methods bk face hair U-clothes

L-arm

R-arm pants L-

legR-leg Dress L-

shoeR-

shoe bag

EM-Adapt [31] 94.50 68.59 57.10 67.75 45.77 41.88 66.16 60.54 59.06 68.33 44.73 45.87 45.66NetWarp-DIS [9] 94.26 69.90 58.50 69.23 45.48 43.29 66.60 60.43 59.73 68.03 44.01 42.55 45.05NetWarp-Flownet [9] 94.34 70.64 58.93 69.67 45.99 43.98 66.79 60.86 60.28 68.08 44.16 42.98 45.80TCNet l+s+c 94.62 72.79 58.79 71.26 48.76 43.68 67.72 63.47 62.28 69.61 42.30 45.05 45.94TCNet l+s+c+f 94.83 71.67 60.09 70.01 48.61 45.43 68.36 63.11 62.39 69.63 44.51 45.51 45.73TCNet l+s+c+uf 95.14 73.99 60.43 71.38 49.39 44.53 68.79 64.18 63.33 70.68 47.18 47.56 45.78

TABLE 9Per-Class Comparison of F-1 scores with state-of-the-arts and some variants of TCNet in Daily dataset. (%).

𝐼𝑡−𝑙 𝑃𝑡−𝑙 𝑃(𝑡−𝑙)→𝑡 𝐶𝑡,𝑡−𝑙 𝐼𝑡 𝑃𝑡−𝑠 𝑃𝑡 𝐺𝑡𝐼𝑡−𝑠 𝑃𝑡−𝑠 𝑃(𝑡−𝑠)→𝑡 𝐶𝑡,𝑡−𝑙

Fig. 8. Step by step illustration of the effectiveness of the pixel-level context fusion. 1∼4 columns: the long-range frame, its parsing result, the warped parsing result and the confidence map. 5∼8 columns: the short-rangeframe, its parsing result, the warped parsing result and the confidence map. 9∼12 columns: test image, the roughparsing result, refined parsing result and ground truth.

[2] Sergi Caelles, Kevis-Kokitsi Maninis, Jordi Pont-Tuset, LauraLeal-Taixe, Daniel Cremers, and Luc Van Gool. One-shot videoobject segmentation. In CVPR, 2017.

[3] Sergi Caelles, Alberto Montes, Kevis-Kokitsi Maninis, YuhuaChen, Luc Van Gool, Federico Perazzi, and Jordi Pont-Tuset.The 2018 davis challenge on video object segmentation. arXivpreprint arXiv:1803.00557, 2018.

[4] Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos,Kevin Murphy, and Alan L Yuille. Semantic image segmen-tation with deep convolutional nets and fully connected crfs.arXiv preprint arXiv:1412.7062, 2014.

[5] Xianjie Chen, Roozbeh Mottaghi, Xiaobai Liu, Sanja Fidler,Raquel Urtasun, and Alan Yuille. Detect what you can:Detecting and representing objects using holistic models andbody parts. In CVPR, 2014.

[6] M. Everingham, S. M. A. Eslami, L. Van Gool, C. K. I. Williams,J. Winn, and A. Zisserman. The pascal visual object classeschallenge: A retrospective. IJCV, 2015.

[7] Mohsen Fayyaz, Mohammad Hajizadeh Saffar, Mohammad

Sabokrou, Mahmood Fathy, Reinhard Klette, and Fay Huang.Stfcn: spatio-temporal fcn for semantic video segmentation.arXiv preprint arXiv:1608.05971, 2016.

[8] Philipp Fischer, Alexey Dosovitskiy, Eddy Ilg, Philip Hausser,Caner Hazırbas, Vladimir Golkov, Patrick van der Smagt,Daniel Cremers, and Thomas Brox. Flownet: Learningoptical flow with convolutional networks. arXiv preprintarXiv:1504.06852, 2015.

[9] Raghudeep Gadde, Varun Jampani, and Peter V Gehler. Se-mantic video cnns through representation warping. arXivpreprint arXiv:1708.03088, 2017.

[10] Shaogang Gong, Marco Cristani, Shuicheng Yan, andChange Loy Chen. Person Re-Identification. 2014.

[11] Joao F. Henriques, Caseiro Rui, Pedro Martins, and JorgeBatista. High-speed tracking with kernelized correlation fil-ters. TPAMI, 2014.

[12] Yangqing Jia, Evan Shelhamer, Jeff Donahue, Sergey Karayev,Jonathan Long, Ross Girshick, Sergio Guadarrama, and TrevorDarrell. Caffe: Convolutional architecture for fast feature

Page 12: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 12

Input EM-AdaptTCNetl+s+c

TCNetl+s+c+f

TCNetl+s+c+uf

GTNetWarp-

DISNetWarp-Flownet

(a) Indoor dataset

Input EM-AdaptTCNetl+s+c

TCNetl+s+c+f

TCNetl+s+c+uf

GTNetWarp-

DISNetWarp-Flownet

(b) Outdoor dataset

(c) iLIDS-Parsing dataset (d) Daily dataset

U-clothes R-shoesPants L-shoesR-leg R-armL-leg HairL-arm FaceDress Bag

Fig. 9. Qualitative results of the EM-Adapt, NetWarp-DIS, and NetWarp-Flownet and several variants of TCNet.

embedding. arXiv preprint arXiv:1408.5093, 2014.[13] Hao Jiang and Kristen Grauman. Detangling people: Individ-

uating multiple close people and their body parts via regionassembly. In CVPR, 2017.

[14] Xiaojie Jin, Xin Li, Huaxin Xiao, Xiaohui Shen, Zhe Lin, JimeiYang, Yunpeng Chen, Jian Dong, Luoqi Liu, Zequn Jie, et al.Video scene parsing with predictive feature learning. In ICCV,2017.

[15] Anna Khoreva, Rodrigo Benenson, Eddy Ilg, Thomas Brox,and Bernt Schiele. Lucid data dreaming for multiple objecttracking. arXiv preprint arXiv:1703.09554, 2017.

[16] Shuang Li, Tong Xiao, Hongsheng Li, Bolei Zhou, Dayu Yue,and Xiaogang Wang. Person search with natural languagedescription. arXiv preprint arXiv:1702.05729, 2017.

[17] Xiaoxiao Li and Chen Change Loy. Video object segmentationwith joint re-identification and attention-aware mask propa-gation. arXiv preprint arXiv:1803.04242, 2018.

[18] Xiaoxiao Li, Yuankai Qi, Zhe Wang, Kai Chen, Ziwei Liu,Jianping Shi, Ping Luo, Xiaoou Tang, and Chen Change Loy.Video object segmentation with re-identification. arXiv preprintarXiv:1708.00197, 2017.

[19] Yule Li, Jianping Shi, and Dahua Lin. Low-latency videosemantic segmentation. In CVPR, 2018.

[20] Xiaodan Liang, Si Liu, Xiaohui Shen, Jianchao Yang, LuoqiLiu, Liang Lin, and Shuicheng Yan. Deep human parsing withactive template regression. TPAMI, 2015.

[21] Xiaodan Liang, Chunyan Xu, Xiaohui Shen, Jianchao Yang,Si Liu, Jinhui Tang, Liang Lin, and Shuicheng Yan. Humanparsing with contextualized convolutional neural network.ICCV, 2015.

[22] Si Liu, Xiaodan Liang, Luoqi Liu, Xiaohui Shen, Jianchao Yang,Changsheng Xu, Liang Lin, Xiaochun Cao, and Shuicheng Yan.Matching-cnn meets knn: Quasi-parametric human parsing.arXiv preprint arXiv:1504.01220, 2015.

[23] Si Liu, Zheng Song, Guangcan Liu, Shuicheng Yan, Chang-sheng Xu, and Hanqing Lu. Street-to-shop: Cross-scenarioclothing retrieval via parts alignment and auxiliary set. InCVPR, 2012.

[24] Si Liu, Changhu Wang, Yao Sun, Ruihe Qian, Han Yu, andRenda Bao. Surveillance video parsing with single framesupervision. CVPR, 2017.

[25] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fullyconvolutional networks for semantic segmentation. CVPR,2015.

[26] Jonathon Luiten, Paul Voigtlaender, and Bastian Leibe. Pre-mvos: Proposal-generation, refinement and merging for thedavis challenge on video object segmentation 2018, 2018.

[27] Ping Luo, Xiaogang Wang, and Xiaoou Tang. Pedestrianparsing via deep decompositional network. In ICCV, 2013.

[28] Wenjie Luo, Alexander G Schwing, and Raquel Urtasun. Effi-cient deep learning for stereo matching. In CVPR, 2016.

[29] Behrooz Mahasseni, Sinisa Todorovic, and Alan Fern. Budget-aware deep semantic video segmentation. In CVPR, 2017.

[30] David Nilsson and Cristian Sminchisescu. Semantic videosegmentation by gated recurrent flow propagation. arXivpreprint arXiv:1612.08871, 2016.

[31] George Papandreou, Liang-Chieh Chen, Kevin P Murphy, andAlan L Yuille. Weakly-and semi-supervised learning of a deepconvolutional network for semantic image segmentation. InICCV, 2015.

[32] F. Perazzi, J. Pont-Tuset, B. McWilliams, L. Van Gool, M. Gross,and A. Sorkine-Hornung. A benchmark dataset and evalua-tion methodology for video object segmentation. In CVPR,2016.

[33] Federico Perazzi, Anna Khoreva, Rodrigo Benenson, BerntSchiele, and Alexander Sorkine-Hornung. Learning videoobject segmentation from static images. In CVPR, 2017.

[34] Tomas Pfister, James Charles, and Andrew Zisserman. Flowingconvnets for human pose estimation in videos. arXiv preprintarXiv:1506.02897, 2015.

[35] Jordi Pont-Tuset, Federico Perazzi, Sergi Caelles, Pablo Ar-belaez, Alexander Sorkine-Hornung, and Luc Van Gool. The2017 davis challenge on video object segmentation. arXivpreprint arXiv:1704.00675, 2017.

[36] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun.Faster r-cnn: Towards real-time object detection with regionproposal networks. In NIPS, 2015.

[37] Evan Shelhamer, Kate Rakelly, Judy Hoffman, and TrevorDarrell. Clockwork convnets for video semantic segmentation.In ECCV, 2016.

[38] Karen Simonyan and Andrew Zisserman. Very deep con-

Page 13: Fine-grained Human-centric Tracklet Segmentation …colalab.org/media/paper/Fine-grained_Human-centric_Track...Abstract—In this paper, we target at the Fine-grAined human-Centric

0162-8828 (c) 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for moreinformation.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI10.1109/TPAMI.2019.2911936, IEEE Transactions on Pattern Analysis and Machine Intelligence

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. X, X 20XX 13

volutional networks for large-scale image recognition. arXivpreprint arXiv:1409.1556, 2014.

[39] Jianyu Wang and Alan L Yuille. Semantic part segmentationusing compositional model combining shape and appearance.In CVPR, 2015.

[40] Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price,and Alan L Yuille. Joint object and part segmentation usingdeep learned potentials. In ICCV, 2015.

[41] Taiqing Wang, Shaogang Gong, Xiatian Zhu, and ShengjinWang. Person re-identification by discriminative selection invideo ranking. TPAMI, 2016.

[42] Fangting Xia, Peng Wang, Liang-Chieh Chen, and Alan LYuille. Zoom better to see clearer: Human part segmentationwith auto zoom net. arXiv preprint arXiv:1511.06881, 2015.

[43] Ning Xu, Linjie Yang, Yuchen Fan, Jianchao Yang, DingchengYue, Yuchen Liang, Brian Price, Scott Cohen, and ThomasHuang. Youtube-vos: Sequence-to-sequence video object seg-mentation. arXiv preprint arXiv:1809.00461, 2018.

[44] Kazuhiro Yamaguchi, Mohammad Hadi Kiapour, and TamaraBerg. Paper doll parsing: Retrieving similar styles to parseclothing items. In ICCV, 2013.

[45] HW Zhang, Z-J Zha, Y Yang, SC Yan, Y Gao, and T-S Chua.Attribute-augmented semantic hierarchy: towards bridgingsemantic gap and intention gap in image retrieval. In ACMMM, 2013.

[46] Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang,and Jiaya Jia. Pyramid scene parsing network. In CVPR, 2017.

[47] Rui Zhao, Wanli Ouyang, and Xiaogang Wang. Learning mid-level filters for person re-identification. In CVPR, 2014.

[48] Qixian Zhou, Xiaodan Liang, Ke Gong, and Liang Lin. Adap-tive temporal encoding network for video instance-level hu-man parsing. arXiv preprint arXiv:1808.00661, 2018.

[49] Tinghui Zhou, Shubham Tulsiani, Weilun Sun, Jitendra Malik,and Alexei A Efros. View synthesis by appearance flow. arXivpreprint arXiv:1605.03557, 2016.

[50] Shizhan Zhu, Sanja Fidler, Raquel Urtasun, Dahua Lin, andChange Loy Chen. Be your own prada: Fashion synthesis withstructural coherence. arXiv preprint arXiv:1710.07346, 2017.

[51] Xizhou Zhu, Yujie Wang, Jifeng Dai, Lu Yuan, and Yichen Wei.Flow-guided feature aggregation for video object detection. InICCV, 2017.

[52] Xizhou Zhu, Yuwen Xiong, Jifeng Dai, Lu Yuan, and YichenWei. Deep feature flow for video recognition. In CVPR, 2017.

Si Liu is an associate professor in BeihangUniversity. She received PHD degree fromInstitute of Automation, Chinese Academy ofSciences. She has been Research Assistantand Postdoc in National University of Sin-gapore. Her research interest includes com-puter vision and multimedia analysis. Shehas published over 40 cutting-edge paperson the human-related analysis including thehuman parsing, face editing and image re-trieval. She was the recipient of Best Paper

of ACM MM 2013, Best demo award of ACM MM 2012. She wasthe Champion of CVPR 2017 Look Into Person Challenge and theorganizer of the ECCV 2018 Person in Context Challenge.

Guanghui Ren is a master student fromInstitute of Information Engineering, ChineseAcademy of Sciences. He received the BSdegree from Xi’an Jiaotong University. Hisresearch interests include object detectionand semantic segmentation.

Sun Yao is an Associate Professor in theInstitute of Information Engineering, ChineseAcademy of Sciences. He received his Ph.D.degree from the Academy of the Mathemat-ics and Systems Science, Chinese Academyof Sciences.

Jinqiao Wang currently a Professor in Na-tional Laboratory of Pattern Recognition,Institute of Automation, Chinese Academyof Sciences. He received the B.E. degreein 2001 from Hebei University of Technol-ogy, China, and the M.S. degree in 2004from Tianjin University, China. He receivedthe Ph.D. degree in pattern recognitionand intelligence systems from the NationalLaboratory of Pattern Recognition, ChineseAcademy of Sciences, in 2008.

Changhu Wang is currently Director ofByteDance AI Lab, Beijing, China. He re-ceived B.E. and Ph.D. degrees from the Uni-versity of Science and Technology of China,in 2004 and 2009, respectively. Before join-ing Toutiao AI Lab, he worked as a leadresearcher in Microsoft Research from 2009to 2017. He worked as a research engineerat the department of Electrical and ComputerEngineering in National University of Singa-pore in 2008.

Bo Li is Changjiang Distinguished Professorof School of Computer Science and Engi-neering, Beihang University. He is a recip-ient of The National Science Fund for Dis-tinguished Young Scholars. He is currentlythe dean of AI Research Institute, BeihangUniversity.He is the chief scientist of National973 Program and the principal investigatorof The National Key Research and Develop-ment Program. He has published over 100papers in top journals and conferences and

held over 50 domestic and foreign patents.

Shuicheng Yan is chief scientist of Qi-hoo/360 company, and also the Dean’s ChairAssociate Professor at National Universityof Singapore. Dr. Yan’s research areas in-clude machine learning, computer visionand multimedia, and he has authored/co-authored hundreds of technical papers overa wide range of research topics, with GoogleScholar citation over 20,000 times and H-index 66. He is ISI Highly-cited Researcherof 2014, 2015 and 2016. His team received

7 times winner or honorable-mention prizes in PASCAL VOC andILSVRC competitions, along with more than 10 times best (student)paper prizes. He is also IAPR Fellow.