9
Weakly Supervised Salient Object Detection Using Image Labels Guanbin Li 1 , Yuan Xie 1 , Liang Lin 1,2* 1 School of Data and Computer Science, Sun Yat-sen University, Guangzhou, China 2 SenseTime Group Limited Abstract Deep learning based salient object detection has recently achieved great success with its performance greatly outper- forms any other unsupervised methods. However, annotating per-pixel saliency masks is a tedious and inefficient proce- dure. In this paper, we note that superior salient object de- tection can be obtained by iteratively mining and correct- ing the labeling ambiguity on saliency maps from traditional unsupervised methods. We propose to use the combination of a coarse salient object activation map from the classifi- cation network and saliency maps generated from unsuper- vised methods as pixel-level annotation, and develop a simple yet very effective algorithm to train fully convolutional net- works for salient object detection supervised by these noisy annotations. Our algorithm is based on alternately exploit- ing a graphical model and training a fully convolutional net- work for model updating. The graphical model corrects the internal labeling ambiguity through spatial consistency and structure preserving while the fully convolutional network helps to correct the cross-image semantic ambiguity and si- multaneously update the coarse activation map for next it- eration. Experimental results demonstrate that our proposed method greatly outperforms all state-of-the-art unsupervised saliency detection methods and can be comparable to the current best strongly-supervised methods training with thou- sands of pixel-level saliency map annotations on all public benchmarks. Introduction Salient object detection is designed to accurately detect dis- tinctive regions in an image that attract human attention. Re- cently, this topic has attracted widespread interest in the re- search community of computer vision and cognitive science as it can be applied to benefit a wide range of artificial intelli- gence and vision applications, such as robot intelligent con- trol (Shon et al. 2005), content-aware image editing (Avidan * Corresponding author is Liang Lin (Email: [email protected]). This work was supported in part by the National Natural Science Foundation of China under Grant 61702565, in part by the Spe- cial Program of the NSFC-Guangdong Joint Fund for Applied Re- search on Super Computation (the second phase), in part by Guang- dong Natural Science Foundation Project for Research Teams un- der Grant 2017A030312006. This work was also sponsored by CCF-Tencent Open Research Fund. Copyright c 2018, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Ours Source GT MST MB+ CAM Figure 1: Two kinds of defects in state-of-the-art unsuper- vised salient object detection methods vs. the results of our proposed weakly supervised optimization framework. and Shamir 2007), visual tracking (Mahadevan and Vascon- celos 2009) and video summarization (Ma et al. 2002). Recently, the deployment of deep convolutional neural networks has resulted in significant progress in salient ob- ject detection (Li and Yu 2016b; Li and Yu 2016a; Liu and Han 2016; Wang et al. 2016). The performance of these CNN based methods, however, comes at the cost of requiring pixel-wise annotations to generate training data. For salient object detection, it is painstaking to annotate mask-level la- bel and takes several minutes for an experienced annotator to label one image. Moreover, as the definition of an object being salient is very subjective, there often exists multiple diverse annotations for a same image between different an- notators. To ensure the quality of training data sets, these im- ages with ambiguous annotations should be removed, which makes the labeling task more laborious and time-consuming. This time-consuming task is bound to limit the total amount of pixel-wise training samples and thus become the bot- tleneck of further development of fully-supervised learning based methods. As a low level vision problem, there exists an ocean of un- supervised salient object detection methods (Wei et al. 2012; Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low- level features such as color, gradient or contrast and some saliency priors, such as the center prior (Liu et al. 2011) and the background prior (Wei et al. 2012). As it is impracti- arXiv:1803.06503v1 [cs.CV] 17 Mar 2018

arXiv:1803.06503v1 [cs.CV] 17 Mar 2018 · Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low-level features such as color,

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: arXiv:1803.06503v1 [cs.CV] 17 Mar 2018 · Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low-level features such as color,

Weakly Supervised Salient Object Detection Using Image Labels

Guanbin Li1, Yuan Xie1, Liang Lin1,2∗1School of Data and Computer Science, Sun Yat-sen University, Guangzhou, China

2SenseTime Group Limited

AbstractDeep learning based salient object detection has recentlyachieved great success with its performance greatly outper-forms any other unsupervised methods. However, annotatingper-pixel saliency masks is a tedious and inefficient proce-dure. In this paper, we note that superior salient object de-tection can be obtained by iteratively mining and correct-ing the labeling ambiguity on saliency maps from traditionalunsupervised methods. We propose to use the combinationof a coarse salient object activation map from the classifi-cation network and saliency maps generated from unsuper-vised methods as pixel-level annotation, and develop a simpleyet very effective algorithm to train fully convolutional net-works for salient object detection supervised by these noisyannotations. Our algorithm is based on alternately exploit-ing a graphical model and training a fully convolutional net-work for model updating. The graphical model corrects theinternal labeling ambiguity through spatial consistency andstructure preserving while the fully convolutional networkhelps to correct the cross-image semantic ambiguity and si-multaneously update the coarse activation map for next it-eration. Experimental results demonstrate that our proposedmethod greatly outperforms all state-of-the-art unsupervisedsaliency detection methods and can be comparable to thecurrent best strongly-supervised methods training with thou-sands of pixel-level saliency map annotations on all publicbenchmarks.

IntroductionSalient object detection is designed to accurately detect dis-tinctive regions in an image that attract human attention. Re-cently, this topic has attracted widespread interest in the re-search community of computer vision and cognitive scienceas it can be applied to benefit a wide range of artificial intelli-gence and vision applications, such as robot intelligent con-trol (Shon et al. 2005), content-aware image editing (Avidan

∗Corresponding author is Liang Lin (Email: [email protected]).This work was supported in part by the National Natural ScienceFoundation of China under Grant 61702565, in part by the Spe-cial Program of the NSFC-Guangdong Joint Fund for Applied Re-search on Super Computation (the second phase), in part by Guang-dong Natural Science Foundation Project for Research Teams un-der Grant 2017A030312006. This work was also sponsored byCCF-Tencent Open Research Fund.Copyright c© 2018, Association for the Advancement of ArtificialIntelligence (www.aaai.org). All rights reserved.

OursSource GTMSTMB+ CAM

Figure 1: Two kinds of defects in state-of-the-art unsuper-vised salient object detection methods vs. the results of ourproposed weakly supervised optimization framework.

and Shamir 2007), visual tracking (Mahadevan and Vascon-celos 2009) and video summarization (Ma et al. 2002).

Recently, the deployment of deep convolutional neuralnetworks has resulted in significant progress in salient ob-ject detection (Li and Yu 2016b; Li and Yu 2016a; Liu andHan 2016; Wang et al. 2016). The performance of theseCNN based methods, however, comes at the cost of requiringpixel-wise annotations to generate training data. For salientobject detection, it is painstaking to annotate mask-level la-bel and takes several minutes for an experienced annotatorto label one image. Moreover, as the definition of an objectbeing salient is very subjective, there often exists multiplediverse annotations for a same image between different an-notators. To ensure the quality of training data sets, these im-ages with ambiguous annotations should be removed, whichmakes the labeling task more laborious and time-consuming.This time-consuming task is bound to limit the total amountof pixel-wise training samples and thus become the bot-tleneck of further development of fully-supervised learningbased methods.

As a low level vision problem, there exists an ocean of un-supervised salient object detection methods (Wei et al. 2012;Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015;Yang et al. 2013). These methods are usually based on low-level features such as color, gradient or contrast and somesaliency priors, such as the center prior (Liu et al. 2011) andthe background prior (Wei et al. 2012). As it is impracti-

arX

iv:1

803.

0650

3v1

[cs

.CV

] 1

7 M

ar 2

018

Page 2: arXiv:1803.06503v1 [cs.CV] 17 Mar 2018 · Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low-level features such as color,

cal to define a set of universal rules for how an object be-ing salient, each of these bottom-up methods works wellfor some images, but none of them can handle all the im-ages. By observing the failure cases, We found that most ofthe saliency detection error lies in the lack of spatial corre-lation inference and image semantic contrast detection. Asshown in Figure 1, the two unsupervised saliency detectionmethods are only able to detect part of the salient objects inthe first two cases as they fail to take into account the spa-tial consistency (e.g. encouraging nearby pixels with similarcolors to take similar saliency scores), while for the last twocases, the two methods completely fail to detect the salientobjects as these objects are of very low contrast in termsof low-level features (they are salient in high semantic con-trast). These two kinds of failure cases are hard to be foundin fully-supervised methods.

During the training and testing of several fully-supervisedsaliency detection methods (Li and Yu 2016a; Liu and Han2016; Wang et al. 2016), we found that a well-trained deep-convolution network without over-fitting can even correctsome user annotation error exists in the training samples.We conjecture that a large amount of model parameters con-tained in deep neural network can be trained to discoverthe universal rules implied in large scale training samplesand thus can help to detect the ambiguity in the annotationmask (noisy annotation) and figure out a “correct” one whichbeing in line with the hidden rules. Moreover, recent workhas shown that CNNs being trained on image-level labelsfor classification have remarkable ability to localize the mostdiscriminative region of an image (Zhou et al. 2016).

Inspired by these observations, in this paper, we addressthe weakly supervised salient object detection task usingonly image-level labels, which in most cases specify salientobjects within the image. We develop an algorithm thatexploits saliency maps generated from any unsupervisedmethod as noisy annotations to train convolutional networksfor better saliency maps. Specifically, we first propose a con-ditional random field based graphical model to correct theinternal label ambiguity by enhancing the spatial coherenceand salient object localization. Meanwhile, a multi-task fullyconvolutional ResNet (He et al. 2015) is learned, which issupervised by the iteratively corrected pixel-level annota-tions as well as image labels (indicating significant objectclass within an image), and in turn provides two probabilitymaps to generate an updated unary potential for the graphi-cal model. The first probability map is called Class Activa-tion Map (CAM) and it highlights the discriminative objectparts detected by the image classification-trained CNN whilethe second one being a more accurate saliency map trainedfrom pixel-wise annotation. Though CAM itself is a rela-tively coarse pixel-level probability map, it shows very ac-curate salient object localization ability and thus can be usedas a guide to generate more precise pixel-wise annotation fora second round training. The proposed method is optimizedalternately until a stopping criteria appears. In our experi-ment, we find that although CAM is trained using imagesfrom a fix number of image classes, it generalizes well toimages of unknown categories, resulting in an intensely ac-curate salient object positioning for generic salient objects.

The proposed optimization framework also theoretically ap-plies to all unsupervised salient object detection methodsand is able to generate more accurate saliency map very ef-ficiently in fewer than one second per image no matter howtime-consuming the original model.

In summary, this paper has the following contributions:

• We introduce a generic alternate optimization frameworkto fill the performance gap between supervised and unsu-pervised salient object detection methods without resort-ing to laborious pixel labeling.

• We propose a conditional random field based graphicalmodel to cleanse the noisy pixel-wise annotation by en-hancing the spatial coherence as well as salient object lo-calization.

• We also design a multi-task fully convolutional ResNet-101 to both generate a coarse class activation map (CAM)and a pixel-wise saliency probability map, the cooperationof which can help to detect and correct the cross-imageannotation ambiguity, generating more accurate saliencyannotation for iterative training.

Alternate Saliency Map OptimizationAs shown in Figure 2, our proposed saliency map optimiza-tion framework consists of two components, a multi-taskfully convolutional network (Multi-FCN) and a graphicalmodel based on conditional random fields (CRF). Giventhe Microsoft COCO dataset (Lin et al. 2014) with mul-tiple image labels corresponding to each image, we ini-tially utilize a state-of-the-art unsupervised salient object de-tection method, i.e. minimum barrier salient object detec-tion (MB+), to generate the saliency maps of all trainingimages. The produced saliency maps as well as their cor-responding image labels are employed to train the Multi-FCN, which simultaneously learns to predict a pixel-wisesaliency map and an image class distribution. When trainingconverged, a class activation mapping technique (Zhou etal. 2016) is applied to the Multi-FCN to generate a seriousof class activation maps (CAMs). Then the initial saliencymap, the predicted saliency map from Multi-FCN as wellas the average map of the top three CAMs (CAM predic-tion corresponding to top 3 classes) are employed to theCRF model to get the corresponding maps with better spatialcoherence and contour localization. We further propose anannotation updating scheme to construct new saliency mapannotations from these three maps with CRF for a seconditeration of Multi-FCN training. Finally, to generalize themodel for saliency detection of unknown image labels, wefurther finetune the saliency map prediction stream of theMulti-FCN guided by generated CAM using salient objectdetection datasets (e.g. MSRA-B and HKUIS) without an-notations.

Multi-Task Fully Convolutional NetworkIn the multi-task fully convolutional stream, we aim to de-sign an end-to-end convolutional network that can be viewedas a combination of the image classification task and thepixel-wise saliency prediction task. To conceive such an

Page 3: arXiv:1803.06503v1 [cs.CV] 17 Mar 2018 · Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low-level features such as color,

image labelMulti-FCN

Euclidean lossorange

class distribution

saliency map

sigmoid cross-entropy loss

update saliency annotationin next iteration

Annotation updating

C R F

Training iteration

saliency annotation

image

CAM

updatedsaliency annotation

Figure 2: Overall framework for alternate saliency map optimization.

end-to-end architecture, we have the following considera-tions. First, the network should be able to correct the noisyinitial saliency annotations as well as possible by mining thesemantic ambiguity between the images. Second, the net-work should be able to be end to end trainable to output asaliency map with appropriate resolution. Last but not theleast, it should also be able to detect visual contrast at differ-ent scales.

We choose ResNet-101 (He et al. 2015) as our pre-trainednetwork and modify it to meet our requirements. We first re-fer to (Chen et al. 2014) and re-purpose it into a dense imagesaliency prediction network by replacing its 1000-way linearclassification layer with a linear convolutional layer with a1×1 kernel and two output channels. The feature maps afterthe final convolutional layer is only 1/32 of that of the origi-nal input image because the original ResNet-101 consists ofone pooling layer and 4 convolutional layers, each of whichhas stride 2. We call these five layers down-sampling layers.As described in (He et al. 2015), the 101 layers in ResNet-101 can be divided into five groups. Feature maps computedby different layers in each group share the same resolution.To make the final saliency map denser, we skip subsamplingin the last two down-sampling layers by setting their stride to1, and increase the dilation rate of subsequent convolutionalkernels using the dilation algorithm to enlarge their recep-tive fields as (Chen et al. 2014). Therefore, all the featuresmaps in the last three groups have the same resolution, 1/8original resolution after network transformation.

As it has been widely verified that feeding multiple scalesof an input image to networks with shared parametersare rewarding for accurately localizing objects of differentscales (Chen et al. 2015; Lin et al. 2015), we replicate thefully convolutional ResNet-101 network three times, eachresponsible for one input scale s (s ∈ {0.5, 0.75, 1}). Eachscale s of the input image is fed to one of the three replicated

ResNet-101, and outputs a two-channel probability map inthe resolution of scale s, denoted as Ms

c , where c ∈ {0, 1}denotes the two classes for saliency detection. The threeprobability maps are resized to the same resolution as theraw input image using bilinear interpolation, summed up andfed to a sigmoid layer to produce the final probability map.The network framework is shown in Figure 3.

For image classification task, as we desire to performobject localization from the classification model, we referto (Zhou et al. 2016) and integrate a global average poolinglayer for generating class activation maps. Specifically, asshown in Figure 3, we rescale the three output feature mapsof the last original convolutional layer in ResNet-101 (cor-responds to three input scale) to the same size (1/8 originalresolution) and concatenate to form feature maps for classi-fication. We further perform global average pooling on theconcatenated convolutional feature maps and use those asfeatures for a fully-connected layer which produces the de-sired classes distribution output. Let fk(x, y) represent theactivation of channel k in the concatenated feature map atspatial location (x, y). DefineMc as the class activation mapfor class c, where each spatial element can be calculated asfollows (Zhou et al. 2016):

Mc (x, y) =∑k

wckfk (x, y) . (1)

wck is the weight corresponding to class c for unit k (afterglobal average pooling, each channel of the concatenatedfeature map becomes a unit activation value).

Graphical Model for Saliency Map RefinementBy observing the saliency maps generated by state-of-the-art unsupervised methods, we find that for images with lowcontrast and complex background, the salient object canhardly be completely detected, with common defects exist

Page 4: arXiv:1803.06503v1 [cs.CV] 17 Mar 2018 · Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low-level features such as color,

ResNet-101(multi-scale + share weights )

concat

GAPbird,squirrel

saliency map

class label

conv

conv

conv

visualize

map 1

map 2

map 3

sum

scale 1

scale 2

scale 3

41×41,1024

input image

Figure 3: The architecture of our multi-task fully convolu-tional network (Multi-FCN).

in spatial consistency and contour preserving. We call thesedefects internal labeling ambiguity in noisy saliency anno-tations. Fully connect CRF model has been widely usedin semantic segmentation (Krahenbuhl and Koltun 2012;Chen et al. 2014) to both refine the segmentation result andbetter capture the object boundaries. It has also been usedas a post-processing step in (Li and Yu 2016a; Li and Yu2016b) for saliency map refinement. In this paper, we referto (Li and Yu 2016a) and utilize the same formulation andsolver of the two classes fully connected CRF model to cor-rect the internal labeling ambiguity. The output of the CRFoperation is a probability map, the value of which denotesthe probability of each pixel being salient. We convert it intoa binary label by thresholding when being used as trainingsample annotations.

Saliency Annotations Updating SchemeWe denote the original input image as I and the correspond-ing saliency map of the specific unsupervised method asSanno. After convergence of the first complete training ofMulti-FCN, we apply the trained model to generate saliencymaps as well as the average map of the top 3 class activa-tion maps for all training images. We denote the predictedsaliency map as Spredict and the average class activationmap as Scam. Furthermore, we also perform fully connectedCRF operation to the initial saliency maps produced by aspecific unsupervised method, the predicted saliency mapfrom Multi-FCN as well as the average class activation map.The resulting saliency maps are denoted as Canno, Cpredictand Ccam respectively. Base on this, we update the train-ing samples as well as their corresponding saliency annota-tions for the next iteration according to Algorithm 1. CRF ()denotes the CRF operation while Supdate refers to the up-dated saliency map annotation, which is further used as thesaliency groundtruth for the next iterative training. MAE()is defined as the average pixelwise absolute difference be-tween two saliency maps (i.e. S1 and S2), which is calcu-lated as follows:

MAE(S1, S2) =1

W ×H

W∑x=1

H∑y=1

|S1(x, y)− S2(x, y)|.

(2)

where W and H being the width and height of the saliencymap.

Algorithm 1 Saliency Annotations UpdatingRequire: Current saliency map annotation Sanno, the pre-

dicted saliency map Spredict, CRF output of currentsaliency map annotation Canno, CRF output of the pre-dicted saliency map Cpredict and CRF output of theclass activation map Ccam

Ensure: The updated saliency map annotation Supdate.1: if MAE(Canno,Cpredict) ≤ α then2: Supdate = CRF

(Sanno+Spredict

2

)3: else if MAE(Canno,Ccam) > β and

MAE(Cpredict,Ccam) > β then4: Discard the training sample in next iteration5: else if MAE(Canno,Ccam) ≤ MAE(Cpredict,Ccam)

then6: Supdate = Canno7: else8: Supdate = Cpredict9: end if

Multi-FCN Training with Weak LabelsThe training of the weakly supervised saliency map opti-mization framework is composed of two stages, both ofwhich are based on an alternative training scheme. In thefirst stage, we train the model using Microsoft COCOdataset (Lin et al. 2014) with multiple image labels perimage. Firstly, we choose a state-of-the-art unsupervisedsalient object detection model and apply it to produce aninitial saliency map for each image in the training. Then wesimply put the saliency maps as initial annotation and trainthe Multi-FCN for a pixel-wise saliency prediction as wellas the classification model for better class activation mapgeneration. While training, we validate on the validation setof the COCO dataset also with generated pixel-wise noisyannotations being the groundtruth. Also note that in orderto speed up the training, we initialize the Multi-FCN witha pre-trained model over the ImageNet dataset (Deng et al.2009) instead of training from scratch. After training con-vergence, we choose the model with lowest validation erroras the final model for this iteration, and apply it to gener-ate saliency maps as well as the average map of top 3 classactivation maps for all training images. Secondly, we ap-ply saliency annotations updating scheme according to Sec-tion to create updated training tuples (images, saliency an-notation and image label) for a second round of training. Wealternately train the model until a stopping criteria appears.After each training round, we evaluate the mean MAE be-tween each pair of saliency annotation (Pseudo Groundtruth)and the predicted saliency map, and the stopping criteriais defined to be the mean MAE gets lower than a specificthreshold or the total number of training rounds reaches5. (Noted that as being a weakly supervised method, wedo not use true annotations to determine the merits of themodel).

Page 5: arXiv:1803.06503v1 [cs.CV] 17 Mar 2018 · Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low-level features such as color,

Finally, in order to generalize the model for genericsalient object detection with unknown image labels, wefurther finetune the saliency map prediction stream of theMulti-FCN guided by offline CAMs using salient object de-tection datasets (e.g. the training images of MSRA-B andHKU-IS) without annotations, until the stopping criteria ap-pears. Here, we calculate the mean of the top 5 CAMs asthe guided CAM, and we discover that although CAM istrained with specific image classification labels, its predictedCAMs of the most similar categories in the category setcan still highlight the most discriminative regions in the im-age and thus still works well as an auxiliary guidance forgeneric salient object detection. The loss function for updat-ing Multi-FCN for pixel-wise saliency prediction is definedas the sigmoid cross entropy between the generated groundtruth (G) and the predicted saliency map (S):

L =− βi|I|∑i=1

Gi logP (Si = 1|Ii,W )

− (1− βi)|I|∑i=1

(1−Gi) logP (Si = 0|Ii,W ) ,

(3)

whereW denotes the collection of corresponding networkparameters in the Multi-FCN, βi is a weight balancing thenumber of salient pixels and unsalient ones, and |I|, |I| and|I|+ denote the total number of pixels, unsalient pixels andsalient pixels in image I , respectively. Then βi =

|I||I| and

1− βi = |I|+|I| . When training for multi-label object classifi-

cation, we simply employ the Euclidean loss as the objectivefunction and only update the parameters of the fully con-nected inference layer with parameters of the main ResNet-101 being unchanged.

Experimental ResultsImplementationOur proposed Multi-FCN has been implemented on the pub-lic DeepLab code base (Chen et al. 2014). A GTX TitanX GPU is used for both training and testing. As describedin Section , the Multi-FCN involves two stages training.In the first stage, we train on Microsoft COCO object de-tection dataset for multi-label recognition, which comprisesa training set of 82,783 images, and a validation set of40,504 images. The dataset covers 80 common object cate-gories, with about 3.5 object labels per image. In the secondstage, we combine the training images of both the MSRA-B dataset (2500 images) (Liu et al. 2011) and the HKU-ISdataset (2500 images) (Li and Yu 2016b) as our trainingset (5000 images), with all original saliency annotations re-moved. The validation sets without annotations in the afore-mentioned two datasets are also combined as our validationset (1000 images). During training, the mini-batch size is setto 2 and we choose to update the loss every 5 iterations. Weset the momentum parameter to 0.9 and the weight decay to0.0005 for both subtasks. The total number of iteration is setto 8K during each training round. During saliency annota-tion updating, the thresholds α and β are set to 15 and 40

respectively. The mean MAE of the training stop criteria isset to 0.05 in our experiment.

DatasetsWe conducted evaluations on six public salient object bench-mark datasets: MSRA-B (Liu et al. 2011), PASCAL-S (Li etal. 2014), DUT-OMRON(Yang et al. 2013), HKU-IS (Li andYu 2016b), ECSSD (Yan et al. 2013) and SOD (Martin et al.2001). Though we do not use any user annotations in train-ing, we get to know the training and validation sets of theMSRA-B and HKU-IS datasets in advance. Therefore, forthe sake of fairness, we evaluate our model on the testingsets of these two datasets and on the combined training andtesting sets of other datasets.

Evaluation MetricsWe adopt precision-recall curves (PR), maximum F-measureand mean absolute error (MAE) as our performance mea-sures. The continuous saliency map is binarized using dif-ferent thresholds varying from 0 to 1. At each thresholdvalue, a pair of precision and recall value can be obtainedby comparing the binarized saliency map against the binarygroundtruth. The PR curve of a dataset is obtained fromall pairs of average precision and recall over saliency mapsof all images in the dataset. The F-measure is defined asFβ = (1+β2)·Precision·Recall

β2·Precision+Recall , where β2 is set to 0.3. We re-port the maximum F-measure computed from all precision-recall pairs. MAE is defined as the average pixelwise ab-solute difference between the binary ground truth and thesaliency map (Perazzi et al. 2012) as described in Equa-tion 2.

Comparison with the UnsupervisedState-of-the-ArtOur proposed alternate saliency map optimization frame-work requires an unsupervised benchmark model as initial-ization. In this section, we choose the state-of-the-art min-imum barrier salient object detection (MB+) method as abaseline and take the optimized model as our final modelwhen compared with other benchmarks. In Section , we willlist more results of our proposed method on other baselinemodels to demonstrate the effectiveness of our proposed al-gorithm.

We compare our method with eight classic or state-of-the-art unsupervised saliency detection algorithms, includ-ing GS (Wei et al. 2012), SF (Perazzi et al. 2012), HS (Yanet al. 2013), MR (Yang et al. 2013), GC (Cheng et al. 2015),BSCA (Qin et al. 2015), MB+ (Zhang et al. 2015) andMST (Tu et al. 2016). For fair comparison, the saliency mapsof different methods are provided by authors or obtainedfrom the available implementations.

A visual comparison is given in Fig. 5. As can be seen,our method generates more accurate saliency maps in var-ious challenging cases, e.g., object in complex backgroundand low contrast between object and background. It is par-ticularly noteworthy that our proposed method employed thesaliency maps generated by MB+ (Zhang et al. 2015) as ini-tial noisy annotations for iterative training, it can learn to

Page 6: arXiv:1803.06503v1 [cs.CV] 17 Mar 2018 · Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low-level features such as color,

0.0 0.2 0.4 0.6 0.8 1.00.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0Pr

ecision

Recall

BSCA GS GC HS MR SF MB+ MST OURS

PASCAL-S

0.0 0.2 0.4 0.6 0.8 1.00.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

Precision

Recall

BSCA GS GC HS MR SF MB+ MST OURS

DUT-OMRON

0.0 0.2 0.4 0.6 0.8 1.00.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

Precision

Recall

BSCA GS GC HS MR SF MB+ MST OURS

HKU-IS

Figure 4: Comparison of precision-recall curves among 9 salient region detection methods on 3 datasets. Our proposed ASMOconsistently outperforms other methods across all the testing datasets.

Data Set Metric GS SF HS MR GC BSCA MB+ MST ASMO ASMO+maxF 0.777 0.700 0.813 0.824 0.719 0.830 0.822 0.809 0.890 0.896MSRA-B MAE 0.144 0.166 0.161 0.127 0.159 0.130 0.133 0.098 0.067 0.068maxF 0.661 0.548 0.727 0.736 0.597 0.758 0.736 0.724 0.837 0.845ECSSD MAE 0.206 0.219 0.228 0.189 0.233 0.183 0.193 0.155 0.110 0.112maxF 0.682 0.590 0.710 0.714 0.588 0.723 0.727 0.707 0.846 0.855HKU-IS MAE 0.166 0.173 0.213 0.174 0.211 0.174 0.180 0.139 0.086 0.088maxF 0.556 0.495 0.616 0.610 0.495 0.617 0.621 0.588 0.722 0.732DUT-OMRON MAE 0.173 0.147 0.227 0.187 0.218 0.191 0.193 0.161 0.101 0.100maxF 0.620 0.493 0.641 0.661 0.539 0.666 0.673 0.657 0.752 0.758PASCAL-S MAE 0.223 0.240 0.264 0.223 0.266 0.224 0.228 0.194 0.152 0.154maxF 0.620 0.516 0.646 0.636 0.526 0.654 0.658 0.647 0.751 0.758SOD MAE 0.251 0.267 0.283 0.259 0.284 0.251 0.255 0.223 0.185 0.187

Table 1: Comparison of quantitative results including maximum F-measure (larger is better) and MAE (smaller is better). Thebest three results on each dataset are shown in red, blue, and green , respectively.

mine the ambiguity inside the original noisy labels and thesemantic annotation ambiguity across different images, cor-rect them, and eventually produce an optimized results farbetter than the original ones. As a part of quantitative eval-uation, we show a comparison of PR curves in Fig. 4, asshown in the figure, our method significantly outperformsall state-of-the-art unsupervised salient object detection al-gorithms. Moreover, a quantitative comparison of maximumF-measure and MAE is listed in Table. 1. Our proposedalternate saliency map optimization (ASMO) improves themaximum F-measure achieved by the best-performing exist-ing algorithm by 8.74%, 11.48%, 17.61%, 17.87%, 12.63%and 15.20% respectively on MSRA-B, ECSSD, HKU-IS,DUT-OMRON, PASCAL-S and SOD. And at the same time,it lowers the MAE by 31.63%, 29.03%, 38.13%, 31.97%,21.65% and 17.04% respectively on MSRA-B, ECSSD,HKU-IS, DUT-OMRON, PASCAL-S and SOD. We alsoevaluate the performance of further applying dense CRF toour proposed method, listed as ASMO+ in the table.

Ablation StudiesEffectiveness of Alternate Saliency Map OptimizationOur proposed Multi-FCN based saliency map optimizationframework is composed of two components, a multi-taskfully convolutional network (Multi-FCN) and a graphicalmodel based on conditional random fields (CRF). To showthe effectiveness of the proposed optimization method, we

compare the saliency map S1 generated from the originalmethod, the saliency map S2 from directly employing denseCRF to the original method, the saliency map S3 from train-ing Multi-FCN with generated saliency maps but withoutCRF or CAM guided, the saliency map S4 from trainingMulti-FCN with generated saliency maps and CRF guidedbut without CAM and the saliency map S5 from our fullpipeline using DUT-OMRON dataset. As shown in Tab. 3,employing dense CRF operation and iterative training onfully convolutional ResNet-101 with original saliency mapas noisy groundtruth can both boost the performance ofthe original unsupervised method. Our alternately updatingscheme can integrate both of these two complementary ad-vantages which further gains 5.22% improvement on max-imum F-measure and 16.6% decrease on MAE. Moreover,CAM guided saliency annotations updating scheme plays aparamount role in our optimization framework which alsogreatly improve the saliency map performance.

Sensitivities to Benchmark Method Selection As de-scribed in Section , our proposed alternate saliency map op-timization method is based on an unsupervised benchmarkmodel as initialization. To demonstrate that our proposedmethod is widely applicable to the optimization of any unsu-pervised salient object detection method, we apply our op-timization method to the other two recently published unsu-pervised saliency detection methods, including BSCA (Qin

Page 7: arXiv:1803.06503v1 [cs.CV] 17 Mar 2018 · Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low-level features such as color,

(j)Ours(a)Source (k)GT(e)MR(d)HS (f)GC(b)GS (c)SF (g)BSCA (h)MB+ (i)MST

Figure 5: Visual comparison of saliency maps from state-of-the-art methods. The ground truth (GT) is shown in the last column.Our proposed method consistently produces saliency maps closest to the ground truth.

Table 2: Comparison with fully supervised salient object detection methods on HKU-IS, ECSSD and PASCAL-S datasets.

Data Set Metric DRFI LEGS MC MDF RFCN DHSNet DCL+ ASMO+ ASMO+(with mask)

maxF 0.776 0.770 0.798 0.861 0.896 0.892 0.904 0.855 0.913HKU-IS MAE 0.167 0.118 0.102 0.076 0.073 0.052 0.049 0.088 0.041maxF 0.782 0.827 0.837 0.847 0.899 0.907 0.901 0.845 0.918ECSSD MAE 0.170 0.118 0.100 0.106 0.091 0.059 0.068 0.112 0.057maxF 0.690 0.752 0.740 0.764 0.832 0.824 0.822 0.758 0.847PASCAL-S MAE 0.210 0.157 0.145 0.145 0.118 0.094 0.108 0.154 0.092

Table 3: Effectiveness evaluation of different componentsof alternate saliency map optimization on DUT-OMRONdataset.

Metric S1 S2 S3 S4 S5 S5+CRFmaxF 0.588 0.630 0.651 0.685 0.722 0.732MAE 0.161 0.178 0.151 0.126 0.101 0.100

et al. 2015) and MST (Tu et al. 2016). Experimental re-sults in Tab. 4 have shown that although these methods haveachieved very good results, the application of our proposedoptimization algorithm can still significantly improve theirperformance.

Evaluation on Semi-Supervised Setting In this section,we aim to compare our proposed method with the state-of-the-art fully supervised methods. As shown in Tab. 2, ourproposed weakly supervised ASMO with CRF already con-sistently outperforms 3 fully supervised methods includingDRFI (Jiang et al. 2013), LEGS (Wang et al. 2015) andMC (Zhao et al. 2015), and it is comparable to MDF (Liand Yu 2016b). Particularly noteworthy that when we addthe groundtruth mask of the training set of MSRA-B datasetto form a semi-supervised setting of our method, it greatlyoutperforms all state-of-the-art fully supervised methodsacross all the three testing datasets (HKU-IS, ECSSD andPASCAL-S). We conjecture that our model considered more

Table 4: Evaluation of different benchmark methods in alter-nate saliency map optimization on DUT-OMRON dataset.

Metric MB+ ASMO(MB+) BSCA ASMO

(BSCA) MST ASMO(MST)

maxF 0.621 0.722 0.617 0.685 0.588 0.691MAE 0.193 0.101 0.191 0.121 0.161 0.126

semantic information than existing fully-supervised modelsas we additionally included the Microsoft COCO dataset inour initial training.

ConclusionsIn this paper, we have introduced a generic alternate opti-mization framework to improve the saliency map quality ofany unsupervised salient object detection methods by alter-nately exploiting a graphical model and training a multi-taskfully convolutional network for model updating. Experimen-tal results demonstrate that our proposed method greatly out-performs all state-of-the-art unsupervised saliency detectionmethods and can be comparable to the current best strongly-supervised methods training with thousands of pixel-levelsaliency map annotations on all public benchmarks.

References[Avidan and Shamir 2007] Avidan, S., and Shamir, A. 2007.Seam carving for content-aware image resizing. ACM Trans-

Page 8: arXiv:1803.06503v1 [cs.CV] 17 Mar 2018 · Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low-level features such as color,

actions on graphics 26(3):10.[Chen et al. 2014] Chen, L.-C.; Papandreou, G.; Kokkinos,I.; Murphy, K.; and Yuille, A. L. 2014. Semantic image seg-mentation with deep convolutional nets and fully connectedcrfs. arXiv preprint arXiv:1412.7062.

[Chen et al. 2015] Chen, L.-C.; Yang, Y.; Wang, J.; Xu,W.; and Yuille, A. L. 2015. Attention to scale:Scale-aware semantic image segmentation. arXiv preprintarXiv:1511.03339.

[Cheng et al. 2015] Cheng, M.-M.; Mitra, N. J.; Huang, X.;Torr, P. H.; and Hu, S.-M. 2015. Global contrast basedsalient region detection. IEEE Transactions on Pattern Anal-ysis and Machine Intelligence 37(3):569–582.

[Deng et al. 2009] Deng, J.; Dong, W.; Socher, R.; Li, L.-J.;Li, K.; and Fei-Fei, L. 2009. Imagenet: A large-scale hier-archical image database. In Proceedings of IEEE ComputerVision and Pattern Recognition, 248–255.

[He et al. 2015] He, K.; Zhang, X.; Ren, S.; and Sun, J. 2015.Deep residual learning for image recognition. arXiv preprintarXiv:1512.03385.

[Jiang et al. 2013] Jiang, H.; Wang, J.; Yuan, Z.; Wu, Y.;Zheng, N.; and Li, S. 2013. Salient object detection: Adiscriminative regional feature integration approach. In Pro-ceedings of IEEE Computer Vision and Pattern Recognition,2083–2090.

[Krahenbuhl and Koltun 2012] Krahenbuhl, P., and Koltun,V. 2012. Efficient inference in fully connected crfs withgaussian edge potentials. arXiv preprint arXiv:1210.5644.

[Li and Yu 2016a] Li, G., and Yu, Y. 2016a. Deep contrastlearning for salient object detection. In Proceedings of theIEEE Conference on Computer Vision and Pattern Recogni-tion, 478–487.

[Li and Yu 2016b] Li, G., and Yu, Y. 2016b. Visual saliencydetection based on multiscale deep cnn features. IEEETransactions on Image Processing 25(11):5012–5024.

[Li et al. 2014] Li, Y.; Hou, X.; Koch, C.; Rehg, J. M.; andYuille, A. L. 2014. The secrets of salient object segmenta-tion. In Proceedings of IEEE Computer Vision and PatternRecognition, 280–287.

[Lin et al. 2014] Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.;Perona, P.; Ramanan, D.; Dollar, P.; and Zitnick, C. L. 2014.Microsoft coco: Common objects in context. In Proceedingsof the European Conference on Computer Vision, 740–755.

[Lin et al. 2015] Lin, G.; Shen, C.; Reid, I.; et al. 2015. Ef-ficient piecewise training of deep structured models for se-mantic segmentation. arXiv preprint arXiv:1504.01013.

[Liu and Han 2016] Liu, N., and Han, J. 2016. Dhsnet: Deephierarchical saliency network for salient object detection. InProceedings of the IEEE Conference on Computer Visionand Pattern Recognition, 678–686.

[Liu et al. 2011] Liu, T.; Yuan, Z.; Sun, J.; Wang, J.; Zheng,N.; Tang, X.; and Shum, H.-Y. 2011. Learning to detect asalient object. IEEE Transactions on Pattern Analysis andMachine Intelligence 33(2):353–367.

[Ma et al. 2002] Ma, Y.-F.; Lu, L.; Zhang, H.-J.; and Li, M.2002. A user attention model for video summarization. In

Proceedings of the tenth ACM international conference onMultimedia, 533–542.

[Mahadevan and Vasconcelos 2009] Mahadevan, V., andVasconcelos, N. 2009. Saliency-based discriminant track-ing. In Proceedings of the IEEE Conference on ComputerVision and Pattern Recognition, 1007–1013.

[Martin et al. 2001] Martin, D.; Fowlkes, C.; Tal, D.; andMalik, J. 2001. A database of human segmented naturalimages and its application to evaluating segmentation algo-rithms and measuring ecological statistics. In Proceedingsof International Conference on Computer Vision, volume 2,416–423.

[Perazzi et al. 2012] Perazzi, F.; Krahenbuhl, P.; Pritch, Y.;and Hornung, A. 2012. Saliency filters: Contrast based fil-tering for salient region detection. In Proceedings of IEEEComputer Vision and Pattern Recognition, 733–740.

[Qin et al. 2015] Qin, Y.; Lu, H.; Xu, Y.; and Wang, H. 2015.Saliency detection via cellular automata. In Proceedings ofIEEE Computer Vision and Pattern Recognition, 110–119.

[Shon et al. 2005] Shon, A. P.; Grimes, D. B.; Baker, C. L.;Hoffman, M. W.; Zhou, S.; and Rao, R. P. 2005. Prob-abilistic gaze imitation and saliency learning in a robotichead. In Proceedings of the IEEE International Conferenceon Robotics and Automation, 2865–2870.

[Tu et al. 2016] Tu, W.-C.; He, S.; Yang, Q.; and Chien, S.-Y. 2016. Real-time salient object detection with a minimumspanning tree. In Proceedings of IEEE Computer Vision andPattern Recognition, 2334–2342.

[Wang et al. 2015] Wang, L.; Lu, H.; Ruan, X.; and Yang,M.-H. 2015. Deep networks for saliency detection via lo-cal estimation and global search. In Proceedings of IEEEComputer Vision and Pattern Recognition, 3183–3192.

[Wang et al. 2016] Wang, L.; Wang, L.; Lu, H.; Zhang, P.;and Ruan, X. 2016. Saliency detection with recurrent fullyconvolutional networks. In Proceedings of the EuropeanConference on Computer Vision, 825–841.

[Wei et al. 2012] Wei, Y.; Wen, F.; Zhu, W.; and Sun, J. 2012.Geodesic saliency using background priors. In Proceedingsof the European Conference on Computer Vision, 29–42.

[Yan et al. 2013] Yan, Q.; Xu, L.; Shi, J.; and Jia, J. 2013. Hi-erarchical saliency detection. In Proceedings of IEEE Com-puter Vision and Pattern Recognition, 1155–1162.

[Yang et al. 2013] Yang, C.; Zhang, L.; Lu, H.; Ruan, X.; andYang, M.-H. 2013. Saliency detection via graph-based man-ifold ranking. In Proceedings of IEEE Computer Vision andPattern Recognition, 3166–3173.

[Zhang et al. 2015] Zhang, J.; Sclaroff, S.; Lin, Z.; Shen, X.;Price, B.; and Mech, R. 2015. Minimum barrier salientobject detection at 80 fps. In Proceedings of InternationalConference on Computer Vision, 1404–1412.

[Zhao et al. 2015] Zhao, R.; Ouyang, W.; Li, H.; and Wang,X. 2015. Saliency detection by multi-context deep learn-ing. In Proceedings of IEEE Computer Vision and PatternRecognition, 1265–1274.

[Zhou et al. 2016] Zhou, B.; Khosla, A.; Lapedriza, A.;Oliva, A.; and Torralba, A. 2016. Learning deep features for

Page 9: arXiv:1803.06503v1 [cs.CV] 17 Mar 2018 · Cheng et al. 2015; Tu et al. 2016; Zhang et al. 2015; Yang et al. 2013). These methods are usually based on low-level features such as color,

discriminative localization. In Proceedings of IEEE Com-puter Vision and Pattern Recognition, 2921–2929.