Log In Sign Up

Weakly Supervised Top-down Salient Object Detection

by   Hisham Cholakkal, et al.

Top-down saliency models produce a probability map that peaks at target locations specified by a task/goal such as object detection. They are usually trained in a fully supervised setting involving pixel-level annotations of objects. We propose a weakly supervised top-down saliency framework using only binary labels that indicate the presence/absence of an object in an image. First, the probabilistic contribution of each image region to the confidence of a CNN-based image classifier is computed through a backtracking strategy to produce top-down saliency. From a set of saliency maps of an image produced by fast bottom-up saliency approaches, we select the best saliency map suitable for the top-down task. The selected bottom-up saliency map is combined with the top-down saliency map. Features having high combined saliency are used to train a linear SVM classifier to estimate feature saliency. This is integrated with combined saliency and further refined through a multi-scale superpixel-averaging of saliency map. We evaluate the performance of the proposed weakly supervised top-down saliency against fully supervised approaches and achieve state-of-the-art performance. Experiments are carried out on seven challenging datasets and quantitative results are compared with 36 closely related approaches across 4 different applications.


page 2

page 3

page 4

page 8

page 9

page 11

page 12

page 14


Joint Learning of Saliency Detection and Weakly Supervised Semantic Segmentation

Existing weakly supervised semantic segmentation (WSSS) methods usually ...

Real Time Image Saliency for Black Box Classifiers

In this work we develop a fast saliency detection method that can be app...

Region Proposals for Saliency Map Refinement for Weakly-supervised Disease Localisation and Classification

The deployment of automated systems to diagnose diseases from medical im...

Bridging Saliency Detection to Weakly Supervised Object Detection Based on Self-paced Curriculum Learning

Weakly-supervised object detection (WOD) is a challenging problems in co...

Energy-Based Generative Cooperative Saliency Prediction

Conventional saliency prediction models typically learn a deterministic ...

Saliency Guided End-to-End Learning for Weakly Supervised Object Detection

Weakly supervised object detection (WSOD), which is the problem of learn...

Transformer Transforms Salient Object Detection and Camouflaged Object Detection

The transformer networks, which originate from machine translation, are ...

I Introduction

The human visual system has the ability to zero-in rapidly onto salient regions in an image. Recently, there has been much interest among computer vision researchers to model this process known as

visual saliency, which is attributed to the phenomenon of visual attention. It is beneficial in applications such as object detection/segmentation [1, 2], image retargeting [3] etc., since identification of salient regions reduces the search space for such high-level tasks. Salient regions in an image are indicated by a probability map called the saliency map. Fig. 1 shows saliency maps in the form of heat maps, where red indicates higher saliency.

In many instances, the salient region corresponds to a specific object in an image, in which case salient object detection becomes a more apt term, wherein pixels belonging to a salient object are assigned high saliency values. Broadly, there are two approaches to salient object detection: bottom-up (BU) [4] and top-down (TD) [5]. The feature contrast at a location plays the central role in BU salient object detection, with no regard to the semantic contents of the scene, although high-level concepts like faces have been used in conjunction with visual cues like color and shape [6]. The assumption that the salient object ‘pops out’ does not hold when there is little or no contrast between the object and the background. Furthermore, the notion of a salient object is not well-defined in BU models as seen in Fig. 1(b, c, d) where BU methods [7], [4] and [8] show the potted plant in the background as salient to a user searching for the cat.

TD salient object detection is task-oriented and utilizes prior knowledge about the object class. For example, in semantic segmentation [9], a pixel is assigned to a particular object class, and a saliency map that aids in this segmentation must invariably be generated by a top-down approach. Fig. 1(e, f, g, h) show the saliency maps produced by the proposed method for person, cat, sofa and potted plant categories, respectively. TD saliency is also viewed as a focus-of-attention mechanism by which BU salient points that are unlikely to be part of the object are pruned [10].

Most methods for TD saliency detection learn object classes in a fully supervised manner using pixel-level labeling of objects [5, 11, 12]

. Weakly supervised learning (WSL) alleviates the need for user-intensive annotation by utilizing only class labels for images. Moosmann 

et al. [13] propose a weakly supervised TD saliency method for image classification that employs iterative refinement of object hypothesis on a training image. Our method does not require any iterations, yet achieves better results compared to even fully supervised approaches [14, 5].

In the preliminary version of this paper presented at CVPR 2016 [15]

, we introduced a novel backtracking strategy on multi-scale spatial pyramid max-pooling (SPP)-based image classifier. The objective is to analyze the contribution of a feature towards the final classifier score which is then utilized to generate the TD saliency map for an object. Following our approach

[15], a weakly supervised saliency approach was proposed in [16]

, which uses excitation backpropagation to identify the task-relevant neurons in a convolutional neural network (CNN). It produces attention maps that highlight discriminative regions of the object as in

[17], [18], and are only useful for applications such as object localization. The proposed method highlights the entire object, which enables it to be used in applications such as weakly supervised semantic segmentation and pixel-accurate salient object detection. We achieve this through the following novel contributions:

(i) A strategy to backtrack multi-scale spatial pyramid pooled (SPP) CNN features; (ii) a feature saliency module that assigns non-zero saliency values at non-discriminative object regions; (iii) a saliency-weighted max-pooling strategy to select a BU saliency map that is better suited for a given task. To the best of our knowledge, this is the first work that demonstrates usefulness of top-down salient object detection for a wide range of applications, starting from coarse-level object localization to pixel-accurate semantic segmentation and category-independent salient object detection.

The major differences in the current version with [15] are: (i) sparse codes of SIFT features in [15] are replaced with CNN features; (ii) since CNN features span larger spatial neighborhood compared to SIFT features, contextual saliency in [15] is replaced with CNN feature saliency; (iii) [15] considers only TD saliency, while the current version proposes a novel strategy to select a BU saliency map among several candidates, which is then combined with TD saliency map; (iv) multi-scale averaging of saliency values within each superpixel is carried out to improve accuracy along object boundaries; and (v) extensive experiments are performed on seven challenging datasets and quantitative results are compared with 40 closely related approaches across 4 different applications. These modifications lead not only to better performance than [15], but also with fully supervised TD approaches as shown in Fig. 2.

We first train a CNN image classifier using image-level representation of CNN features, that gives a confidence score on the presence of an object in an image. The probabilistic contribution of each discriminative feature to this confidence score is represented in a TD saliency map, which is combined with a BU saliency map that is selected from several candidate BU maps through a novel selection strategy. Next, the saliency of each feature is separately evaluated using a dedicated feature classifier, as a means to assign non-zero saliency values to features from non-discriminative object regions, based on their dissimilarity with the background features. Saliency inference at a pixel involves combining the image classifier-based saliency map and the feature classifier-based saliency map.

Fig. 1: Comparison of proposed top-down salient object detection with bottom-up methods. (a) Input image, bottom-up saliency maps of (b) MB [7], (c) MST [4], and (d) HC [8]; proposed top-down saliency maps for (e) person (f) cat (g) sofa and (h) potted plant categories.

(a)                    (b)                   (c)                    (d)                    (e)                   (f)

Fig. 2: Visual comparison of the proposed weakly supervised approach with fully supervised top-down saliency approaches. (a) Input image, top-down saliency maps of (b) Kocak et al. [11], (c) LCCSC [12], (d) Yang and Yang [5], (e) Exemplar [14] and (f) proposed method for cat (top row) and cow (bottom row) categories.

Ii Related Work

We review related work in top-down saliency and relevant applications of CNN under weak supervision.

Ii-a Top-down saliency frameworks

Kanan et al. [19] proposed a TD saliency approach which uses object appearance in conjunction with a location prior. Ineffectiveness of this prior largely affects the accuracy based on the position of the object within the image. Closer to our framework, Yang and Yang [5] proposed a fully supervised TD saliency model that jointly learns a conditional random field (CRF) and dictionary using sparse codes of SIFT features as latent variables. The inability to discriminate between objects having similar parts (e.g. wheels of car and motorbike) causes a large number of false detections. Kocak et al. [11] improved upon this by replacing SIFT features with the first and second order statistics of color, edge orientation and pixel location within a superpixel, along with objectness [20]. Although this improved the accuracy in distinguishing objects from background, it failed to discriminate between object categories, causing large number of false detections if the test image contained objects from other categories, as shown in Fig. 2(b). Blocking artifacts are also observed in the saliency map at the superpixel boundaries because the superpixels are extracted on a single scale alone. Khan and Tappen [21] used label and location-dependent smoothness constraint in a sparse code formulation to produce a smooth saliency map compared to conventional sparse coding, but with additional computational cost. A joint framework for image classification and TD saliency is proposed in [22].

Zhu et al. [23] proposed a contextual-pooling based approach where LLC [24] codes of SIFT features are max-pooled in a local neighborhood followed by log-linear model learning. By replacing LLC codes with locality-constrained contextual sparse coding (LCCSC), Cholakkal et al. [12] improved on [23] with a carefully chosen category-specific dictionary learned from the annotated object area. Discriminative models [25, 26, 10] often represent a few patches on the object as salient and not the entire object. Hence, such models end up with low recall rates compared to [11, 27]. In [26], the task of image classification is improved using discriminative spatial saliency to weight visual features.

In [14], a fully supervised, CNN-based TD saliency method was proposed that utilized visual association of query images with multiple object exemplars. They followed a two-stage deep model where the first stage learnt object-to-object association and the second stage learnt object-to-background discrimination. Each patch, extracted using a sliding window, is resized to and input separately to the CNN. There are approximately 500 patches in an image of size , resulting in 500 forward passes through the network. Training the model required more than a week on a GPU. Our approach needs only one forward pass to extract CNN features for the entire image, which reduces the computation time significantly. It is still able to produce better saliency maps (Fig. 2(f)) compared to [14] (Fig. 2(e)). CNN-based saliency approaches [28, 29, 30, 31] learn category-independent salient features [32] from a large number of fully annotated training images [8]. Training or fine-tuning these saliency models  [29, 31]

took multiple days, even after initializing their models with convolutional filter weights pre-trained for image classification on ImageNet  


The use of weak supervision in TD saliency has largely been left unexamined. Gao et al. [10] used a weakly supervised setting where bottom-up features are combined with discriminative features that maximize the mutual information to the category label. In [13], a joint framework using classifier and TD saliency is used for object categorization by sampling representative windows containing the object. Their iterative strategy leads to inaccurate saliency estimation if the initialized windows do not contain the object. In [18, 16, 34], different variants of CNN backpropagation are used to identify the image regions responsible for activations corresponding to an object class (salient regions). In these discriminative saliency approaches [10, 18, 16], higher saliency values are assigned only to the features that are discriminative for a category in an image classification task, limiting their use in applications such as object segmentation, where all pixels of the object need to be identified accurately. [35] improved upon [18] by evaluating CNN derivatives with respect to feature maps of the intermediate convolutional layer instead of input image. Additionally, they train a fully-connected CRF for semantic segmentation by using the saliency maps as unary potentials. We use the pre-trained CNN only as a feature extractor, and the proposed backtracking strategy is applied only from the linear SVM weights to the multi-scale spatial pyramid pooling layer. Backpropagation of multi-scale spatial pyramid pooling layer is not yet explored for saliency estimation [16, 18, 35].

Ii-B CNN-based weakly supervised frameworks

Recently, CNN has been used in a number of weakly supervised object localization approaches [36, 37, 38, 39]. Multiple-instance learning is applied on CNN features in [38]. In [36], image regions are masked out to identify regions causing maximal activation. The outputs of CNN on multiple overlapping patches are utilized for object localization in [39]. All these approaches need multiple forward passes on a network to localize objects, which makes them computationally less efficient. Oquab et al. [37] applied global max-pooling to localize a point on objects. Global max-pooling is replaced by average pooling in [40] to help identify the full extent of the object as well. The underlying assumption is that the loss for average pooling enables the network to identify discriminative object regions. However, the spatial information is lost, whereas it is retained in our framework via multi-scale spatial pyramid pooling in the image classifier. The image classifier weights are reused for localization in [40]. We learn an additional feature classifier to better estimate saliency at non-discriminative object regions.

A weakly supervised, end-to-end CNN architecture is proposed in [41] for simultaneous object detection and image classification. Object detection requires classification of a large number of category-independent object proposals [42], [43]

. On a test image, the CNN features are extracted on the original and flipped image at five scales totaling to 10 feature extraction iterations. In

[44], category-specific object proposals are shown to be effective for weakly supervised object detection. Recent semantic segmentation, co-saliency and co-segmentation frameworks [45, 9, 46, 47] also train their CNN models in a weakly supervised setting.

Internal representations learned by CNN are visualized in [48, 49, 50, 51, 17, 34] for better understanding of its properties. [48] and [51] analyze the convolution layers using techniques such as deconvolutional networks. In [49] and [50], CNN features are inverted at different layers of the network including the fully connected layers, to analyze the visual encoding of CNN.

Iii Proposed method

Fig. 3: Illustration of combined saliency estimation for dog category. Red arrows indicate the proposed backtracking strategy for top-down saliency (B-cSPP). From a set of BU saliency maps (MB [7], HC [8], MST [4] and 3-Max), the best one is selected and is integrated with B-cSPP saliency to produce combined saliency. 3-Max is the saliency map obtained by taking the maximum saliency at each pixel across the 3 BU saliency maps.

In this section, we first describe the CNN image classifier and how the backtracking mechanism generates the TD saliency map. Next, a novel strategy to select a BU map is described followed by its integration with the TD saliency map to obtain combined saliency. A dedicated feature saliency model is learned on features extracted using this combined saliency to evaluate saliency at each feature independently. Finally, inference involves combining the feature saliency with the combined saliency to obtain the TD map.

Iii-a CNN Image Classifier

The CNN features of an image are extracted using VGG-16 [52] that has been pre-trained on the ImageNet ILSVRC 2012 data [33]

with only image-level supervision (no bounding box annotation). All the layers of VGG-16 upto relu5_3 (third rectified linear unit in the fifth layer) are used for feature extraction and the convolution weights pre-trained for image classification are used without any fine-tuning. However, unlike

[52, 53], we do not crop/resize the input image for feature extraction, but use the original image at its full resolution. A fixed-length, image-level representation of CNN features is obtained through a multi-scale spatial pyramid max-pooling as in Spatial Pyramid Pooling (SPP-net) [54] image classifier. We use a binary linear SVM after the spatial pyramid pooling layer, instead of fully connected layers in [54]. In order to reduce computations, we extract CNN features at a single image scale instead of multiple scales.

The spatial dimensions of an input image are down sampled by a factor of 16 at the relu5_3 feature map due to spatial max-pooling in the first four layers. There are (=512) convolution filters in conv5_3 (third convolution in the fifth layer) which results in activation maps at the output of succeeding relu5_3. We consider the activation at each spatial location of relu5_3 activation maps as a

dimensional feature vector, which we refer to as relu5 feature. The relu5 feature represents the overall response of multiple pixels from its receptive field in the original image.

Let denote relu5 features each of dimension . The spatial distribution of the features in the image is encoded in the spatial pyramid max-pooled image vector through a multi-scale max-pooling operation of the relu5 features on a 3-level spatial pyramid [55] as shown in Fig. 3. The element of is a max-pooled value derived using maximum operation on element of all relu5 features in a spatial pyramid region defined by , and . i.e,


Let the label indicate the presence or absence of an object in the image. If , it is a positive image, else it is a negative image. Image-label pairs () of training images are used to train a binary linear SVM classifier [56, 57]

that minimizes the following loss function,


where and are the SVM weight vector and bias respectively and are learnt for each object category. is the length of the max-pooled image vector and is a regularization constant empirically set to 0.01 through cross validations.

Given a validation/test image with max-pooled vector , the classifier score indicates the confidence of the presence of object in it. Our CNN image classifier is referred as cSPP in the following discussions.

Iii-B Backtracking image classifier for saliency estimation

In the cSPP image classifier, both the linear-SVM and multi-scale max-pooling operations can be traced back to the relu5 feature locations. This enables us to analyze the contribution of each feature towards the final classifier score which is then utilized to generate the TD saliency map for an object.

First, we estimate the ability of a relu5 feature to represent a spatial pyramid region by counting the number of times the elements of that feature made it to the spatial pyramid max-pooled vector. We call this the representativeness, , of a feature . Representative features may either contribute positively or negatively to the classifier score with higher contribution indicating more relevance of the feature to an object . The relevance of the feature to the object is denoted .

For a positively contributing relu5 feature, it is possible that there are some negatively contributing elements among the total elements. For example, let be the feature with its element being a local maximum in its spatial pyramid region. Although contributes positively to the classifier confidence , the other non-zero elements or may contribute negatively, indicating absence of the object. So, the relevance of a feature to the object requires its contribution to be computed in the absence of other features; this relevance is denoted . We define the top-down saliency of the feature to the object class as the joint probability of three parameters - and as


The representative elements of the feature are identified as


where is the Kronecker delta function and the location of in identifies the region in the spatial pyramid and its position in the relu5 feature . is the inverse operation of multi-scale spatial pyramid max-pooling illustrated in Fig. 3 and written as


The probability of representativeness of the feature to the image is then defined as


where is the cardinality.

The linear SVM classifier confidence is a score indicating the presence of the object in the image, which increases from a definite absence () to definite presence (). The confidence of is

where is the contribution of feature to the image classifier confidence and is the length of .

Given that the feature is representative of the image, the probability of it belonging to the object is



is the sigmoid function that maps the confidence score to [0, 1] and ensures that the probabilities of

belonging to object and not belonging to object sums to 1.

Using the above probabilities, we select a set of all features that contribute positively to the classifier confidence as


The net contribution of a feature in the absence of other features is


where is the spatial pyramid max-pooling operation performed by replacing all features except with a zero vector of size to form max-pooled vector .

Implementation. Fig. 3 illustrates three relu5 features , and . The confidence of the presence of object in an image is indicated by the classifier score as mentioned in the previous section and the element of has a corresponding weight . The elements from the Hadamard product with mark the features and that contribute positively to the classifier confidence through a operation, i.e the set . The contribution of feature in the absence of other features is evaluated using max-pooling operations in which all features except are replaced with forming max-pooled vector . The saliency of a feature is given by


Since this TD saliency of a feature is arrived at by backtracking the cSPP classifier, we call it B-cSPP saliency and the corresponding saliency map as B-cSPP saliency map. The feature from the object (dog) region is assigned high B-cSPP saliency while from background is assigned zero B-cSPP saliency.

Iii-C Selection of bottom-up saliency map

Contextual background regions help in improving image classification tasks [16, 18, 35]. For example, grass or sky in the background can help in classifying cow or bird category, respectively. Salient object detection should exclude such background regions. Bottom-up saliency assigns low saliency to these uniform regions that have low contrast with the surroundings. Hence, we multiply the B-cSPP saliency with a bottom-up saliency selected through a novel selection strategy, to form a combined saliency. In this section, we explain our novel strategy to select a saliency map which is best suited for the task at hand, from a set of BU maps based on a saliency weighted max-pooling.

State-of-the-art BU saliency approaches [7, 4] can produce a category-independent saliency map for an image within 40 milliseconds. They assume image boundaries as the background while approaches such as [8] focus on feature contrast to estimate saliency. These approaches do not require any training and give reasonably good results. Since BU saliency maps are task-independent from a user’s perspective, the definition of ‘good saliency map’ varies based on the application. For example, consider Fig. 1, where four different objects are present. If a user searches for a ‘person’ in the image, BU approaches [7, 4] that assume image boundary as the background fail to produce a ‘good saliency map’. In such scenarios, an approach [8] that does not use such assumptions can produce better results. Thus, our objective is to develop a strategy to select a BU saliency method for a particular image that is best suited for the task at hand.

Our cSPP image classifier which was trained to estimate the presence of object in an image is employed to select a BU saliency map suitable for the task of identifying image regions that belong to object . To achieve a one-to-one correspondence between pixels in the BU saliency map and the relu5 features, we downsample the saliency maps to the spatial resolution of feature map at relu5_3, i.e, by a factor of 16. From BU saliency maps, we need to select one for which features that belong to an object are assigned high saliency and those that do not belong to an object are assigned low saliency. For a max-pooled vector of an image, the SVM predicts a confidence score which is proportional to the confidence of object presence in that image. i.e,



Ideally, features belonging to object contribute positively to the classifier confidence and hence they correspond to elements in whose indices belong to , while the background features result in indices. It is to be noted that is non-negative since it is derived from relu5 through max-pooling operation.

First, the feature is weighted with , the BU saliency value for that feature estimated by approach. i.e, . The saliency-weighted relu5 features are used to estimate the saliency-weighted max-pooled vector and similar to Eq. (13), the modified confidence score due to the BU map is computed as,


If higher values in the saliency map produced by algorithm falls exactly on the object regions, the second summation will be largely reduced, due to weighting background indices with low saliency values and hence will be high. If some of the background also garners high saliency, then will be relatively low. In order to reinforce the above assertion, we invert the saliency map (by subtracting saliency values from the maximum saliency value in the image), and recompute the saliency-weighted relu5 features (), saliency-weighted max-pooled vector , and .

Let be the maximum value in the saliency map of an image by approach, then the inverted saliency value for location is


Then the feature is weighted with , resulting in , i.e.,


Repeating the above procedure for all relu5 features we obtain , the relu5 features weighted with inverted saliency map. Multi-scale spatial pyramid pooling ( ) of will result in saliency-weighted max-pooled vector . i.e,


Note that is a non-linear operation, hence . Finally, is


If all object regions are assigned with higher saliency values in Eq. (16), higher weights are assigned to the background regions and lower weights to the salient regions, leading to a lower score of . Combining the above two observations, an ideal saliency map should maximize


In order to prevent the selection of a map that assigns high saliency to the entire image, we impose a penalty of on saliency map with a mean saliency . Combining the above observations, the final objective function to select a BU saliency map is


If the saliency map of algorithm is not aligned with the object, then the false positives will increase and decrease in , thus increasing the second term of Eq. (18). False negatives will reduce and increase reducing the first term. Hence an inaccurate BU saliency map will result in low . The saliency map that maximizes Eq. (18) is selected. In right side of eqs. (16), (17) and (18), the saliency map of algorithm is implicit in and through and .

In addition to choosing individual BU saliency maps, we also analyze whether a combination of these maps has an effect on improving TD saliency. To this end, we combine saliency maps by picking the maximum saliency for each pixel and use Eq. (18) to select the best map from a set of saliency maps that includes the maximum map. In this section, we have assumed that the SVM weights learnt for an object is accurate and that the object appears only at locations where are positive. Although this may not be always true, we retain this assumption since object locations are not available in a weakly supervised setting.

Off-the-shelf BU methods can be used to generate the BU saliency maps. We used MB [7], MST [4] and HC [8] in our framework to improve the runtime performance. The B-cSPP saliency map and the selected bottom up saliency map are combined through a simple multiplication as shown in Fig. 3. We denote this combined saliency map as . Following [5, 10, 11], we also characterize our category-specific saliency inference framework as TD saliency even though there is a bottom-up component.

Iii-D Feature saliency training

Image classifiers trained on image-level representation of features have shown to be effective in discriminative TD saliency estimation [15, 26, 10]. The combined saliency map takes non-zero values only at discriminative image regions whose features make positive contribution to the image classifier confidence. The assumption is that the object appears only at grids in the spatial pyramid where are positive, which may not be true across all images. Our objective is not limited to identifying the discriminative image regions, but to assign higher saliency values to all pixels belonging to the salient object. In order to independently estimate the saliency value of each relu5 feature, we also learn a top-down feature saliency model that uses a linear SVM learnt on positive and negative relu5 features from the training images. Since feature-level annotation is not available, we use object features extracted using the combined saliency map to train the model.

From positive training images of object , relu5 features with saliency greater than are selected as positive features with label . In order to prevent training features from non-discriminative object regions of positive images with negative label, only those features at which both B-cSSP and BU saliency are selected as negative features with label . Additionally, random features are selected from negative images with label . A linear SVM model with weight and bias is learned. Since the relu5 features are already computed for B-cSPP, learning of linear SVM is the only additional computation required to train this top-down model. The saliency map obtained from feature saliency is denoted .

Iii-E Saliency inference

For inference on a test image, the combined saliency and feature saliency are first integrated followed by multi-scale superpixel averaging and finally associated with the confidence of the image classifier to obtain the saliency at a pixel. While the combined saliency is obtained as described in Section III-C, the feature saliency for a feature is the probability of the feature belonging to an object computed by applying a sigmod function to the linear SVM score,


The feature saliency and combined saliency values are integrated using a mean operation to form the saliency map, .

Iii-E1 Multi-scale superpixel-averaging of saliency map

The low resolution saliency map

is upsampled to the original image size using bicubic interpolation. As a consequence, saliency values may not be uniform within a superpixel. Also, the saliency map will not be edge-aware with object regions spreading to the background. Hence, a multi-scale superpixel-averaging strategy is employed. The mean saliency at a superpixel (obtained by SLIC segmentation  

[58]) is assigned to every pixel in it. This process is repeated at multiple scales by varying the SLIC parameters. The resulting maps are averaged to produce a smooth, pixel-level saliency map that uniformly highlights the salient object and also produces a sharp transition at object boundaries.

Iii-E2 Integrating with image classifier confidence

For a given image, the TD saliency map indicates the probable pixels that belong to object . Since the presence of a specific object in a test image is not known apriori for applications such as semantic segmentation and object detection, the saliency map needs to be estimated for both positive and negative images. Hence, it is beneficial to integrate with a confidence score that indicates the presence of object in at least one pixel in the image. For this, we use the same cSPP image classifiers learnt earlier for each category. The SVM associated with the cSPP image classifier gives a confidence score for a particular object as . These scores are scaled between and as


where, is the total number of categories. Unlike soft-max that sums to 1, we normalize the score with the maximum because multiple categories can simultaneously appear in an image such as in PASCAL VOC-2012 [59]. In such scenarios, softmax will end up assigning a lower value to all positive categories. However, our objective is to identify the relative confidence across categories, and assign to the most probable category. To reduce false detections from less probable categories, we assume values of that are less than as less important, and replace it with . This limits the number of probable object categories per image to less than 5 categories in most images, and hence the category-specific saliency map needs to be computed only for these few probable object categories. We compute the classifier-weighted, category-specific score for each object ,


Iii-E3 Category-independent salient object detection

The proposed category-specific TD saliency map in Eq. (21) can be used to compute the category-independent saliency value , by computing the maximum saliency value at each pixel (,) as

Since the bottom-up information is integrated to through the combined saliency map , the gives an accurate estimate of saliency maps under free-viewing condition.

Iv Experimental evaluation

We evaluate our weakly supervised saliency model on Graz-02 [60], PASCAL VOC-2012, PASCAL VOC-2007 [61] and PASCAL-S [32] datasets. Additionally, we use PASCAL VOC-2012 segmentation test set and validation set to evaluate weakly supervised semantic segmentation, Object Discovery dataset [62] to compare with semantic object selection and co-segmentation approaches, and validation set of PASCAL VOC-2012 detection challenge to evaluate object localization and object detection performance.

Graz-02 dataset contains 3 object categories and a background category with 300 images per category. We split the images into training and testing sets following [27]

, i.e., 150 odd numbered images from each category are used for training and 150 even numbered images from each category for testing. PASCAL VOC-2012 is another challenging dataset with category-specific annotations for 20 object categories. It has 5717 training images and 5823 validation images for image classification/object detection challenge. There are 1464 training images, 1449 validation images and 1456 test images for segmentation challenge. Our PASCAL VOC-2012 saliency models are trained using 5717 training images for image classification task. There are 210 test images in the segmentation challenge of PASCAL VOC-2007.

PASCAL-S is a widely used dataset to evaluate category-independent saliency models. It has 850 images picked from the validation set of PASCAL VOC-2010 [63] segmentation images. Given the segmented objects in an image, the ground truth salient objects are marked by twelve subjects under free-viewing condition. We use Object Discovery dataset [62] to evaluate object segmentation. The dataset has three object categories, namely airplane, car and horse. Apart from 100 test images per category, there are 461, 1206 and 779 additional images for airplane, car and horse, respectively.

Iv-a Analysis of proposed framework

(a)           (b)          (c)            (d)            (e)

Fig. 4: Qualitative results at individual stages of the proposed method. (a) Input image, (b) B-cSPP saliency map, (c) (b) + bottom-up saliency, (d) (c) + feature saliency, (e) (d) + superpixel averaging.

Iv-A1 Contribution of individual modules

Fig. 4 shows the visual comparison of the effect of each stage in the proposed method. For the input images in Fig. 4(a), image regions containing bird’s head and cow’s legs make positive contribution to their image classifiers and are, therefore, assigned high B-cSPP saliency in Fig. 4(b). Combining B-cSPP saliency with BU saliency removed false detections in B-cSPP saliency as shown in Fig. 4(c). Integration of feature saliency assigns higher saliency value to the non-discriminative object regions (Fig. 4(d)). Finally, the addition of the multi-scale superpixel-averaging improved the accuracy along object boundaries as shown in Fig. 4(e).

Fig. 5: Evaluation of individual stages of the proposed framework across 20 categories of PASCAL VOC-2012 using pixel-level precision rate at EER. The improvement in accuracy by the addition of each module is shaded.
Evaluation of individual components
Segmentation validation set
Graz-02 PASCAL-S
Without BU and without super pixel averaging.
Only top-down modules (B-cSPP+feature saliency)
50.2 69.1 42.8
With BU and without super pixel averaging.
(B-cSPP+BU+feature saliency)
59.6 78.1 55.0
Without BU and With super pixel averaging
(B-cSPP+feature saliency+superpixel averaging)
57.1 79.9 48.5
Full framework
(B-cSPP+BU+feature saliency+super pixel averaging)
61.4 82.5 67.5
TABLE I: Pixel-level precision rates at EER (%) to analyze usefulness of BU components

We evaluate the improvement in the mean precision rate (%) at EER at each stage of our framework. The evaluation is done across 20 object categories of PASCAL VOC-2012 segmentation-validation set. The contribution of each component in the proposed saliency model to the final accuracy is shown shaded in Fig. 5. The accuracy of B-cSPP saliency is 34.1%. On adding the BU map to yield combined saliency, the accuracy increased to 43.6%, demonstrating the effectiveness of the proposed BU selection strategy.

Training the saliency model using negative patches from positive training images improve the accuracy by 5% in [15]. In the proposed framework, the accuracy of the combined saliency map is improved by weighting B-cSPP with the selected BU map, which enabled us to train the feature saliency using negative patches from positive images. This resulted in an additional improvement of 3% in accuracy, totaling to 16% with the addition of feature saliency. This demonstrates that (i) learning a dedicated feature classifier plays an important role for TD saliency and (ii) combined saliency map and feature saliency map complement each other. A similar trend in improvement was observed in [15], where a feature classifier is learnt using contextual max-pooled sparse codes. Since relu5 features span larger spatial context compared to SIFT features computed on patches, contextual max-pooling on relu5 features is not required.

The feature saliency map and combined saliency map are integrated as . Other combinations such as and gave similar results with less than 1% variation in accuracy. Taking the product of both saliency maps reduced the accuracy by 6% as the combined saliency is often 0 in non-discriminative object regions and multiplication causes 0 values in such locations of the integrated map, disregarding feature saliency.

Finally, superpixel-averaging is applied at 6 different scales, by extracting 8, 16, 32, 64, 128 and 256 superpixels from an image. The saliency values at each pixel are further averaged across these 6 scales to get the saliency map . The quality of the saliency map at object boundaries is improved leading to 1.8% improvement in the accuracy, to obtain an accuracy of 61.4%. Since superpixel computation at multiple scales is time consuming relative to other modules, inference speed can be largely improved in applications such as object localization that do not require exact object boundaries by removing this step.

Iv-A2 Evaluation of category-independent components

There are two category-independent components in our framework: bottom-up selection and superpixel averaging. Effectiveness of these components are evaluated in this section. We evaluated the performance of top-down modules (B-cSPP, feature saliency) by removing both BU selection and superpixel averaging modules. From row 1 and row 2 of Table I, it can be observed that addition of BU component improves the accuracy by 9% in category-specific datasets Graz-02 and PASCAL VOC 2012, while it has larger impact on category-independent dataset PASCAL-S (12.2%).

Addition of multi-scale superpixel averaging to the framework with BU component helped to improve the accuracy by 1.8%, 4.4% and 12.5% on PASCAL VOC 2012, Graz-02 and PASCAL-S datasets (row 2 vs row 4), respectively. These results shows that having category-independent components in the framework helps in improving both category-specific and category-independent saliency detection tasks and it has higher impact on category-independent saliency detection.

We further analyzed the performance of super-pixel averaging on a framework without BU, and the results indicate that superpixel averaging has larger role in removing false positives in the top-down saliency detection through multi-scale averaging of saliency maps as seen by increased accuracy of 6.9% and 10.8% on PASCAL VOC 2012 and Graz-02 datasets respectively (row 1 vs row 3). It is to be noted that the proposed bottom-up selection strategy uses fast and training-free bottom-up saliency approaches and hence it is faster compared to super-pixel averaging (the other category-independent component in our framework) as explained in Section II of the supplementary material.

Iv-B Comparison with other approaches

Method SV Test set Bike Car Person Mean
1 - Yang and Yang [27] FS 59.4 47.4 49.8 52.2
2 - Kocak et al. [11] FS 59.9 45.2 51.5 52.2
3 - LCCSC [12] FS 69.1 58.4 58.2 61.9
4 - CG-TD [22] FS 64.4 50.9 56.4 57.2
5 - WS-SC[15] WS 64.0 45.1 55.2 54.8
6 - Proposed WS
80.5 61.4 75 .0 72.3
7 - MB [7] TF 54.67 39.03 52.04 48.58
8 - Aldavert et al. [64] FS 71.9 64.9 58.6 65.13
9 - Fulkerson et al. [65] FS 72.2 72.2 66.1 70.16
10 - Shape mask [66] FS 61.8 53.8 44.1 53.23
11 - Yang and Yang [27] FS 62.4 60 62 61.33
12 - Khan and Tappen [21] FS 72.1 - - -
13 - CG-TD [22] FS 67.3 59.8 57.1 61.4
14 - WS-SC[15] WS 67.5 56.5 57.56 60.5
15 - Proposed WS
84.1 81.5 81.8 82.5
TABLE II: Pixel-level precision rates at EER (%) on Graz-02.

Iv-B1 Graz-02 dataset

Method SV plane bike bird boat botl bus car cat chair cow table dog horse moto pers plant sheep sofa train tv Mean
Yang [27] FS 14.7 28.1 9.8 6.1 2.2 24.1 30.2 17.3 6.2 7.6 10.3 11.5 12.5 24.1 36.7 2.2 20.4 12.3 26.1 10.2 15.6
Kocak [11] FS 46.5 45.0 33.1 60.2 25.8 48.4 31.4 64.4 19.8 32.2 44.7 30.1 41.8 72.1 33.0 40.5 38.6 12.2 64.6 23.6 40.4
Exemplar [14] FS 55.9 37.9 45.6 43.8 47.3 83.6 57.8 69.4 22.7 68.5 37.1 72.8 63.7 69.0 57.5 43.9 66.6 38.3 75.1 56.7 56.2
Oquab [37] WS 48.9 42.9 37.9 47.1 31.4 68.4 39.9 66.2 27.2 54.0 38.3 48.5 56.5 70.1 43.2 42.6 52.2 34.8 68.1 43.4 48.1
Proposed WS 71.2 22.3 74.9 39.9 52.5 82.7 58.9 83.4 27.1 81.1 49.3 82.4 77.9 74.2 69.8 31.9 81.4 49.8 63.2 53.3 61.4
TABLE III: Pixel-level precision rates at EER on validation set of PASCAL VOC-2012 segmentation dataset.The proposed weakly supervised approach outperforms all fully supervised approaches including [14], which is based on CNN, in 14 out of 20 classes and in mean accuracy.

We report our pixel-level results on different test set configurations of Graz-02. First, the proposed saliency model is compared with other TD saliency algorithms [27, 11, 15, 22, 12] on all 600 test images. Second, for comparison with related approaches [66, 65], each object category is evaluated on test images from its respective category. Finally, to compare with [10, 19], results on 300 test images are evaluated, where 150 test images are from a single category and the remaining 150 are from the background.

The pixel-level comparisons in the first two test set configurations are shown in Table II, where SV indicates supervision level with WS, FS and TF referring to weakly supervised, fully supervised and training-free approaches respectively. [27, 11] and [22] are fully supervised (FS), needing multiple iterations of CRF learning with sparse codes relearned at each iteration. Separate dictionaries are used for each object category. On the contrary, the proposed weakly supervised method does not require any iterative learning and the relu5 features are extracted with a single forward pass on the CNN. [15] does not require any iterative learning and uses a smaller dictionary of 1536 atoms, compared to 2048 atoms used in [12]. Despite incorporating objectness [20] and superpixel features to [27], the discriminative capability of [11] did not improve (row 2 vs row 1). The proposed weakly supervised method (row 6 and row 15) outperforms all other fully supervised TD saliency approaches [27, 11, 22] .

With respect to the second test configuration, [66] requires images to be marked as difficult or truncated in addition to the object annotation for training of shape mask. [64] uses 500,000 dictionary atoms in their fully supervised framework to obtain 65.13% (row 8), whereas the dimension of our relu5 feature is only 512. In this test setting, the proposed method achieves a mean accuracy of 82.5% outperforming the previous weakly supervised approach [15] by 22%.

DSD [10] has limited capability to remove background clutter, resulting in poor performance of their model. DSD [10] and SUN [19] did not evaluate their model on Graz-02 dataset, but Yang and Yang [27]

reported their patch-level precision rates at EER on 300 test images as 49.4% and 53.3%, respectively. Feature learning using independent component analysis helped SUN to perform better than DSD, but substantially poorer than

[15] (65.4%) and the proposed method. It is to be noted that the performance of [10, 19, 27, 15] deteriorates while converting their patch-level results to pixel-level. The proposed weakly supervised method gives a mean pixel-level precision rate at EER of 73.1% which is better than the 70.16% and 70.49% reported by [11] and [12] respectively in this test setting. In all the three test settings, the proposed modifications enabled our current model to outperform [15] by more than 18% in accuracy, achieving state-of-the art performance. The use of CNN features contributes mainly to this performance boost. We could not compare with [18] and [16], since they did not report their salient object detection results. Qualitative comparisons with fully supervised and weakly supervised TD saliency approaches are shown in Fig. 2 and in the supplementary material, respectively.

Iv-B2 PASCAL VOC-2012 segmentation dataset

In Table III, we compare a CNN-based fully supervised TD saliency [14] with our method by evaluating on PASCAL VOC-2012 segmentation-validation set consisting of 1449 images. Similar to [14], each object category is evaluated only on positive images of that category. We did not fine-tune the convolution layers for this dataset, which took nearly 8 days on a GPU in [14]. The presence of multiple, visually similar object classes in a single image is challenging for a weakly supervised approach. Inspite of this, we outperform the state-of-the art fully supervised approach [14] and the CNN-based weakly supervised object localization approach [37] in mean accuracy by 5% and 13%, respectively. We outperform [37] in 15 out of the 20 categories. The top-down selection of BU approach along with feature saliency plays an important role in this improved performance, especially in classes like aeroplane and sheep.

Iv-B3 PASCAL VOC-2007 segmentation dataset

Following [12, 11] and [27], the saliency models are evaluated on 210 segmentation test images. We used the models trained on PASCAL VOC-2012 training set in this experiment. Separate sparse codes of size 512 are computed for each category in  [27, 11] and [22]. [15] uses sparse coding on a common dictionary of 1536 atoms for all object classes. Similarly, a common feature code of elements is used in [12]. In our method, we compute 512 dimensional relu5 features which are common for all object categories.

Table IV compares the pixel-level performance of the proposed WS method and patch-level results of FS top-down saliency approaches [27, 12] (these approaches did not report their pixel-level results on this dataset). We outperform [27], [22] and [12] in almost all categories and in mean precision rate at EER across 20 classes. A performance drop of 5 to 10% is reported by [21] while converting patch-level results of  [27] to pixel-level, which further increases the performance gain of the proposed approach.

Yang and
Yang [27]
LCCSC [12]
CG-TD [22]
Supervision FS FS FS WS WS
Mean of
20 classes
16.7 23.4 23.81 18.6 42.1
TABLE IV: Precision rates at EER(%) on PASCAL VOC-2007.

Khan and Tappen [21] report pixel-level precision rates at EER only for cow category (8.5%) which is much lower than the proposed weakly supervised approach (52.3%).

Iv-B4 Category-independent salient object detection

Fig. 6:

Comparison of the proposed weakly supervised approach with state-of-the-art category-independent saliency approaches on PASCAL-S dataset. We achieve a performance comparable with deep learning-based fully supervised approaches.

Category-independent saliency maps are obtained using the top-down models trained on PASCAL VOC-2012 training set through simple pixel-level maximum operation as explained in Section III-E3. Saliency values less than 0.5 are considered as background, and those between 0.5 and 1 are normalized to . The performance is evaluated on PASCAL-S dataset. Fig. 6 compares the proposed method against state-of-the-art category-independent approaches that include deep learning based fully supervised approaches such as MCDL [29], LEGS [30], MDF [31] and DHSNet [28]. The performance metric, F-measure is

where  [32]. Following [28]

, precision and recall are computed by binarizing each saliency map at an image adaptive threshold, which is twice the average value of the saliency map.

The proposed weakly supervised method achieves an f-measure of 67.5, which is comparable with fully supervised LEGS, MCDL and MDF. We use only 5717 images from PASCAL VOC-2012 training set, which is much smaller compared to the training data used by fully supervised approaches shown in Fig. 6. For example, DHSNet uses nearly 10,000 fully annotated images from multiple datasets such as MSRA 10K [67] and DUT-OMRON [68]. Data augmentation is used to further increase the number of training images. With less supervision and lesser training data, we achieve a performance comparable with these fully supervised approaches. Qualitative results are present in the supplementary material.

Iv-C Applications

(a) Semantic segmentation
(b) Object segmentation
(c) Object localization
(d) Object detection

Fig. 7: Applications of proposed saliency map for (a) Semantic segmentation, (b) object segmentation, (c) Object localization and (d) object detection.

TD saliency [5, 11, 15, 14] and tasks like object detection, localization and segmentation mainly differ in their granularity of representation. Object detection produces a tight rectangular bounding box around all instances of objects belonging to user-defined categories. It is necessary to identify both the location as well as the extent of each object. The process of identifying the location of a particular object in an image, without marking the extent of the object, is referred to as object localization [37]. Object segmentation, also referred to as semantic object selection produces a binary mask with ‘1’ indicating all pixels that belong to a user-defined object category. It differs from the task of semantic segmentation, where the objective is to classify each pixel in the image to one of predefined classes. In this section, we evaluate the use of our TD saliency framework for various applications in a weakly supervised setting.

Iv-C1 Weakly supervised semantic segmentation

The category-specific saliency maps in the proposed framework can be easily adapted for semantic segmentation. In the saliency map, a pixel with is less likely to belong to an object . The pixels at which the maximum saliency across all categories is less than 0.5 is more likely to be background. Hence, an additional map corresponding to the background category is generated as a uniform map with . We assign to each pixel the category for which its saliency is the maximum.

In PASCAL VOC-2012 semantic segmentation task, each pixel in the image needs to be classified to one of 21 categories comprising background and 20 object categories. Our results are reported from PASCAL VOC-2012 evaluation server which uses intersection-over-union (IoU) ) as the evaluation metric. The mean intersection-over-union (mIoU) on the validation set and test set are reported in Table 

V. The proposed approach requires the least number of training images (5k), yet it achieves a performance comparable with the state-of-the-art approaches [69, 70, 71, 35, 72] that use 10k to 50k training images. Unlike  [69, 70, 71, 35], we do not use CRF or Grab-cut [73] or similar energy minimization techniques to post-process the segmentation result.

Fig. 7 compares the qualitative results obtained by the proposed method against the ground truth. Majority of the person pixels are classified correctly, despite the size of person being small compared to motorbike in the top row.

mIoU on
val set
mIoU on
test set
MIL-FCN [74] 10k 25.7 24.9
CCNN [45] 10k 35.3 35.6
MIL-sppxl [75] 700k 36.6 35.8
MIL-bb [75] 700k 37.8 37.0
EM-Adapt [76] 10k 38.2 39.6
STC [72] 50k 49.8 51.2
DCSM w/o CRF [35] 10k 40.5 41.0
DCSM w/ CRF [35] 10k 44.1 45.1
BFBP w/ CRF [71] 10k 46.6 48.0
SEC w/ CRF [70] 10k 50.7 51.7
AF-SS w/ CRF [69] 10k 52.6 52.7
Proposed 5k 43.5 44.1
TABLE V: mean Intersection over union (mIOU) for semantic segmentation on validation and test sets of PASCAL VOC-2012.

Iv-C2 Weakly supervised object segmentation

Conventional object segmentation approaches use scribbles or rectangular boxes to indicate the object of interest, while in our approach, only the semantic label of the object of interest is input to the system, similar to the semantic object selection [77]. We threshold our TD saliency map to identify definite foreground and background regions in an image, followed by Grab-cut [73] to accurately segment out the object of interest. Being a weakly supervised approach, framework is comparable to co-segmentation approaches that segment out a common object from a given set of images. We learn a model for the common object, which helps to achieve faster inference for a newly added test image, whereas co-segmentation approaches need to re-segment every image in the set upon encountering a new image.

Object segmentation accuracy is evaluated on 100 test images from each category of Object Discovery dataset [62]. 300 images from each category are used to train our saliency model, along with 300 negative images from Graz-02 dataset. Qualitative results are shown in Fig. 7. Multiple instances of car are accurately segmented out and the proposed approach could accurately segment out the horse. Quantitative comparisons with state-of-the-art co-segmentation approaches are shown in Table VI. The Jaccard similarity, i.e, intersection over union () with the ground-truth is evaluated as in  [62]. In all the three categories, we achieve state-of-the-art performance compared to related co-segmentation [62, 46] and co-saliency  [47] approaches. The semantic object selection [77] uses additional supervision by collecting positive training images with white background using an internet search. Inspite of this modification, they could only achieve an average accuracy of 63.73%, which is lower than our mean accuracy of 68.0% across 3 categories.

Method Airplane Car Horse Mean
Joulin et al. [78] 15.4 37.2 30.2 27.6
Joulin et al. [79] 11.7 35.2 29.5 25.5
Kim et al.[80] 7.9 0.04 6.43 4.79
Object Discovey [62] 55.8 64.4 51.6 57.3
Koteshwar et al. [81] 56 69 55 60
Zhang et al.[47] 53.5 58.8 52.2 54.8
Quan et al.[46] 56.3 66.8 58.1 60.4
WS-SC[15] 57.3 67.4 50.51 58.4
Object selection [77] 64.3 71.8 55.1 63.7
Proposed 65.0 77.3 61.6 68.0
TABLE VI: Comparison of proposed weakly supervised approach with object segmentation approaches on Object Discovery dataset, evaluated using Jaccard similarity.
Method SV Exact 18 Pixel
RCNN [1] FS 67.7 74.8
Fast RCNN [82] FS - 81.9
Exemplar [14] FS 73.4 -
Oquab et al. [37], WS - 74.5
ProNet [44] WS 69.8 74.8
ProNet + classifier [44] WS 73.1 77.7
Bency et al.,[83] WS - 79.7
Proposed WS 82.2 84.7
TABLE VII: Average precision of object localization on PASCAL VOC-2012 detection validation set.

Iv-C3 Weakly supervised object localization

Object localization deals with locating object within a positive image. Here, only the location of the object needs to be identified, not its extent. The peaks of our saliency map, indicates the location of object ,

Presence of multiple objects in an image makes object localization on PASCAL VOC-2012 detection set a challenging task, especially in a weakly supervised setting. Since an accurate estimate of object boundaries are not required, we replaced the multi-scale superpixel averaging with an averaging filter on a rectangular window of size pixels for faster inference.

The location that falls exactly within any ground truth bounding box associated for a given category is assumed correct and the average precision is calculated as in [14]. In [37], average precision is evaluated by giving an error tolerance of 18 pixels to the predicted location. We evaluated our model in both these settings denoted Exact and 18 Pix and corresponding results are compared with state-of-the art approaches as shown in Table VII. In both the evaluation settings, we achieve a performance which is comparable to fully supervised TD saliency approaches and dedicated object detectors such as fast RCNN [82]. It is to be noted that object localization is inherently a simpler task compared to object detection. In object detection applications, fully supervised object detectors [84, 85, 82] outperform the proposed method, as expected. Fig. 7 shows some qualitative results obtained using the proposed method in localizing multiple objects. Partially occluded objects such as motorbike and car are localized accurately despite the presence of other distracting objects.

Iv-C4 Weakly supervised object detection

In object detection, multiple instances of the same object category need to be identified separately. This is more challenging than localization and especially so in a weakly supervised setting. Conventional object detectors such as R-CNN [1, 41] need to classify thousands of category-independent object proposals generated using selective search [42], [43]. This incurs a huge computational cost. The proposed TD saliency framework simplifies object detection by generating less than 5 proposals for an object category per image. First, the category-specific saliency is binarized by applying a threshold at . The smallest rectangular box enclosing each disconnected region is the detection box for object . With this simple strategy, we achieve a performance which is comparable to dedicated weakly supervised object detectors [44].

Method Oquab et al. [37] ProNet [44] ProNet+Classifier [44] Proposed
(Mean of 20 Classes)
11.74 13 15.5 20.4
TABLE VIII: Comparison with weakly supervised object detection approaches on PASCAL VOC-2012 validation dataset, measured by average precision.

The object detection boxes produced by a simple binarization of our saliency maps is shown to be comparable with dedicated weakly supervised object detectors in Table VIII. We outperform [44] which uses an additional box classifier to classify their object proposal boxes. We consider all category-specific object boxes as positive detections. PASCAL VOC 2012 evaluation server is used to estimate object detection accuracy, where a detection having an with the ground truth rectangular bounding box is considered as true positive. [86] requires thousands of forward passes through the network to identify the class-specific object proposals, which is time consuming. Moreover, fine-tuning of convolution layers is required for detection task, whereas we use the CNN trained for image classification, without fine-tuning. Inspite of all these computational requirements, their framework based on AlexNet achieved a mAP of 22.4% and their framework based on [53] achieved a mAP of 29.1 on PASCAL VOC 2012 object detection dataset. In Fig. 7, multiple overlapping objects are accurately detected by the proposed strategy. Multiple instances of person, motorbike and car are also detected. The person and bicycle are accurately detected despite the presence of other categories in the image. Similarly, an accurate bounding box around the cat is marked in an image that also contains a TV monitor.

Limitations. Similar to other weakly supervised approaches [44, 41], the proposed approach has limited ability to discriminate among multiple instances of an object which are spatially adjacent. This causes low performance for object detection, compared to state-of-the-art fully supervised object detectors [85, 82, 84]. Examples are provided in the supplementary material.

Supplementary material. Additional experiments to evaluate (i) selection of bottom-up saliency map [87, 88], (ii) performance comparison on different CNN architectures, (iii) computation time, and (iv) additional qualitative results and comparisons are included in the supplementary material.

V Conclusion

In this paper, a CNN feature-based weakly supervised salient object detection approach is proposed. A novel strategy to select a BU saliency map that suits a top-down task is proposed. Contribution of relu5 features at different spatial locations are estimated to compute a novel B-cSPP saliency. The top-down B-cSPP saliency is integrated with the BU saliency map and produces a combined saliency which is further integrated with feature saliency. The proposed weakly supervised top-down saliency model achieves state-of-the-art performance in top-down salient object detection across multiple datasets, by outperforming even fully supervised CNN-based approaches. Moreover, the top-down saliency maps of different object categories are combined to produce a category-independent saliency map that can estimate salient objects under free-viewing condition. Finally, through quantitative comparisons, we demonstrated the usefulness of proposed saliency map for four different applications. We plan to extend our framework to videos for weakly supervised salient object detection.


  • [1] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in CVPR, 2014, pp. 580–587.
  • [2] Y. Jia and M. Han, “Category-independent object-level saliency detection,” in ICCV, 2013.
  • [3] D. Vaquero, M. Turk, K. Pulli, M. Tico, and N. Gelfand, “A survey of image retargeting techniques,” in SPIE Optical Engineering + Applications, 2010, pp. 779–814.
  • [4] W.-C. Tu, S. He, Q. Yang, and S.-Y. Chien, “Real-time salient object detection with a minimum spanning tree,” in CVPR, June 2016.
  • [5] J. Yang and M.-H. Yang, “Top-down visual saliency via joint crf and dictionary learning,” PAMI, 2016.
  • [6] T. Liu, Z. Yuan, J. Sun, J. Wang, N. Zheng, X. Tang, and H.-Y. Shum, “Learning to detect a salient object,” PAMI, vol. 33, no. 2, pp. 353–367, 2011.
  • [7] J. Zhang, S. Sclaroff, Z. Lin, X. Shen, B. Price, and R. Mĕch, “Minimum barrier salient object detection at 80 fps,” in ICCV, 2015.
  • [8] M. M. Cheng, N. J. Mitra, X. Huang, P. H. S. Torr, and S. M. Hu, “Global contrast based salient region detection,” PAMI, vol. 37, no. 3, pp. 569–582, March 2015.
  • [9] P. O. Pinheiro and R. Collobert, “Weakly supervised semantic segmentation with convolutional networks,” in CVPR, vol. 2, no. 5, 2015, p. 6.
  • [10] D. Gao, S. Han, and N. Vasconcelos, “Discriminant saliency, the detection of suspicious coincidences, and applications to visual recognition,” PAMI, vol. 31, no. 6, pp. 989–1005, 2009.
  • [11] A. Kocak, K. Cizmeciler, A. Erdem, and E. Erdem, “Top down saliency estimation via superpixel-based discriminative dictionaries,” in BMVC, 2014.
  • [12] H. Cholakkal, D. Rajan, and J. Johnson, “Top-down saliency with locality-constrained contextual sparse coding,” in BMVC, 2015.
  • [13] F. Moosmann, E. Nowak, and F. Jurie, “Randomized clustering forests for image classification,” PAMI, vol. 30, no. 9, pp. 1632–1646, Sept 2008.
  • [14] S. He, R. W. Lau, and Q. Yang, “Exemplar-driven top-down saliency detection via deep association,” in CVPR, 2016.
  • [15] H. Cholakkal, J. Johnson, and D. Rajan, “Backtracking scspm image classifier for weakly supervised top-down saliency,” in CVPR, 2016.
  • [16] J. Zhang, Z. Lin, S. X. Brandt, Jonathan, and S. Sclaroff, “Top-down neural attention by excitation backprop,” in ECCV, 2016.
  • [17] J. T. Springenberg, A. Dosovitskiy, T. Brox, and M. Riedmiller, “Striving for simplicity: The all convolutional net,” arXiv preprint arXiv:1412.6806, 2014.
  • [18] K. Simonyan, A. Vedaldi, and A. Zisserman, “Deep inside convolutional networks: Visualising image classification models and saliency maps,” in ICLR Workshop, 2014.
  • [19] C. Kanan, M. H. Tong, L. Zhang, and G. W. Cottrell, “Sun: Top-down saliency using natural statistics,” Vis. cogn., vol. 17, no. 6-7, pp. 979–1003, 2009.
  • [20] B. Alexe, T. Deselaers, and V. Ferrari, “What is an object?” in CVPR, 2010.
  • [21] N. Khan and M. F. Tappen, “Discriminative dictionary learning with spatial priors.” in ICIP, 2013.
  • [22] H. Cholakkal, J. Johnson, and D. Rajan, “A classifier-guided approach for top-down salient object detection,” Signal Process. Image Commun., vol. 45, pp. 24–40, 2016.
  • [23] J. Zhu, Y. Qiu, R. Zhang, J. Huang, and W. Zhang, “Top-down saliency detection via contextual pooling,” J. Signal Process. Syst., vol. 74, no. 1, pp. 33–46, 2014.
  • [24] J. Wang, J. Yang, K. Yu, F. Lv, T. Huang, and Y. Gong, “Locality-constrained linear coding for image classification,” in CVPR, 2010.
  • [25] B. Yao, A. Khosla, and L. Fei-Fei, “Combining randomization and discrimination for fine-grained image categorization,” in CVPR, 2011.
  • [26] G. Sharma, F. Jurie, and C. Schmid, “Discriminative spatial saliency for image classification,” in CVPR, 2012, pp. 3506–3513.
  • [27] J. Yang and M.-H. Yang, “Top-down visual saliency via joint crf and dictionary learning,” in CVPR, 2012.
  • [28] N. Liu and J. Han, “Dhsnet: Deep hierarchical saliency network for salient object detection,” in CVPR, June 2016.
  • [29] R. Zhao, W. Ouyang, H. Li, and X. Wang, “Saliency detection by multi-context deep learning,” in CVPR, 2015, pp. 1265–1274.
  • [30] L. Wang, H. Lu, X. Ruan, and M.-H. Yang, “Deep networks for saliency detection via local estimation and global search,” in CVPR, June 2015.
  • [31]

    G. Li and Y. Yu, “Visual saliency based on multiscale deep features,” in

    CVPR, June 2015.
  • [32] Y. Li, X. Hou, C. Koch, J. M. Rehg, and A. L. Yuille, “The secrets of salient object segmentation,” in CVPR, 2014.
  • [33] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei, “ImageNet Large Scale Visual Recognition Challenge,” Int. J. Comput. Vision, 2015.
  • [34] A. Mahendran and A. Vedaldi, “Salient deconvolutional networks,” in ECCV, 2016.
  • [35] W. Shimoda and K. Yanai, “Distinct class-specific saliency maps for weakly supervised semantic segmentation,” in ECCV, 2016.
  • [36] A. Bergamo, L. Bazzani, D. Anguelov, and L. Torresani, “Self-taught object localization with deep networks,” arXiv preprint arXiv:1409.3964, 2014.
  • [37] M. Oquab, L. Bottou, I. Laptev, and J. Sivic, “Is object localization for free?-weakly-supervised learning with convolutional neural networks,” in CVPR, 2015, pp. 685–694.
  • [38] R. G. Cinbis, J. Verbeek, and C. Schmid, “Weakly supervised object localization with multi-fold multiple instance learning,” PAMI, 2016.
  • [39] M. Oquab, L. Bottou, I. Laptev, and J. Sivic, “Learning and transferring mid-level image representations using convolutional neural networks,” in CVPR, 2014, pp. 1717–1724.
  • [40] Z. Bolei, K. Aditya, L. Agata, O. Aude, and T. Antonio, “Learning deep features for discriminative localization,” in CVPR, 2016.
  • [41] H. Bilen and A. Vedaldi, “Weakly supervised deep detection networks,” in CVPR, 2016.
  • [42] J. R. Uijlings, K. E. van de Sande, T. Gevers, and A. W. Smeulders, “Selective search for object recognition,” Int. J. Comput. Vision, vol. 104, no. 2, pp. 154–171, 2013.
  • [43] C. L. Zitnick and P. Dollár, “Edge boxes: Locating object proposals from edges,” in ECCV, 2014, pp. 391–405.
  • [44] C. Sun, M. Paluri, R. Collobert, R. Nevatia, and L. Bourdev, “Pronet: Learning to propose object-specific boxes for cascaded neural networks,” in CVPR, 2016.
  • [45] D. Pathak, P. Krahenbuhl, and T. Darrell, “Constrained convolutional neural networks for weakly supervised segmentation,” in ICCV, 2015, pp. 1796–1804.
  • [46] R. Quan, J. Han, D. Zhang, and F. Nie, “Object co-segmentation via graph optimized-flexible manifold ranking,” in CVPR, June 2016.
  • [47] D. Zhang, J. Han, C. Li, J. Wang, and X. Li, “Detection of co-salient objects by looking deep and wide,” Int. J. Comput. Vision, pp. 1–18, 2016.
  • [48] M. D. Zeiler and R. Fergus, “Visualizing and understanding convolutional networks,” in ECCV, 2014, pp. 818–833.
  • [49] A. Mahendran and A. Vedaldi, “Understanding deep image representations by inverting them,” in CVPR, 2015, pp. 5188–5196.
  • [50] A. Dosovitskiy and T. Brox, “Inverting convolutional networks with convolutional networks,” CoRR abs/1506.02753, 2015.
  • [51] B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, and A. Torralba, “Object detectors emerge in deep scene cnns,” arXiv preprint arXiv:1412.6856, 2014.
  • [52] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” CoRR, vol. abs/1409.1556, 2014.
  • [53] K. Chatfield, K. Simonyan, A. Vedaldi, and A. Zisserman, “Return of the devil in the details: Delving deep into convolutional nets,” in BMVC, 2014.
  • [54] K. He, X. Zhang, S. Ren, and J. Sun, “Spatial pyramid pooling in deep convolutional networks for visual recognition,” PAMI, vol. 37, no. 9, pp. 1904–1916, 2015.
  • [55] S. Lazebnik, C. Schmid, and J. Ponce, “Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories,” in CVPR, 2006.
  • [56] A. Vedaldi and B. Fulkerson, “VLFeat: An open and portable library of computer vision algorithms,”, 2008.
  • [57] S. Shalev-Shwartz and T. Zhang, “Stochastic dual coordinate ascent methods for regularized loss minimization,” J. Mach. Learn. Res., vol. 14, no. Feb, pp. 567–599, 2013.
  • [58] R. Achanta, A. Shaji, K. Smith, A. Lucchi, P. Fua, and S. Susstrunk, “Slic superpixels compared to state-of-the-art superpixel methods,” PAMI, vol. 34, no. 11, pp. 2274–2281, 2012.
  • [59] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman, “The pascal visual object classes challenge 2012 (voc2012) results,”
  • [60] A. Opelt, A. Pinz, M. Fussenegger, and P. Auer, “Generic object recognition with boosting,” PAMI, vol. 28, no. 3, pp. 416–431, 2006.
  • [61] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman, “The PASCAL Visual Object Classes Challenge 2007 (VOC2007) Results,”
  • [62] M. Rubinstein, A. Joulin, J. Kopf, and C. Liu, “Unsupervised joint object discovery and segmentation in internet images,” in CVPR, 2013.
  • [63] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman, “The PASCAL Visual Object Classes Challenge 2010 (VOC2010) Results,”
  • [64] D. Aldavert, A. Ramisa, R. L. de Mantaras, and R. Toledo, “Fast and robust object segmentation with the integral linear classifier,” in CVPR, 2010.
  • [65] F. Brian, V. Andrea, and S. Stefano, “Class segmentation and object localization with superpixel neighborhoods,” in ICCV, 2009.
  • [66] M. Marcin and S. Cordelia, “Accurate object recognition with shape masks,” Int. J. Comput. Vision, vol. 97, no. 2, pp. 191–209, 2012.
  • [67] B. Ali, C. Ming-Ming, J. Huaizu, and L. Jia, “Salient object detection: A benchmark,” TIP, vol. 24, no. 12, pp. 5706–5722, 2015.
  • [68] C. Yang, L. Zhang, H. Lu, X. Ruan, and M.-H. Yang, “Saliency detection via graph-based manifold ranking,” in CVPR, 2013, pp. 3166–3173.
  • [69] X. Qi, Z. Liu, J. Shi, H. Zhao, and J. Jia, “Augmented feedback in semantic segmentation under image level supervision,” in ECCV, 2016, pp. 90–105.
  • [70] A. Kolesnikov and C. H. Lampert, “Seed, expand and constrain: Three principles for weakly-supervised image segmentation,” in ECCV, 2016, pp. 695–711.
  • [71] F. Saleh, M. S. A. Akbarian, M. Salzmann, L. Petersson, S. Gould, and J. M. Alvarez, “Built-in foreground/background prior for weakly-supervised semantic segmentation,” in ECCV.   Springer, 2016, pp. 413–432.
  • [72] Y. Wei, X. Liang, Y. Chen, X. Shen, M.-M. Cheng, J. Feng, Y. Zhao, and S. Yan, “Stc: A simple to complex framework for weakly-supervised semantic segmentation,” PAMI, 2016.
  • [73] C. Rother, V. Kolmogorov, and A. Blake, “Grabcut: Interactive foreground extraction using iterated graph cuts,” ACM Trans. Graphics, vol. 23, no. 3, pp. 309–314, 2004.
  • [74] D. Pathak, E. Shelhamer, J. Long, and T. Darrell, “Fully convolutional multi-class multiple instance learning,” in Proc. ICLR, 2015.
  • [75] P. O. Pinheiro and R. Collobert, “From image-level to pixel-level labeling with convolutional networks,” June 2015.
  • [76]

    G. Papandreou, L.-C. Chen, K. Murphy, and A. L. Yuille, “Weakly-and semi-supervised learning of a dcnn for semantic image segmentation,” in

    ICCV, 2015.
  • [77] E. Ahmed, S. Cohen, and B. Price, “Semantic object selection,” in CVPR, 2014.
  • [78] A. Joulin, F. Bach, and J. Ponce, “Discriminative clustering for image co-segmentation,” in CVPR, 2010.
  • [79] J. Armand, B. Francis, and P. Jean, “Multi-class cosegmentation,” in CVPR, 2012.
  • [80] G. Kim, E. P. Xing, L. Fei-Fei, and T. Kanade, “Distributed cosegmentation via submodular optimization on anisotropic diffusion,” in ICCV, 2011.
  • [81] K. R. Jerripothula, J. Cai, and J. Yuan, “Object detection with discriminatively trained part-based models,” TMM, 2016.
  • [82] R. Girshick, “Fast r-cnn,” in ICCV, 2015, pp. 1440–1448.
  • [83] A. J. Bency, H. Kwon, H. Lee, S. Karthikeyan, and B. S. Manjunath, “Weakly supervised localization using deep feature maps,” in ECCV, 2016.
  • [84] S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: Towards real-time object detection with region proposal networks,” in NIPS., 2015, pp. 91–99.
  • [85] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in CVPR, 2016.
  • [86] D. Li, J. B. Huang, Y. Li, S. Wang, and M. H. Yang, “Weakly supervised object localization with progressive domain adaptation,” in CVPR, June 2016, pp. 3512–3520.
  • [87] P. Jiang, N. Vasconcelos, and J. Peng, “Generic promotion of diffusion-based salient object detection,” in ICCV, 2015, pp. 217–225.
  • [88] Ç. Aytekin, E. C. Ozan, S. Kiranyaz, and M. Gabbouj, “Visual saliency by extended quantum cuts,” in ICIP, 2015, pp. 1692–1696.