Md Amirul Islamcs.ryerson.ca/ amirul1,4
\addauthorMatthew Kowalmkowal2.github.io2,4
\addauthorSen Jiagithub.com/SenJia5
\addauthorKonstantinos G. Derpaniswww.eecs.yorku.ca/ kosta2,4,6
\addauthorNeil D. B. Brucesocs.uoguelph.ca/ brucen3,4
\addinstitution
Ryerson University, Canada
\addinstitution
York University, Canada
\addinstitution
University of Guelph, Canada
\addinstitution
Vector Institute for AI, Canada
\addinstitution
Toronto AI Lab, LG
\addinstitution
Samsung AI Centre Toronto
Pseudo-label generator
Simpler Does It: Generating Semantic Labels with Objectness Guidance
Abstract
Existing weakly or semi-supervised semantic segmentation methods utilize image or box-level supervision to generate pseudo-labels for weakly labeled images. However, due to the lack of strong supervision, the generated pseudo-labels are often noisy near the object boundaries, which severely impacts the network’s ability to learn strong representations. To address this problem, we present a novel framework that generates pseudo-labels for training images, which are then used to train a segmentation model. To generate pseudo-labels, we combine information from: (i) a class agnostic ‘objectness’ network that learns to recognize object-like regions, and (ii) either image-level or bounding box annotations. We show the efficacy of our approach by demonstrating how the objectness network can naturally be leveraged to generate object-like regions for unseen categories. We then propose an end-to-end multi-task learning strategy, that jointly learns to segment semantics and objectness using the generated pseudo-labels. Extensive experiments demonstrate the high quality of our generated pseudo-labels and effectiveness of the proposed framework in a variety of domains. Our approach achieves better or competitive performance compared to existing weakly-supervised and semi-supervised methods.
1 Introduction
State-of-the-art methods for semantic segmentation [Long et al.(2015)Long, Shelhamer, and Darrell, Chen et al.(2015)Chen, Papandreou, Kokkinos, Murphy, and Yuille, Noh et al.(2015)Noh, Hong, and Han, Badrinarayanan et al.(2017)Badrinarayanan, Kendall, and Cipolla, Islam et al.(2017b)Islam, Rochan, Bruce, and Wang, Ghiasi and Fowlkes(2016), Islam et al.(2017a)Islam, Naha, Rochan, Bruce, and Wang, Yu and Koltun(2016), Ghiasi and Fowlkes(2016), Zhao et al.(2017)Zhao, Shi, Qi, Wang, and Jia, Lin et al.(2017)Lin, Milan, Shen, and Reid, Chen et al.(2017)Chen, Papandreou, Schroff, and Adam, Chen et al.(2018)Chen, Papandreou, Kokkinos, Murphy, and Yuille, Karim et al.(2020)Karim, Islam, and Bruce, Karim et al.(2019)Karim, Islam, and Bruce, Islam et al.(2020)Islam, Kowal, Derpanis, and Bruce, Takikawa et al.(2019)Takikawa, Acuna, Jampani, and Fidler, Islam et al.(2021a)Islam, Kowal, Derpanis, and Bruce] are founded on fully convolutional networks (FCN) [Long et al.(2015)Long, Shelhamer, and Darrell] to segment semantic objects in an end-to-end manner. A caveat of such training is that it requires supervision with an extensive amount of pixel-level annotations. Since the expense for generating semantic segmentation annotations is large, a natural solution is to address the problem of semantic segmentation with one of two common supervision settings, weakly or semi-supervised.
In the weakly supervised semantic segmentation (WSSS) setting, labels used during training contain only partial information. Recently proposed WSSS methods utilize image-level labels [Fan et al.(2020b)Fan, Zhang, and Tan, Chen et al.(2020)Chen, Wu, Fu, Han, and Zhang, Chang et al.(2020)Chang, Wang, Hung, Piramuthu, Tsai, and Yang, Fan et al.(2020a)Fan, Zhang, Song, and Tan, Ahn and Kwak(2018), Huang et al.(2018)Huang, Wang, Wang, Liu, and Wang, Hou et al.(2018)Hou, Jiang, Wei, and Cheng, Lee et al.(2019a)Lee, Kim, Lee, Lee, and Yoon, Zeng et al.(2019)Zeng, Zhuge, Lu, and Zhang, Lee et al.(2019b)Lee, Kim, Lee, Lee, and Yoon, Ahn et al.(2019)Ahn, Cho, and Kwak, Jiang et al.(2019)Jiang, Hou, Cao, Cheng, Wei, and Xiong, Wei et al.(2018)Wei, Xiao, Shi, Jie, Feng, and Huang], scribbles [Lin et al.(2016)Lin, Dai, Jia, He, and Sun], or bounding box [Dai et al.(2015)Dai, He, and Sun, Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele, Song et al.(2019)Song, Huang, Ouyang, and Wang] supervision to learn semantic masks.

Most of these methods rely on incorporating additional guidance to obtain the location and shape information. A common way to obtain location cues from class labels is by using Class Activation Maps (CAMs) [Zhou et al.(2016)Zhou, Khosla, Lapedriza, Oliva, and Torralba] as it roughly localizes semantic regions of each class. However, utilizing CAMs directly as supervision can be problematic as they roughly localize objects and cannot capture detailed object boundaries between different semantic regions. Recent works have addressed this issue in a variety of ways [Pinheiro and Collobert(2015), Kwak et al.(2017)Kwak, Hong, and Han, Ahn and Kwak(2018), Ahn et al.(2019)Ahn, Cho, and Kwak], one of the most effective being the use of object guidance via the use of class agnostic saliency [Huang et al.(2018)Huang, Wang, Wang, Liu, and Wang, Lee et al.(2019a)Lee, Kim, Lee, Lee, and Yoon, Zeng et al.(2019)Zeng, Zhuge, Lu, and Zhang, Jiang et al.(2019)Jiang, Hou, Cao, Cheng, Wei, and Xiong]. Similarly, bounding box based methods [Kulharia et al.(2020)Kulharia, Chandra, Agrawal, Torr, and Tyagi, Song et al.(2019)Song, Huang, Ouyang, and Wang, Li et al.(2018)Li, Arnab, and Torr, Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele, Huang et al.(2018)Huang, Wang, Wang, Liu, and Wang] typically rely on generating rough pseudo-labels by applying the unsupervised CRF [Krähenbühl and Koltun(2011)], MCG [Pont-Tuset et al.(2016)Pont-Tuset, Arbelaez, Barron, Marques, and Malik], or GrabCut [Rother et al.(2004)Rother, Kolmogorov, and Blake] methods to remove irrelevant regions from the semantic region proposal in an iterative way to obtain stronger pseudo-labels at each iteration. However, the quality gap between the pseudo-labels and groundtruth is typically large for the CAM-based and bounding box-based approaches. Furthermore, iterative procedures and complex pipelines can make the data generation process for these methods computationally expensive and time consuming.
In the semi-supervised semantic segmentation (SSSS) setting, the groundtruth annotations are used but only for a fraction of the total number of training examples, e.g., 10% of the labels [Souly et al.(2017)Souly, Spampinato, and Shah]. Similar to the techniques used in WSSS methods, pseudo-labels are then generated for the unlabelled data (e.g., by using additional image-level annotations [Xiao et al.(2018a)Xiao, Wei, Liu, Zhang, and Feng, Wei et al.(2018)Wei, Xiao, Shi, Jie, Feng, and Huang, Lee et al.(2019a)Lee, Kim, Lee, Lee, and Yoon, Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele, Huang et al.(2018)Huang, Wang, Wang, Liu, and Wang]). Recent work [Hu et al.(2018)Hu, Dollár, He, Darrell, and Girshick] introduced a partially supervised training paradigm which learns to segment everything using a portion of box and mask annotations. However, these methods still require labour-intensive pixel-level semantic annotations and the performance heavily depends on the quantity of the labeled data and the quality of the generated pseudo-labels.
In the light of the highlighted issues that arise in WSSS and SSSS methods, we propose a novel simple yet effective pipeline which transfers ‘objectness’ knowledge to weakly labeled images for learning semantic segmentation. The intuition behind using the objectness guidance instead of widely used saliency-based approaches [Oh et al.(2017)Oh, Benenson, Khoreva, Akata, Fritz, and Schiele, Wei et al.(2018)Wei, Xiao, Shi, Jie, Feng, and Huang, Zeng et al.(2019)Zeng, Zhuge, Lu, and Zhang, Yao and Gong(2020)] is that groundtruth saliency masks inherently ignore objects near the border of the image due to the well-known centre bias [Bruce et al.(2015)Bruce, Wloka, Frosst, Rahman, and Tsotsos, Açık et al.(2014)Açık, Bartel, and Koenig]. Recent works [Xiao et al.(2018a)Xiao, Wei, Liu, Zhang, and Feng, Wang et al.(2016)Wang, Liu, Li, Yan, and Lu] also utilize the objectness prior to refine the semantic proposals. There are two key differences between our work and [Xiao et al.(2018a)Xiao, Wei, Liu, Zhang, and Feng]. First is the use of a source dataset. [Xiao et al.(2018a)Xiao, Wei, Liu, Zhang, and Feng] obtains the objectness prior strictly from the target data distribution, which is arguably an easier problem to solve. However in our work, we strictly prohibit the use of per-pixel labels from the target dataset and only use a source dataset (i.e., COCOStuff) for the objectness prior. We argue that using COCOStuff as the source data (instead of VOC like in [Xiao et al.(2018a)Xiao, Wei, Liu, Zhang, and Feng]) will allow our objectness network to generate better pseudo-labels for a more diverse set of categories and can be generalized well to different target datasets. Second, during the segmentation network training, [Xiao et al.(2018a)Xiao, Wei, Liu, Zhang, and Feng] uses the semantic segmentation labels for the strong categories (i.e., the classes used to train their objectness network), while in our settings we only use pseudo-labels when training the semantic segmentation network.
The key component of our pipeline is the pseudo-label generation approach (see Fig. 1), where we first train an objectness network on a source dataset which generates a class agnostic objectness prior. We then combine this prior with weak semantic proposals (e.g., image or box-level) to generate semantic segmentation labels for a target dataset. We further show that the objectness prior is robust enough to generalize the objectness knowledge onto categories that have never been seen by the objectness network; when the source dataset has no class overlap with the target dataset (i.e., the non-overlapping case). We view the non-overlapping setting as comparable with weak-supervision, as the objectness model has no direct understanding of the shape of the target domain classes (unlike previous methods [Oh et al.(2017)Oh, Benenson, Khoreva, Akata, Fritz, and Schiele, Wei et al.(2018)Wei, Xiao, Shi, Jie, Feng, and Huang, Zeng et al.(2019)Zeng, Zhuge, Lu, and Zhang, Yao and Gong(2020)] which use overlapping groundtruth saliency annotations). In contrast, the overlapping setting (i.e., the class agnostic source dataset contains objects found in the target dataset) is comparable (but has less supervision) to semi-supervision as class-agnostic (i.e., binary) segmentation annotations are used. Finally, for segmentation learning, we adopt a multi-task joint-learning [Cheng et al.(2017)Cheng, Tsai, Wang, and Yang, Islam et al.(2018a)Islam, Kalash, and Bruce, Islam et al.(2018b)Islam, Kalash, and Bruce, Zeng et al.(2019)Zeng, Zhuge, Lu, and Zhang, He et al.(2021)He, Lu, Wang, Song, and Zhou] based Semantic Objectness Network (denoted as SONet), with the addition of an ‘objectness branch’, that explicitly models the relationship between semantics and objectness. We summarize our main contributions as follows:
-
1.
We introduce a simple yet effective pseudo-label generation technique that combines a class agnostic ‘objectness’ prior with semantic region proposals. The flexibility of our technique is demonstrated by its ability to incorporate either image or box-level labels into the pseudo-label generation pipeline.
-
2.
We propose a joint learning based Semantic Objectness Network, SONet, that improves the semantic segmentation quality through objectness guidance.
-
3.
We present an extensive set of experimental results which demonstrates the effectiveness of our proposed method in both the simplicity of the pseudo-label generation process as well as the quality of the pseudo-labels. Our proposed approach achieves competitive performance compared to existing WSSS methods and outperforms SSSS methods without ever using groundtruth semantic segmentation supervision.
2 Proposed Framework
Our pipeline consists of two key components. First, we generate pseudo-labels for training images by combining our generated objectness prior with weak semantic proposals, which are produced from either image labels or box annotations (Sec. 2.1). Second, we introduce our multi-task model, SONet, that jointly learns to segment both semantic categories and a binary ‘objectness’ mask, which enforces richer boundary detail and semantic information (Sec. 2.2).
2.1 Semantic Pseudo-Label Generation
Our pseudo-label generation process consists of two separate components. We first describe the procedure behind training the ‘objectness’ network which is designed to obtain detailed boundary information for any object-like region. Next, we describe two different techniques for generating semantic pseudo-labels by combining the output of the objectness network with semantic region proposals, which are obtained from either image-level class labels or bounding box annotations.
Training an Objectness Network. Pixel objectness [Xiong et al.(2018)Xiong, Jain, and Grauman] quantifies how likely a pixel belongs to an object of any class (i.e., other than “stuff” classes like background, grass, sky, sidewalks, etc.), and should be high even for objects unseen during training. We use DeepLabv3 network [Chen et al.(2017)Chen, Papandreou, Schroff, and Adam], , on a source dataset, , to learn an objectness prior from the ‘things’ label. We use a weak form of the COCOStuff dataset, denoted as COCO-Binary and consider it as the source dataset, . More specifically, we generate COCO-Binary by removing all semantic labels from the COCOStuff dataset so what remains is binary maps where all the things categories are assigned to the label one, and the stuff categories to zero. We then train the objectness network, , on the source dataset, , under two different settings which outputs a pixel-wise ‘objectness score’ (similar to the saliency detection models). In the first setting, we include all the images from the source data, , regardless of whether the objects found in images overlap with target data, . In the second setting, we create a subset of by excluding those images containing any categories which overlap with categories. We can formalize the overlapping and non-overlapping settings as follows:
|
(1) |
where denotes the set of object classes contained in COCO-Binary used to train the objectness model, . represents the non-overlapping subset where there is no semantic category overlap between and . Note that the semantic annotations are solely used to generate the subset of non-overlapping data, , and is not required for training the objectness model, . We believe the non-overlapping setting is more challenging than saliency-based WSSS [Oh et al.(2017)Oh, Benenson, Khoreva, Akata, Fritz, and
Schiele, Wei et al.(2018)Wei, Xiao, Shi, Jie, Feng, and
Huang, Zeng et al.(2019)Zeng, Zhuge, Lu, and Zhang, Yao and Gong(2020)], because those methods contain semantic overlap within the source and target data. In both settings, we train the objectness classifier using the class-agnostic segmentation groundtruth and use the binary cross entropy loss function. The main goal of the objectness classifier is to learn a strong objectness representation [Islam et al.(2021b)Islam, Kowal, Esser, Jia, Ommer,
Derpanis, and Bruce] that contributes towards creating pseudo-labels for semantic supervision.
Class-Driven Pseudo-labels. CAM [Zhou et al.(2016)Zhou, Khosla, Lapedriza, Oliva, and Torralba] is widely used as a weak source of supervision as it roughly localizes semantic object areas. Following previous works [Ahn and Kwak(2018), Ahn et al.(2019)Ahn, Cho, and Kwak], we first generate CAMs for training images by adopting the method of [Zhou et al.(2016)Zhou, Khosla, Lapedriza, Oliva, and Torralba] using a multi-label image classification network. For a fair comparison, we use a ResNet-50 [He et al.(2016)He, Zhang, Ren, and Sun] model as the classification network, as used in other CAM-based methods [Ahn and Kwak(2018), Ahn et al.(2019)Ahn, Cho, and Kwak, Huang et al.(2018)Huang, Wang, Wang, Liu, and Wang, Yao and Gong(2020)]. We directly utilize the raw CAMs to generate pseudo-labels by thresholding their confidence scores for each class label at every pixel predicted to be an object by the class agnostic objectness network (see Fig. 1(B)). We can formalize this procedure as follows:
|
(2) |
where denotes the pseudo-label value at pixel , is the set of class indices, is the objectness score, is the non-thresholded CAM proposals, and is a threshold (we use in all experiments).
Box-Driven Pseudo-labels. The simplest box-driven pseudo-labels can be obtained by filling the bounding box annotations with the corresponding class label. Some methods [Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele, Song et al.(2019)Song, Huang, Ouyang, and Wang] use semi-automatic segmentation techniques (e.g., CRF [Papandreou et al.(2015)Papandreou, Chen, Murphy, and Yuille], GrabCut [Rother et al.(2004)Rother, Kolmogorov, and Blake]) to further refine the box annotations, as rectangular regions contain a significant number of incorrectly labeled background pixels. However, these techniques are time consuming and the quality of the pseudo-label is lacking. To address this challenge, we propose an approach to generate pseudo-labels using the class agnostic objectness masks, , and the box annotations, .
Following common practice [Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele, Ibrahim et al.(2018)Ibrahim, Vahdat, and Macready, Song et al.(2019)Song, Huang, Ouyang, and Wang], if any two bounding boxes overlap, we assume the box with smaller area appears in front. Additionally, if the overlap between any box and the largest box in the image is greater than some threshold, we only keep the inner 60% of the box and fill the rest of the box as 255 (which is ignored during training). The intuition behind the ignoring strategy is simply trading-off lower recall (ignore more pixels where high-degree of overlap occurs) for higher precision (more pixels are correctly labelled). We then mask the resulting box proposal, , with the objectness map, , to filter out the irrelevant regions from and , and only keep the regions overlapping the object of interest. We set any pixel to the background class if it does not overlap any boxes. Formally, for each bounding box, , , in an image:
|
(3) |
where denotes the largest box, is the number of boxes in each image, is the outer 40% of the bounding box’s area, is the inner 60% of the bounding box, ‘’ calculates the area of intersection between two bounding boxes, and is a threshold (we set ).
2.2 Semantic Objectness Network: SONet
The Semantic Objectness Network (SONet) consists of a segmentation network and an objectness module. The objectness module receives the output of the segmentation network as input, and predicts a binary mask (see Fig. 2).

Network Architecture. We use DeepLabv3 [Chen et al.(2017)Chen, Papandreou, Schroff, and Adam] as our segmentation network, , which outputs feature maps of 1/16 of the input image size. Given an input image, , generates a semantic segmentation map, , using the pseudo-label as supervision. The objectness module, , takes as input and consists of a stack of five convolutional layers that includes batch normalization and ReLU layers (see Table S1 in the supplementary for architectural details). We use a 33 kernel in the first four convolution layers and use a 11 kernel in the last layer which outputs the objectness map, . The procedure for obtaining the semantic and objectness maps can be described as:
(4) |
where and refer to trainable weights for the and modules, respectively.
Joint Learning of Semantics and Objectness. We train our proposed SONet method using the generated pixel-level semantic and objectness pseudo-labels in an end-to-end manner (see Fig. 2). Let denote the target semantic segmentation dataset with images , pseudo-labels , and is the objectness prior. More specifically, let be a training image from with semantic segmentation pseudo-label and the objectness prior . We denote the pixel-wise cross entropy loss function and between () and , respectively. The final loss function of the network is the sum of the segmentation and objectness losses as follows:
(5) |
where and denote the multi-class and binary cross entropy loss function, respectively. The joint training allows our network to propagate objectness information together with semantics, and suppress erroneous decisions which allows for more accurate final predictions for both outputs. During inference, we simply take the segmentation map to measure the overall performance of our proposed approach.
3 Experiments
We evaluate our proposed framework on the PASCAL VOC 2012 [Everingham et al.(2015)Everingham, Eslami, Van Gool, Williams, Winn, and Zisserman] and Cityscapes [Cordts et al.(2016)Cordts, Omran, Ramos, Rehfeld, Enzweiler, Benenson, Franke, Roth, and Schiele] semantic segmentation benchmarks. We generate objectness masks for the VOC12 target dataset from two different objectness trained models on COCO-Stuff: overlapping (all images) and non-overlapping (images with no overlapping objects with target data). We also report experiments under domain adaptation settings by training on a different target dataset, OpenV5 [Kuznetsova et al.(2018)Kuznetsova, Rom, Alldrin, Uijlings, Krasin, Pont-Tuset, Kamali, Popov, Malloci, and Duerig], and evaluating on VOC12. OpenV5 [Kuznetsova et al.(2018)Kuznetsova, Rom, Alldrin, Uijlings, Krasin, Pont-Tuset, Kamali, Popov, Malloci, and Duerig] is a recently released dataset consisting of image-level, bounding box, and semantic segmentation annotations for over 600 classes. For these experiments, we randomly select 42,621 images from the same 21 classes as VOC12 and generate pseudo-labels using our box-driven approach. We train SONet with the generated pseudo-labels from OpenV5 and then evaluate on VOC12 (denoted as O V). We also finetune SONet on VOC12 before evaluation (denoted as O + V). We report experimental results with different backbone networks for a fair comparison.
3.1 Analysis of Generated Pseudo-labels
We first evaluate the quality of our generated pseudo-labels to explore the upper bound for different types of weak supervision and report the results in Table 1(a). We consider the generated pseudo-labels as predictions and obtain the upper bound for each supervision type by calculating the mIoU between the pseudo-label and the groundtruth. To generate CAMs pseudo-labels, , we simply threshold the scores of raw CAMs. When we apply our objectness mask, , to improve the boundary of CAMs (), we obtain 70.6% mIoU. Further, using the bounding boxes and objectness map () achieves 76.6% mIoU that further improves the upper bound mIoU by 6%. In addition, applying the non-overlapping objectness mask, , substantially improves the CAMs () or bounding box () proposals. As shown in Table 1(a), exploiting
Sup. | Train |
---|---|
48.3 | |
63.5 | |
70.6 | |
60.2 | |
68.6 | |
76.6 | |
(a) |
Method | Sup. | Val. |
SONet | 50.2 | |
65.3 | ||
70.5 | ||
54.6 | ||
67.9 | ||
73.8 | ||
(b) |
Method | Sup. | mIoU |
O V | ||
SONet | 51.5 | |
71.0 | ||
O + V | ||
SONet | 75.9 | |
76.9 | ||
(c) |
an objectness map with CAMs or bounding boxes significantly improves the quality of pseudo-labels as it removes incorrectly labeled pixels from the CAM and bounding box proposals. Next, we evaluate the performance of our proposed SONet (Table 1(b)) with CAMs, box annotations, and the generated pseudo-labels. SONet trained with box proposals achieves 54.6% mIoU which outperforms the same model trained using CAM proposals (50.2%). When we use our generated pseudo-labels during training, SONet achieves 70.5% mIoU () and 73.8% () on the VOC12 val set. In the domain adaptation settings (see Table 1(c)), SONet trained only with OpenV5 groundtruth boxes achieves 51.5% mIoU accuracy on the VOC12 val set. When SONet is trained on OpenV5 with as supervision, it drastically improves the overall mIoU to 71.0% (note that in this setting we only use OpenV5 images to train SONet). Additionally, fine-tuning SONet on the VOC12 training set with supervision increases the mIoU to 76.9%. These experiments indicate that our pseudo-label generation technique achieves good upper bound performance compared to the groundtruth.
3.2 Image Segmentation Results
Method | Backbone | Guidance | mIoU | |
---|---|---|---|---|
val | test | |||
Weakly-Supervised Approaches | ||||
Image-Level Supervision (CAM) | ||||
FlickleNet [Lee et al.(2019a)Lee, Kim, Lee, Lee, and Yoon] | Res-101 | Saliency | 64.9 | 65.3 |
OAA∗ [Jiang et al.(2019)Jiang, Hou, Cao, Cheng, Wei, and Xiong] | Res-101 | Saliency | 65.2 | 66.4 |
ME [Fan et al.(2020b)Fan, Zhang, and Tan] | Res101 | Saliency | 67.2 | 66.7 |
ICD [Fan et al.(2020a)Fan, Zhang, Song, and Tan] | Res101 | Saliency | 67.8 | 68.0 |
SGAN∗ | Res-101 | Saliency | 67.1 | 67.2 |
SONet-O∗ | Res-101 | Objectness | 64.5 | 65.8 |
SONet∗ | Res-101 | Objectness | 68.1 | 69.7 |
SONet | Res-101 | Objectness | 70.5 | 71.5 |
Box-Level Supervision | ||||
SDI∗ [Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele] | Res-101 | BSDS | 69.4 | - |
BCM∗ [Song et al.(2019)Song, Huang, Ouyang, and Wang] | Res-101 | CRF | 70.2 | - |
SONet∗ | Res-101 | Objectness | 72.2 | 73.7 |
SONet | Res-101 | Objectness | 74.8 | 76.0 |
Semi-Supervised Approaches | ||||
WSSL [Papandreou et al.(2015)Papandreou, Chen, Murphy, and Yuille] | VGG-16 | 1.4k GT | 64.6 | - |
SDI [Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele] | VGG-16 | 1.4k GT | 65.8 | 66.9 |
FickleNet [Lee et al.(2019a)Lee, Kim, Lee, Lee, and Yoon] | VGG-16 | 1.4k GT | 65.8 | - |
SONet | VGG-16 | - | 66.1 | 67 |
In this section, we compare our proposed SONet method with previous state-of-the-art WSSS and SSSS methods [Papandreou et al.(2015)Papandreou, Chen, Murphy, and Yuille, Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele, Huang et al.(2018)Huang, Wang, Wang, Liu, and Wang, Song et al.(2019)Song, Huang, Ouyang, and Wang, Lee et al.(2019a)Lee, Kim, Lee, Lee, and Yoon, Zeng et al.(2019)Zeng, Zhuge, Lu, and Zhang, Jiang et al.(2019)Jiang, Hou, Cao, Cheng, Wei, and Xiong, Fan et al.(2020b)Fan, Zhang, and Tan, Fan et al.(2020a)Fan, Zhang, Song, and Tan, Chang et al.(2020)Chang, Wang, Hung, Piramuthu, Tsai, and Yang]. Table 2 presents a comparison with recent weakly and semi supervised methods using image and bounding box-level supervision. For fair comparison in WSSS setting, we compare with other methods that use ResNet-101 as the backbone and additional guidance (e.g., saliency maps and optical flow) as supervision. SONet∗ outperforms the current state-of-the-art image-level + extra guidance based methods by a reasonable margin, achieving 68.1% mIoU on the VOC12 val set. Interestingly, SONet-O∗, which is trained on the pseudo-labels generated under the non-overlapping settings, also achieves comparable performance with the baseline WSSS methods. When compared to methods that use bounding box-level supervision with extra guidance, SONet∗ also improves upon the state-of-the-art [Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele, Song et al.(2019)Song, Huang, Ouyang, and Wang] by 2.0%. Note that both BCM [Song et al.(2019)Song, Huang, Ouyang, and Wang] and SDI [Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele] take much longer to produce pseudo-labels than our approach due to their iterative procedures and use complex training protocols. We do not include results for a recent box based method, Box2Seg [Kulharia et al.(2020)Kulharia, Chandra, Agrawal, Torr, and Tyagi], as they use a higher capacity network architecture [Xiao et al.(2018b)Xiao, Liu, Zhou, Jiang, and Sun] for segmentation learning without publicly available code. Our SONet method achieves 74.8% mIoU on the VOC12 val set which is very close (2.4% lower) to the fully supervised trained baseline [Chen et al.(2017)Chen, Papandreou, Schroff, and Adam] model (77.2%). In the SSSS setting, we use a similar backbone network as existing methods to ensure a fair comparison. Note, existing SSSS methods use a portion of the target semantic segmentation groundtruth while we solely use our generated pseudo-labels to train the network. Surprisingly, SONet (VGG-16 backbone) marginally outperforms the existing SSSS methods (66.1% vs. 65.8% mIoU). These results demonstrate that our pseudo-label generation procedure is flexible and achieves substantial improvements or competitive performance compared to existing methods in WSSS and SSSS.
Method |
Sup. |
aero |
bike |
bird |
boat |
bottle |
bus |
car |
cat |
chair |
cow |
table |
dog |
horse |
mbike |
person |
plant |
sheep |
sofa |
train |
tv |
mIoU |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
SONet | 84.4 | 37.6 | 83.7 | 63.9 | 51.6 | 88.4 | 84.7 | 68.3 | 30.8 | 81.4 | 57.9 | 68.0 | 79.6 | 83.6 | 74.2 | 58.5 | 84.6 | 53.4 | 81.5 | 57.1 | 69.7 | |
SONet(OV) | 91.3 | 39.9 | 89.7 | 68.5 | 68.6 | 89.8 | 77.0 | 80.8 | 21.4 | 71.9 | 34.2 | 81.3 | 83.4 | 82.4 | 71.4 | 58.6 | 82.9 | 53.2 | 86.4 | 66.5 | 71.1 | |
SONet | 91.4 | 37.3 | 88.7 | 68.0 | 66.0 | 94.1 | 88.0 | 79.9 | 32.3 | 83.4 | 64.3 | 77.5 | 86.3 | 78.0 | 74.4 | 59.4 | 86.3 | 57.3 | 84.8 | 66.2 | 74.1 | |
SONet(O+V) | 90.7 | 40.0 | 90.2 | 69.7 | 72.7 | 94.1 | 87.4 | 79.2 | 32.7 | 86.7 | 62.6 | 80.1 | 88.5 | 81.3 | 74.2 | 62.6 | 91.9 | 58.5 | 89.2 | 69.5 | 76.0 | |
SONet | 92.1 | 40.9 | 90.6 | 68.5 | 74.0 | 94.1 | 87.1 | 83.2 | 31.3 | 86.4 | 67.2 | 78.2 | 84.6 | 84.0 | 77.1 | 61.6 | 90.6 | 55.3 | 85.4 | 69.2 | 76.0 | |
SONet (O+V) | 91.8 | 39.9 | 89.9 | 71.3 | 74.8 | 94.6 | 88.2 | 80.9 | 33.0 | 89.5 | 62.8 | 82.5 | 89.7 | 83.8 | 76.9 | 62.8 | 90.3 | 59.9 | 89.3 | 70.0 | 77.0 | |
DeepLabV3 | 92.9 | 60.3 | 93.0 | 70.5 | 73.3 | 94.1 | 88.1 | 90.9 | 35.3 | 83.4 | 65.7 | 86.3 | 87.5 | 85.2 | 86.5 | 63.8 | 88.1 | 57.6 | 85.0 | 72.3 | 78.8 |
Table 3 presents a class-wise IoU comparison of SONet with different training strategies as well as with the fully supervised baseline DeepLabv3 model on the VOC12 test set. Notably, although the fully supervised model achieves the highest mIoU, SONet (O+V) trained using outperforms the fully supervised model on half of the categories, and is competitive in many others. In general, when trained using bounding box-based pseudo-labels, SONet performs well on rectangular shaped classes, e.g., bus, car, tv, cow, bottle, bus, and train. However, it is still difficult for any training protocol combined with SONet to achieve comparable performance with classes like bike, motorbike, cat, dog or person, where the objects have complex boundary information or are occluded with other classes, e.g., person on a horse or bike. Furthermore, using the ignore strategy ( vs. ) improves the performance notably for both the normal and domain adaptation settings. The quantitative results indicate that our SONet model can achieve competitive performance with the fully supervised model, showing the effectiveness of the proposed pseudo-label generation and the joint learning techniques.
Method | Sup. | mIoU |
---|---|---|
DLabv3 (Things) | Full | 81.5 |
SONet (Things) | 76.6 |
We further use Cityscapes [Cordts et al.(2016)Cordts, Omran, Ramos, Rehfeld, Enzweiler, Benenson, Franke, Roth, and Schiele] as our target dataset and report results in Table 4. Cityscapes consists of eight things classes and 11 stuff classes. Similar to the VOC12, we first generate class agnostic objectness masks for the Cityscapes train set and combine it with the bounding box annotations to generate semantic pseudo-labels. Since our objectness network is not trained for generating masks for stuff classes, we only consider the things classes from Cityscapes during pseudo-label generation, training, and evaluation. Next, we train DeepLabv3-ResNet50 [Chen et al.(2017)Chen, Papandreou, Schroff, and Adam] with full supervision as a baseline and SONet (DeepLabv3-ResNet50 as backbone) using the generated pseudo-labels (). Table 4 shows that, our SONet performs well (76.6% mIoU) and obtains 94% relative to the baseline (similar to our results on VOC12). This result further confirms the generalizability of our pseudo-label generation technique despite the significant distribution gap between the target (Cityscapes [Cordts et al.(2016)Cordts, Omran, Ramos, Rehfeld, Enzweiler, Benenson, Franke, Roth, and Schiele]) and the source (COCOStuff [Caesar et al.(2018)Caesar, Uijlings, and Ferrari]) dataset.
3.3 Ablation Studies
Effectiveness of Objectness Branch.
We first validate the effect of the objectness branch by comparing the results of SONet trained in both multi- and single-task settings. We train SONet with and without the objectness branch. Note that SONet without the objectness branch is equivalent to DeepLabv3 [Chen et al.(2018)Chen, Papandreou, Kokkinos, Murphy, and
Yuille]. The result of these comparisons are summarized in Table 5 (a). It is clear that the models trained with the objectness branch achieve superior performance compared to the models trained only for the task of semantic segmentation. Interestingly, the objectness branch improves the boundary details to bring more smoothness (see Fig. in Table 5 (top row)) as expected, as well as the semantic information (see the figure in Table 5 (bottom row)). SONet’s multi-task objective not only provides it with the ability to robustly predict both binary and semantic segmentation, but the objectness-based learning naturally provides the segmentation network a significant performance boost.
Effectiveness of Ignoring Strategy. In Table 5(b), we compare our ignore strategy to the strategy in SDI [Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and
Schiele] when trained using SONet. We show that our ignoring strategy outperforms both SDI [Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and
Schiele] and SONet trained without an ignore strategy.
Sup. | VOC12 | O V | ||
---|---|---|---|---|
[Chen et al.(2018)Chen, Papandreou, Kokkinos, Murphy, and Yuille] | SONet | [Chen et al.(2018)Chen, Papandreou, Kokkinos, Murphy, and Yuille] | SONet | |
52.1 | 54.6 | 50.5 | 51.5 | |
73.1 | 73.8 | 70.4 | 71.0 | |
(a) |
Sup. | Val. | Test |
---|---|---|
[Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele] | 67.9 | - |
[Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele] | 67.6 | - |
73.8 | 74.1 | |
74.8 | 76.0 | |
(b) |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
Image | GT | DeepLabv3 | Mask | SONet |
Improving Semantic Proposals: Objectness or Saliency Guidance? It is common to utilize pixel-level saliency information as additional guidance to be combined with the CAM proposals [Oh et al.(2017)Oh, Benenson, Khoreva, Akata, Fritz, and Schiele, Wei et al.(2018)Wei, Xiao, Shi, Jie, Feng, and Huang, Zeng et al.(2019)Zeng, Zhuge, Lu, and Zhang, Yao and Gong(2020)]. Specifically, DHSNet [Liu and Han(2016)] and DSS [Hou et al.(2017)Hou, Cheng, Hu, Borji, Tu, and Torr] have been used in [Hou et al.(2018)Hou, Jiang, Wei, and Cheng, Chaudhry et al.(2017)Chaudhry, Dokania, and Torr, Wei et al.(2018)Wei, Xiao, Shi, Jie, Feng, and Huang] to generate a saliency mask for each training sample. This guidance of saliency can deliver non-semantic pixel-level supervision for a better boundary segmentation. However, the saliency information used in the previous studies only focus on the most salient object due to the problem of centre bias [Bruce et al.(2015)Bruce, Wloka, Frosst, Rahman, and Tsotsos, Açık et al.(2014)Açık, Bartel, and Koenig]. For instance, as shown in Fig. 3(a), the masks generated by saliency models can only detect the objects near the centre of an image, the “ship” near the corner will be incorrectly labelled as background (top row). Furthermore, the region of the “train” can only be partially labelled because the back of the train is not salient. This problem introduces outliers (incorrectly labelled regions) when training a segmentation model. In contrast, our proposed objectness model learns to recognize objects in all image locations, even if they are not salient or near the image boundary, see Fig. 3(a). Figure 3(b) further illustrates that the objectness network is equally likely to make errors in all image locations, while the saliency detection network is biased towards making erroneous predictions near the image border. To validate this claim quantitatively, we conduct experiments (see Table in Fig. 3(c)) by replacing the objectness mask with a saliency mask for creating semantic pseudo-labels. We use two recent saliency detectors, PiCaNet [Liu et al.(2018)Liu, Han, and Yang] and BASNet [Qin et al.(2019)Qin, Zhang, Huang, Gao, Dehghan, and Jagersand] to generate the saliency mask for VOC12 training images. Combining saliency masks with and achieves performance which is significantly lower than the quality of our pseudo-labels generated using the objectness guidance.
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
Image | Ours | BASNet [Qin et al.(2019)Qin, Zhang, Huang, Gao, Dehghan, and Jagersand] | PicaNet [Liu et al.(2018)Liu, Han, and Yang] |
(a) |
![]() |
![]() |
![]() |
|
Saliency [Qin et al.(2019)Qin, Zhang, Huang, Gao, Dehghan, and Jagersand] | Objectness |
(b) |
* | ||
---|---|---|
PicaNet | 53.7 | 64.7 |
BASNet | 58.5 | 64.0 |
Ours | 70.6 | 76.6 |
(c) |
4 Discussion and Conclusion
Existing saliency-based WSSS [Papandreou et al.(2015)Papandreou, Chen, Murphy, and Yuille, Huang et al.(2018)Huang, Wang, Wang, Liu, and Wang, Lee et al.(2019a)Lee, Kim, Lee, Lee, and Yoon, Zeng et al.(2019)Zeng, Zhuge, Lu, and Zhang, Jiang et al.(2019)Jiang, Hou, Cao, Cheng, Wei, and Xiong, Fan et al.(2020b)Fan, Zhang, and Tan, Fan et al.(2020a)Fan, Zhang, Song, and Tan] and SSSS [Lee et al.(2019a)Lee, Kim, Lee, Lee, and Yoon] methods utilize both saliency detectors (trained on the DUT-S [Wang et al.(2017)Wang, Lu, Wang, Feng, Wang, Yin, and Ruan] or MSRA-B [Jiang et al.(2013)Jiang, Wang, Yuan, Wu, Zheng, and Li] datasets which have pixel-level binary segmentation ground-truth for a large number of overlapping instances in the VOC12 dataset) and a portion of semantic segmentation GT, respectively. Following this line of work, we choose the objectness-based dataset to introduce a better proposal model which addresses the severe center bias issue of saliency detectors (see Fig. 3 (a, b)) for WSSS (e.g., saliency inherently ignores objects near the border). We compare with both WSSS and SSSS techniques since we do not fall neatly within either category of supervision (i.e., comparing against methods which use only CAM is unfair but we also do not use any semantic segmentation GT). Moreover, in contrast to the previous methods [Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele, Xiao et al.(2018a)Xiao, Wei, Liu, Zhang, and Feng, Ahn and Kwak(2018), Ahn et al.(2019)Ahn, Cho, and Kwak], our framework does not require multiple stages of label inference and training for pseudo-label generation, but instead operates in a single stage. Additionally, the objectness branch improves the performance of the segmentation network by propagating boundary and semantic information back through the network. We believe the objectness branch helps with semantics because it forces the model to treat objects more uniformly (since the objectness label is binary). This can guide the segmentation model to treat nearby pixels as the same semantic object class and promote more spatially uniform predictions, which is correct in many cases.
In summary, we have presented a pseudo-label generation and joint learning strategy for the tasks of both WSSS and SSSS. We first introduced a novel technique to generate high quality pseudo-labels that combines class agnostic objectness priors with either image-level labels or bounding box annotations. Next, we proposed a model that jointly learns semantics and objectness to guide the network to encode more accurate boundary information and better semantic representations. We conducted an extensive set of experiments under different settings and supervision strategies to validate the effectiveness of the proposed methods. The ablation studies isolated the improvements due to the proposed objectness branch, and validated the efficacy of our ignoring strategy. Furthermore, the pseudo-label generation pipeline is simple, efficient, and can be used for large-scale data annotation.
Simpler Does It: Generating Semantic Labels with Objectness Guidance
–Supplementary Materials–

S1 Examples of Generated Pseudo-labels on VOC 2012
Figure S1 shows additional examples of the generated pseudo-labels by combining the class agnostic objectness priors with either CAM [Zhou et al.(2016)Zhou, Khosla, Lapedriza, Oliva, and Torralba] or bounding box proposals. Our pseudo-label generation technique successfully extracts boundary information from the objectness prior and class information from the CAM or bounding box proposals, resulting in high-quality pseudo-labels with fine-grained details about the object’s shape. For the ignore strategy, we assign values of 255 to the outer regions of a bounding box if it overlaps (above a certain threshold) with the largest bounding box in an image. These overlapped semantic regions have a high degree of uncertainty due to the inherent structure of bounding boxes and ignoring these regions during training results in better predictions (see Fig. S4).
S2 Details of SONet Architecture
We discussed the SONet architecture in Sec. 2.2 of the main manuscript. The details of the objectness module in SONet architecture are shown in Table S1. The input to the objectness module is the segmentation map, which is generated by the DeepLabv3-ResNet101 network. The objectness module consists of five convolution layers where first four layers gradually increase the depth (i.e., channel) of the feature map. The last convolution layer predicts the desired objectness map, . Note that we apply batch normalization and ReLU layers after each convolution layers except the last one which predicts the objectness map. The newly introduced convolution layers are trained from scratch.
Input: Segmentation Map |
Conv2d (), Batch Norm, ReLU |
Conv2d (), Batch Norm, ReLU |
Conv2d (), Batch Norm, ReLU |
Conv2d (), Batch Norm, ReLU |
Conv2d (), Batch Norm, ReLU |
Output: Objectness Map |
S3 Supplementary Experiments
In this section, we first provide implementation details of our proposed SONet (Sec. S3.1) and a description of the OpenV5 dataset (Sec. S3.2). Then, we provide anonymous links to the PASCAL VOC 2012 test set results and additional qualitative examples predicted by SONet with different levels of supervision (Sec. S3.3). Further we conduct experiments on video object segmentation (Sec. S3.4). We also show the generality of our proposed pseudo-label generation technique on the Berkeley DeepDrive dataset [Yu et al.(2018)Yu, Xian, Chen, Liu, Liao, Madhavan, and Darrell] (Sec. S3.5). Finally, we report a series of ablation studies (Sec. S3.6).
S3.1 Implementation Details
We implement our method using the PyTorch [Paszke et al.(2019)Paszke, Gross, Massa, Lerer, Bradbury, Chanan, Killeen, Lin, Gimelshein, Antiga, et al.] framework trained end-to-end on two NVIDIA GeForce GTX 1080 Ti GPUs. We use the SGD optimizer to train our network. We train all the variants of our SONet for 40 epochs with an initial learning rate of 2e-3. We use a random crop of 513513 and 321321 during training for SONet and SONet∗, respectively. Similarly, we use a output stride of 16 and 8 during training for SONet and SONet∗, respectively. During inference, we use a crop of 513513 and rescale to the original size using simple bilinear interpolation before calculating the mIoU. Following the current practice [Chen et al.(2018)Chen, Papandreou, Kokkinos, Murphy, and Yuille, Zhao et al.(2017)Zhao, Shi, Qi, Wang, and Jia, Noh et al.(2015)Noh, Hong, and Han], to report test set results on PASCAL VOC 2012, we first train on the augmented training set followed by fine-tuning on the original trainval set with the generated pseudo-labels.

S3.2 OpenV5 Dataset
We have shown experimental results using the OpenV5 dataset for the task of semantic segmentation in Table 1(c) of the main paper. We compare against the state-of-the-art by using the standard protocol (training on PASCAL VOC 2012 augmented train set and evaluate on PASCAL VOC 2012 val/test set). As mentioned in Sec. 3.1 of the main manuscript, we use a subset of the OpenV5 dataset, where each semantic category is contained in a large number of images, consisting of 42,621 total images and 20 semantic categories. Figure S2 shows the comparison of the object instance distribution of the OpenV5 subset and PASCAL VOC 2012 dataset. It is evident from the table that there are a considerable number of instances for each semantic category and person is by far the most dominant category as expected since it co-occurs with most of other categories. Figure S3 shows examples of the generated pseudo-labels for OpenV5, by combining the class agnostic objectness priors with bounding box proposals.


S3.3 PASCAL VOC 2012 Test Set Results
We illustrate additional visual examples predicted by SONet with different levels (CAMs and box-driven) of supervision on PASCAL VOC 2012 validation images in Fig. S4. The segmentation mask generated by SONet produces more accurate results when trained with CAMs or box-driven pseudo-labels than SONet trained solely with CAMs or bounding box annotations.
S3.4 Video Object Segmentation Results
We also experiment on the YouTube-Object (YTO) dataset [Prest et al.(2012)Prest, Leistner, Civera, Schmid, and Ferrari] to show the effectiveness of our method in segmenting objects from videos by simply evaluating the results produced by SONet. Following prior works [Tang et al.(2013)Tang, Sukthankar, Yagnik, and Fei-Fei, Papazoglou and Ferrari(2013), Lee et al.(2019b)Lee, Kim, Lee, Lee, and Yoon], we use the groundtruth segmentation masks provided by [Jain and Grauman(2014)] to evaluate the performance of SONet and also compare our method with recent video segmentation methods with weak supervision in Table S2. Note that all the baseline methods are explicitly trained on video datasets and use temporal cues, while our method is trained on static images without temporal information. Our SONet method outperforms the existing methods which use different levels of supervision. This may be because objectness-driven pseudo-labels provide more fine-grained localization with sharper object boundaries than coarse bounding boxes. Samples of the predicted masks for the YTO dataset are shown in Fig. S5.
OVS [Drayer and Brox(2016)] |
SONet |
|||||||
Temporal | X | |||||||
Sup. | ||||||||
mIoU | 54.1 | 56.2 | 61.7 | 53.3 | 58.6 | 61.9 | 62.1 | 64.3 |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
S3.5 Generalization to Different Domains: Berkeley DeepDrive
We further apply our bounding box-driven pseudo-label generation technique on a recent driving dataset, Berkeley DeepDrive [Yu et al.(2018)Yu, Xian, Chen, Liu, Liao, Madhavan, and Darrell], to validate whether our procedure can generalize well on a dataset from a different domain. The Berkeley DeepDrive dataset [Yu et al.(2018)Yu, Xian, Chen, Liu, Liao, Madhavan, and Darrell] is composed of images of diverse road scenes (with motion blur) taken from various locations throughout the USA. We generate pseudo-labels for 100k frames which have bounding box annotation available for the 10 different categories: bus, light, sign, person, bike, truck, motor, car, train, and rider. Figure S6 presents examples of generated pseudo-labels of DeepDrive video frames. It is clear that our class agnostic objectness model can generate masks with sharp boundaries in complex driving scenarios, resulting in high-quality pseudo-labels. Since the DeepDrive dataset does not provide pixel-wise annotation for these 100k frames we can not evaluate the quality of generated pseudo-labels in terms of mIoU.

S3.6 Ablation Studies
We conduct further ablation studies to analyze our design and the effectiveness of the objectness branch (Sec. S3.6.1 & Sec. S3.6.2).
S3.6.1 Design Choices of Objectness Branch.
We vary the design of the objectness branch, , of SONet and compare the architectures against each other. The results are reported in Table S3. We evaluate three different variants: (v1) a single 11 convolutional layer which predicts the objectness and takes as input the final feature representation (res5C), (v2) a single 11 convolution layer which takes as input the semantic prediction (), and (v3) a smaller network is applied (as discussed in Sec. 3.2 of the main manuscript) but takes as input the features from res5C.
S3.6.2 Effectiveness of Objectness Branch in SONet
We provide additional qualitative examples in Fig. S7 to show the objectness branch’s effect on SONet’s semantic segmentation predictions. Note that SONet without the objectness branch is equivalent to DeepLabv3 [Chen et al.(2018)Chen, Papandreou, Kokkinos, Murphy, and Yuille]. As can be seen from the examples, the objectness branch can guide the segmentation network to produce more accurate and smooth predictions.
Name | Sup. | Architecture () | Input | mIoU |
SONet | smaller network discussed in Sec. 3.3 | semantic () | 73.8 | |
v1 | single 11 convolution layer | res5C | 72.2 | |
v2 | single 11 convolution layer | semantic () | 73.5 | |
v3 | smaller network discussed in Sec. 3.3 | res5C | 73.6 |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
Image | GT | DeepLabv3 [Chen et al.(2018)Chen, Papandreou, Kokkinos, Murphy, and Yuille] | Objectness | SONet |
S3.6.3 Transferring Semantic Knowledge from Source to Target Dataset.
As an additional baseline, we directly transfer the semantic information from COCOStuff to the VOC12 dataset. Towards this goal, we first train DeepLabv3 [Chen et al.(2017)Chen, Papandreou, Schroff, and Adam] on COCOStuff to output semantic segmentation (i.e., multi-class) masks instead of objectness masks (i.e., binary). Note, similar to the objectness training, we only consider the things classes and use the pretrained model to generate pseudo-label (quality: 50.8% mIoU) for the VOC12 train set. Then, we train DeepLabv3 using the generated pseudo-labels, resulting in 53.4% mIoU on VOC12 val set.
References
- [Açık et al.(2014)Açık, Bartel, and Koenig] Alper Açık, Andreas Bartel, and Peter Koenig. Real and implied motion at the center of gaze. Journal of Vision, 2014.
- [Ahn and Kwak(2018)] Jiwoon Ahn and Suha Kwak. Learning pixel-level semantic affinity with image-level supervision for weakly supervised semantic segmentation. In CVPR, 2018.
- [Ahn et al.(2019)Ahn, Cho, and Kwak] Jiwoon Ahn, Sunghyun Cho, and Suha Kwak. Weakly supervised learning of instance segmentation with inter-pixel relations. In CVPR, 2019.
- [Badrinarayanan et al.(2017)Badrinarayanan, Kendall, and Cipolla] Vijay Badrinarayanan, Alex Kendall, and Roberto Cipolla. Segnet: A deep convolutional encoder-decoder architecture for scene segmentation. TPAMI, 2017.
- [Bruce et al.(2015)Bruce, Wloka, Frosst, Rahman, and Tsotsos] Neil DB Bruce, Calden Wloka, Nick Frosst, Shafin Rahman, and John K Tsotsos. On computational modeling of visual saliency: Examining what’s right, and what’s left. Vision research, 2015.
- [Caesar et al.(2018)Caesar, Uijlings, and Ferrari] Holger Caesar, Jasper Uijlings, and Vittorio Ferrari. COCO-Stuff: Thing and stuff classes in context. In CVPR, 2018.
- [Chang et al.(2020)Chang, Wang, Hung, Piramuthu, Tsai, and Yang] Yu-Ting Chang, Qiaosong Wang, Wei-Chih Hung, Robinson Piramuthu, Yi-Hsuan Tsai, and Ming-Hsuan Yang. Weakly-supervised semantic segmentation via sub-category exploration. In CVPR, 2020.
- [Chaudhry et al.(2017)Chaudhry, Dokania, and Torr] Arslan Chaudhry, Puneet K Dokania, and Philip HS Torr. Discovering class-specific pixels for weakly-supervised semantic segmentation. arXiv preprint arXiv:1707.05821, 2017.
- [Chen et al.(2015)Chen, Papandreou, Kokkinos, Murphy, and Yuille] Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L. Yuille. Semantic image segmentation with deep convolutional nets and fully connected CRFs. In ICLR, 2015.
- [Chen et al.(2017)Chen, Papandreou, Schroff, and Adam] Liang-Chieh Chen, George Papandreou, Florian Schroff, and Hartwig Adam. Rethinking atrous convolution for semantic image segmentation. arXiv:1706.05587, 2017.
- [Chen et al.(2018)Chen, Papandreou, Kokkinos, Murphy, and Yuille] Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L Yuille. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. TPAMI, 2018.
- [Chen et al.(2020)Chen, Wu, Fu, Han, and Zhang] Liyi Chen, Weiwei Wu, Chenchen Fu, Xiao Han, and Yuntao Zhang. Weakly supervised semantic segmentation with boundary exploration. In ECCV, 2020.
- [Cheng et al.(2017)Cheng, Tsai, Wang, and Yang] Jingchun Cheng, Yi-Hsuan Tsai, Shengjin Wang, and Ming-Hsuan Yang. SegFlow: Joint learning for video object segmentation and optical flow. In ICCV, 2017.
- [Cordts et al.(2016)Cordts, Omran, Ramos, Rehfeld, Enzweiler, Benenson, Franke, Roth, and Schiele] Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, Stefan Roth, and Bernt Schiele. The Cityscapes dataset for semantic urban scene understanding. In CVPR, 2016.
- [Dai et al.(2015)Dai, He, and Sun] Jifeng Dai, Kaiming He, and Jian Sun. Boxsup: Exploiting bounding boxes to supervise convolutional networks for semantic segmentation. In ICCV, 2015.
- [Drayer and Brox(2016)] Benjamin Drayer and Thomas Brox. Object detection, tracking, and motion segmentation for object-level video segmentation. arXiv preprint arXiv:1608.03066, 2016.
- [Everingham et al.(2015)Everingham, Eslami, Van Gool, Williams, Winn, and Zisserman] Mark Everingham, SM Ali Eslami, Luc Van Gool, Christopher KI Williams, John Winn, and Andrew Zisserman. The Pascal visual object classes challenge: A retrospective. IJCV, 2015.
- [Fan et al.(2020a)Fan, Zhang, Song, and Tan] Junsong Fan, Zhaoxiang Zhang, Chunfeng Song, and Tieniu Tan. Learning integral objects with intra-class discriminator for weakly-supervised semantic segmentation. In CVPR, 2020a.
- [Fan et al.(2020b)Fan, Zhang, and Tan] Junsong Fan, Zhaoxiang Zhang, and Tieniu Tan. Employing multi-estimations for weakly-supervised semantic segmentation. In ECCV, 2020b.
- [Ghiasi and Fowlkes(2016)] Golnaz Ghiasi and Charless C Fowlkes. Laplacian pyramid reconstruction and refinement for semantic segmentation. In ECCV, 2016.
- [He et al.(2016)He, Zhang, Ren, and Sun] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, 2016.
- [He et al.(2021)He, Lu, Wang, Song, and Zhou] Lei He, Jiwen Lu, Guanghui Wang, Shiyu Song, and Jie Zhou. Sosd-net: Joint semantic object segmentation and depth estimation from monocular images. Neurocomputing, 2021.
- [Hong et al.(2017)Hong, Yeo, Kwak, Lee, and Han] Seunghoon Hong, Donghun Yeo, Suha Kwak, Honglak Lee, and Bohyung Han. Weakly supervised semantic segmentation using web-crawled videos. In CVPR, 2017.
- [Hou et al.(2017)Hou, Cheng, Hu, Borji, Tu, and Torr] Qibin Hou, Ming-Ming Cheng, Xiaowei Hu, Ali Borji, Zhuowen Tu, and Philip Torr. Deeply supervised salient object detection with short connections. In CVPR, 2017.
- [Hou et al.(2018)Hou, Jiang, Wei, and Cheng] Qibin Hou, PengTao Jiang, Yunchao Wei, and Ming-Ming Cheng. Self-erasing network for integral object attention. In NIPS, 2018.
- [Hu et al.(2018)Hu, Dollár, He, Darrell, and Girshick] Ronghang Hu, Piotr Dollár, Kaiming He, Trevor Darrell, and Ross Girshick. Learning to segment every thing. In CVPR, 2018.
- [Huang et al.(2018)Huang, Wang, Wang, Liu, and Wang] Zilong Huang, Xinggang Wang, Jiasi Wang, Wenyu Liu, and Jingdong Wang. Weakly-supervised semantic segmentation network with deep seeded region growing. In CVPR, 2018.
- [Ibrahim et al.(2018)Ibrahim, Vahdat, and Macready] Mostafa S Ibrahim, Arash Vahdat, and William G Macready. Weakly supervised semantic image segmentation with self-correcting networks. arXiv preprint arXiv:1811.07073, 2018.
- [Islam et al.(2017a)Islam, Naha, Rochan, Bruce, and Wang] Md Amirul Islam, Shujon Naha, Mrigank Rochan, Neil Bruce, and Yang Wang. Label refinement network for coarse-to-fine semantic segmentation. arXiv:1703.00551, 2017a.
- [Islam et al.(2017b)Islam, Rochan, Bruce, and Wang] Md Amirul Islam, Mrigank Rochan, Neil D. B. Bruce, and Yang Wang. Gated feedback refinement network for dense image labeling. In CVPR, 2017b.
- [Islam et al.(2018a)Islam, Kalash, and Bruce] Md Amirul Islam, Mahmoud Kalash, and Neil D.B. Bruce. Revisiting salient object detection: Simultaneous detection, ranking, and subitizing of multiple salient objects. In CVPR, 2018a.
- [Islam et al.(2018b)Islam, Kalash, and Bruce] Md Amirul Islam, Mahmoud Kalash, and Neil DB Bruce. Semantics meet saliency: Exploring domain affinity and models for dual-task prediction. In BMVC, 2018b.
- [Islam et al.(2020)Islam, Kowal, Derpanis, and Bruce] Md Amirul Islam, Matthew Kowal, Konstantinos G Derpanis, and Neil DB Bruce. Feature binding with category-dependant mixup for semantic segmentation and adversarial robustness. In BMVC, 2020.
- [Islam et al.(2021a)Islam, Kowal, Derpanis, and Bruce] Md Amirul Islam, Matthew Kowal, Konstantinos G Derpanis, and Neil DB Bruce. Segmix: Co-occurrence driven mixup for semantic segmentation and adversarial robustness. arXiv preprint arXiv:2108.09929, 2021a.
- [Islam et al.(2021b)Islam, Kowal, Esser, Jia, Ommer, Derpanis, and Bruce] Md Amirul Islam, Matthew Kowal, Patrick Esser, Sen Jia, Björn Ommer, Konstantinos G. Derpanis, and Neil Bruce. Shape or texture: Understanding discriminative features in CNNs. In ICLR, 2021b.
- [Jain and Grauman(2014)] Suyog Dutt Jain and Kristen Grauman. Supervoxel-consistent foreground propagation in video. In ECCV, 2014.
- [Jiang et al.(2013)Jiang, Wang, Yuan, Wu, Zheng, and Li] Huaizu Jiang, Jingdong Wang, Zejian Yuan, Yang Wu, Nanning Zheng, and Shipeng Li. Salient object detection: A discriminative regional feature integration approach. In CVPR, 2013.
- [Jiang et al.(2019)Jiang, Hou, Cao, Cheng, Wei, and Xiong] Peng-Tao Jiang, Qibin Hou, Yang Cao, Ming-Ming Cheng, Yunchao Wei, and Hong-Kai Xiong. Integral object mining via online attention accumulation. In ICCV, 2019.
- [Karim et al.(2019)Karim, Islam, and Bruce] Rezaul Karim, Md Amirul Islam, and Neil DB Bruce. Recurrent iterative gating networks for semantic segmentation. In WACV, 2019.
- [Karim et al.(2020)Karim, Islam, and Bruce] Rezaul Karim, Md Amirul Islam, and Neil DB Bruce. Distributed iterative gating networks for semantic segmentation. In WACV, 2020.
- [Khoreva et al.(2017)Khoreva, Benenson, Hosang, Hein, and Schiele] Anna Khoreva, Rodrigo Benenson, Jan Hosang, Matthias Hein, and Bernt Schiele. Simple does it: Weakly supervised instance and semantic segmentation. In CVPR, 2017.
- [Krähenbühl and Koltun(2011)] Philipp Krähenbühl and Vladlen Koltun. Efficient inference in fully connected crfs with gaussian edge potentials. In NIPS, 2011.
- [Kulharia et al.(2020)Kulharia, Chandra, Agrawal, Torr, and Tyagi] Viveka Kulharia, Siddhartha Chandra, Amit Agrawal, Philip Torr, and Ambrish Tyagi. Box2Seg: Attention weighted loss and discriminative feature learning for weakly supervised segmentation. In ECCV, 2020.
- [Kuznetsova et al.(2018)Kuznetsova, Rom, Alldrin, Uijlings, Krasin, Pont-Tuset, Kamali, Popov, Malloci, and Duerig] Alina Kuznetsova, Hassan Rom, Neil Alldrin, Jasper Uijlings, Ivan Krasin, Jordi Pont-Tuset, Shahab Kamali, Stefan Popov, Matteo Malloci, and Tom Duerig. The open images dataset v4: Unified image classification, object detection, and visual relationship detection at scale. arXiv preprint arXiv:1811.00982, 2018.
- [Kwak et al.(2017)Kwak, Hong, and Han] Suha Kwak, Seunghoon Hong, and Bohyung Han. Weakly supervised semantic segmentation using superpixel pooling network. In AAAI, 2017.
- [Lee et al.(2019a)Lee, Kim, Lee, Lee, and Yoon] Jungbeom Lee, Eunji Kim, Sungmin Lee, Jangho Lee, and Sungroh Yoon. FickleNet: Weakly and semi-supervised semantic image segmentation using stochastic inference. In CVPR, 2019a.
- [Lee et al.(2019b)Lee, Kim, Lee, Lee, and Yoon] Jungbeom Lee, Eunji Kim, Sungmin Lee, Jangho Lee, and Sungroh Yoon. Frame-to-frame aggregation of active regions in web videos for weakly supervised semantic segmentation. In ICCV, 2019b.
- [Li et al.(2018)Li, Arnab, and Torr] Qizhu Li, Anurag Arnab, and Philip HS Torr. Weakly-and semi-supervised panoptic segmentation. In ECCV, 2018.
- [Lin et al.(2016)Lin, Dai, Jia, He, and Sun] Di Lin, Jifeng Dai, Jiaya Jia, Kaiming He, and Jian Sun. Scribblesup: Scribble-supervised convolutional networks for semantic segmentation. In CVPR, 2016.
- [Lin et al.(2017)Lin, Milan, Shen, and Reid] G. Lin, A. Milan, C. Shen, and I. Reid. RefineNet: Multi-path refinement networks for high-resolution semantic segmentation. In CVPR, 2017.
- [Liu and Han(2016)] Nian Liu and Junwei Han. Dhsnet: Deep hierarchical saliency network for salient object detection. In CVPR, 2016.
- [Liu et al.(2018)Liu, Han, and Yang] Nian Liu, Junwei Han, and Ming-Hsuan Yang. PicaNet: Learning pixel-wise contextual attention for saliency detection. In CVPR, 2018.
- [Long et al.(2015)Long, Shelhamer, and Darrell] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In CVPR, 2015.
- [Noh et al.(2015)Noh, Hong, and Han] Hyeonwoo Noh, Seunghoon Hong, and Bohyung Han. Learning deconvolution network for semantic segmentation. In ICCV, 2015.
- [Oh et al.(2017)Oh, Benenson, Khoreva, Akata, Fritz, and Schiele] Seong Joon Oh, Rodrigo Benenson, Anna Khoreva, Zeynep Akata, Mario Fritz, and Bernt Schiele. Exploiting saliency for object segmentation from image level labels. In CVPR, 2017.
- [Papandreou et al.(2015)Papandreou, Chen, Murphy, and Yuille] George Papandreou, Liang-Chieh Chen, Kevin P Murphy, and Alan L Yuille. Weakly-and semi-supervised learning of a deep convolutional network for semantic image segmentation. In ICCV, 2015.
- [Papazoglou and Ferrari(2013)] Anestis Papazoglou and Vittorio Ferrari. Fast object segmentation in unconstrained video. In ICCV, 2013.
- [Paszke et al.(2019)Paszke, Gross, Massa, Lerer, Bradbury, Chanan, Killeen, Lin, Gimelshein, Antiga, et al.] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. In NeurIPS, 2019.
- [Pinheiro and Collobert(2015)] Pedro O Pinheiro and Ronan Collobert. From image-level to pixel-level labeling with convolutional networks. In CVPR, 2015.
- [Pont-Tuset et al.(2016)Pont-Tuset, Arbelaez, Barron, Marques, and Malik] Jordi Pont-Tuset, Pablo Arbelaez, Jonathan T Barron, Ferran Marques, and Jitendra Malik. Multiscale combinatorial grouping for image segmentation and object proposal generation. TPAMI, 2016.
- [Prest et al.(2012)Prest, Leistner, Civera, Schmid, and Ferrari] Alessandro Prest, Christian Leistner, Javier Civera, Cordelia Schmid, and Vittorio Ferrari. Learning object class detectors from weakly annotated video. In CVPR, 2012.
- [Qin et al.(2019)Qin, Zhang, Huang, Gao, Dehghan, and Jagersand] Xuebin Qin, Zichen Zhang, Chenyang Huang, Chao Gao, Masood Dehghan, and Martin Jagersand. Basnet: Boundary-aware salient object detection. In CVPR, 2019.
- [Rother et al.(2004)Rother, Kolmogorov, and Blake] Carsten Rother, Vladimir Kolmogorov, and Andrew Blake. Grabcut: Interactive foreground extraction using iterated graph cuts. In TOG, 2004.
- [Saleh et al.(2017)Saleh, Aliakbarian, Salzmann, Petersson, and Alvarez] Fatemeh Sadat Saleh, Mohammad Sadegh Aliakbarian, Mathieu Salzmann, Lars Petersson, and Jose M Alvarez. Bringing background into the foreground: Making all classes equal in weakly-supervised video semantic segmentation. In ICCV, 2017.
- [Song et al.(2019)Song, Huang, Ouyang, and Wang] Chunfeng Song, Yan Huang, Wanli Ouyang, and Liang Wang. Box-driven class-wise region masking and filling rate guided loss for weakly supervised semantic segmentation. In CVPR, 2019.
- [Souly et al.(2017)Souly, Spampinato, and Shah] Nasim Souly, Concetto Spampinato, and Mubarak Shah. Semi supervised semantic segmentation using generative adversarial network. In ICCV, 2017.
- [Takikawa et al.(2019)Takikawa, Acuna, Jampani, and Fidler] Towaki Takikawa, David Acuna, Varun Jampani, and Sanja Fidler. Gated-SCNN: Gated shape cnns for semantic segmentation. In ICCV, 2019.
- [Tang et al.(2013)Tang, Sukthankar, Yagnik, and Fei-Fei] Kevin Tang, Rahul Sukthankar, Jay Yagnik, and Li Fei-Fei. Discriminative segment annotation in weakly labeled video. In CVPR, 2013.
- [Wang et al.(2017)Wang, Lu, Wang, Feng, Wang, Yin, and Ruan] Lijun Wang, Huchuan Lu, Yifan Wang, Mengyang Feng, Dong Wang, Baocai Yin, and Xiang Ruan. Learning to detect salient objects with image-level supervision. In CVPR, 2017.
- [Wang et al.(2016)Wang, Liu, Li, Yan, and Lu] Yuhang Wang, Jing Liu, Yong Li, Junjie Yan, and Hanqing Lu. Objectness-aware semantic segmentation. In ICM, 2016.
- [Wei et al.(2018)Wei, Xiao, Shi, Jie, Feng, and Huang] Yunchao Wei, Huaxin Xiao, Honghui Shi, Zequn Jie, Jiashi Feng, and Thomas S Huang. Revisiting dilated convolution: A simple approach for weakly-and semi-supervised semantic segmentation. In CVPR, 2018.
- [Xiao et al.(2018a)Xiao, Wei, Liu, Zhang, and Feng] Huaxin Xiao, Yunchao Wei, Yu Liu, Maojun Zhang, and Jiashi Feng. Transferable semi-supervised semantic segmentation. In AAAI, 2018a.
- [Xiao et al.(2018b)Xiao, Liu, Zhou, Jiang, and Sun] Tete Xiao, Yingcheng Liu, Bolei Zhou, Yuning Jiang, and Jian Sun. Unified perceptual parsing for scene understanding. In ECCV, 2018b.
- [Xiong et al.(2018)Xiong, Jain, and Grauman] Bo Xiong, Suyog Dutt Jain, and Kristen Grauman. Pixel objectness: learning to segment generic objects automatically in images and videos. TPAMI, 2018.
- [Yang et al.(2018)Yang, Han, Zhang, Liu, and Zhang] Le Yang, Junwei Han, Dingwen Zhang, Nian Liu, and Dong Zhang. Segmentation in weakly labeled videos via a semantic ranking and optical warping network. TIP, 2018.
- [Yao and Gong(2020)] Qi Yao and Xiaojin Gong. Saliency guided self-attention network for weakly and semi-supervised semantic segmentation. IEEE Access, 2020.
- [Yu and Koltun(2016)] Fisher Yu and Vladlen Koltun. Multi-scale context aggregation by dilated convolutions. In ICLR, 2016.
- [Yu et al.(2018)Yu, Xian, Chen, Liu, Liao, Madhavan, and Darrell] Fisher Yu, Wenqi Xian, Yingying Chen, Fangchen Liu, Mike Liao, Vashisht Madhavan, and Trevor Darrell. Bdd100k: A diverse driving video database with scalable annotation tooling. arXiv preprint arXiv:1805.04687, 2018.
- [Zeng et al.(2019)Zeng, Zhuge, Lu, and Zhang] Yu Zeng, Yunzhi Zhuge, Huchuan Lu, and Lihe Zhang. Joint learning of saliency detection and weakly supervised semantic segmentation. In CVPR, 2019.
- [Zhang et al.(2015)Zhang, Chen, Li, Wang, and Xia] Yu Zhang, Xiaowu Chen, Jia Li, Chen Wang, and Changqun Xia. Semantic object segmentation via detection in weakly labeled video. In CVPR, 2015.
- [Zhang et al.(2017)Zhang, Chen, Li, Wang, Xia, and Li] Yu Zhang, Xiaowu Chen, Jia Li, Chen Wang, Changqun Xia, and Jun Li. Semantic object segmentation in tagged videos via detection. TPAMI, 2017.
- [Zhao et al.(2017)Zhao, Shi, Qi, Wang, and Jia] Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. Pyramid scene parsing network. In CVPR, 2017.
- [Zhou et al.(2016)Zhou, Khosla, Lapedriza, Oliva, and Torralba] Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, and Antonio Torralba. Learning deep features for discriminative localization. In CVPR, 2016.