Learning Transferrable Knowledge for Semantic Segmentation with Deep Convolutional Neural Network

Learning Transferrable Knowledge for Semantic Segmentation
with Deep Convolutional Neural Network

Seunghoon Hong            Junhyuk Oh            Bohyung Han            Honglak Lee
Dept. of Computer Science and Engineering Dept. of Electrical Engineering and Computer Science POSTECH, Pohang, Korea University of Michigan, Ann Arbor, MI, USA {maga33,bhhan}@postech.ac.kr {junhyuk,honglak}@umich.edu

We propose a novel weakly-supervised semantic segmentation algorithm based on Deep Convolutional Neural Network (DCNN). Contrary to existing weakly-supervised approaches, our algorithm exploits auxiliary segmentation annotations available for different categories to guide segmentations on images with only image-level class labels. To make the segmentation knowledge transferrable across categories, we design a decoupled encoder-decoder architecture with attention model. In this architecture, the model generates spatial highlights of each category presented in an image using an attention model, and subsequently generates foreground segmentation for each highlighted region using decoder. Combining attention model, we show that the decoder trained with segmentation annotations in different categories can boost the performance of weakly-supervised semantic segmentation. The proposed algorithm demonstrates substantially improved performance compared to the state-of-the-art weakly-supervised techniques in challenging PASCAL VOC 2012 dataset when our model is trained with the annotations in 60 exclusive categories in Microsoft COCO dataset.

1 Introduction

Semantic segmentation refers to the task assigning dense class labels to each pixel in image. Although pixel-wise labels provide richer descriptions of images than bounding box labels or image-level tags, inferring such labels is a much more challenging task as it involves a highly complicated structured prediction problem.

Recent breakthrough in semantic segmentation has been mainly accelerated by the approaches based on Convolutional Neural Networks (CNNs) [21, 4, 11, 10, 25]. Given a classification network pre-trained on a large image collection, they learn a network for segmentation based on strong supervision—pixel-wise class labels. Although the approaches substantially improve the performance over the prior arts, training CNN requires a large number of fine-quality segmentation annotations, which are difficult to collect due to extensive labeling cost. For this reason, scaling up the semantic segmentation task to a large number of categories is very challenging in practice.

Weakly-supervised semantic segmentation [5, 31, 27, 29] is an alternative approach to alleviate annotation efforts in supervised methods. They infer latent segmentation labels from training images given weak annotations such as image-level class labels [31, 27, 29] or bounding boxes [5]. Since such annotations are easy to collect and even already available in existing datasets [6], it is straightforward to apply those approaches to large-scale problems with many categories. However, the segmentation quality by the weakly-supervised techniques is typically much worse than the one by supervised methods since there is no direct supervision for segmentation such as object shapes and locations during training.

The objective of this paper is to reduce the gap between semantic segmentation algorithms based on strong supervisions (e.g., semi- and fully-supervised approaches) and weak supervisions (e.g., weakly-supervised approaches). Our key idea is to employ segmentation annotations available for different categories to compensate for missing supervisions in weakly annotated images. No additional cost is required to collect such data since there are already several datasets publicly available with pixel-wise annotations, e.g., BSD [22], Microsoft COCO [20], and LabelMe [32]. These datasets have not been actively explored yet for semantic segmentation due to the mismatches in semantic categories with the popular benchmark datasets, e.g. PASCAL VOC [7]. The critical challenge in this problem is to learn common prior knowledge for segmentation transferrable across categories. It is not a trivial task with existing architectures, since they simply pose the semantic segmentation as pixel-wise classification and it is is difficult to exploit examples from the unseen classes.

We propose a novel encoder-decoder architecture with attention model, which is conceptually appropriate to transfer segmentation knowledge from one category to another. In this architecture, the attention model generates category-specific saliency on each location of an image, while the decoder performs foreground segmentation using the saliency map based on category-independent segmentation knowledge. Our model trained on one dataset is transferable to another by adapting the attention model to focus on unseen categories. Since the attention model is trainable with only image-level class labels, our algorithm is applicable to semantic segmentation on weakly-annotated images through transfer learning.

The contributions of this paper are summarized below.

  • We propose a new paradigm for weakly-supervised semantic segmentation, which exploits segmentation annotations from different categories to guide segmentations with weak annotations. To our knowledge, this is the first attempt to tackle the weakly-supervised semantic segmentation problem by transfer learning.

  • We propose a novel encoder-decoder architecture with attention model, which is appropriate to transfer the segmentation knowledge across categories.

  • The proposed algorithm achieved substantial performance improvement over existing weakly-supervised approaches with segmentation annotations in exclusive categories.

The rest of the paper is organized as follows. We briefly review related work and introduce our algorithm in Section 2 and 3, respectively. The detailed configuration of the proposed network is described in Section 4. Training and inference procedures are presented in Section 5. Section 6 illustrates experimental results on a benchmark dataset.

2 Related Work

Recent success in CNN has brought significant progress on semantic segmentation in the past few years [21, 4, 11, 10, 25]. By posing the semantic segmentation as region-based classification problem, they train the network to produce pixel-wise class labels using segmentation annotations as training data [21, 11, 10, 25]. Based on this framework, some approaches improve segmentation performance by learning deconvolution network to capture accurate object boundaries [26] or adopting fully connected CRF as post-processing [38, 4]. However, the performance of the supervised approaches depends heavily on the size and quality of training data, which limits the scalability of the algorithms.

To reduce the efforts for annotations, weakly-supervised approaches attempt to learn the model for semantic segmentation only with weak annotations [31, 27, 29, 5]. To infer latent segmentation labels, they often rely on the techniques such as Multiple Instance Learning (MIL) [31, 29] or Expectation-Maximization (EM) [27]. Unfortunately, they are not sufficient to make up missing supervision and lead to significant performance degradation compared to fully-supervised approaches. In the middle, semi-supervised approaches [27, 13] exploit a limited number of strong annotations to reduce performance gap between fully- and weakly-supervised approaches. Notably, [13] proposed a decoupled encoder-decoder architecture for segmentation, where it divides semantic segmentation into two separate problems—classification and segmentation—and learns a decoder to perform binary segmentation for each class identified in the encoder. Although this semi-supervised approach improves performance by sharing the decoder for all classes, it still needs strong annotations in the corresponding classes for segmentation. We avoid this problem by using segmentation annotations available for other categories.

In the field of computer vision, the idea of employing external data to improve performance of target task has been explored in context of domain adaptation [33, 15, 9, 8] or transfer learning [19, 36]. However, the approaches in domain adaptation often assume that there are shared categories across domains, and the techniques with transfer learning are often limited to simple tasks such as classification. We refer [30] for comprehensive surveys on domain adaptation and transfer learning. Hoffman et al[12] proposed a large-scale detection system by transferring knowledge for object detection between categories. Our work shares the motivations with this work, but aims to solve a highly complicated structured prediction problem, semantic segmentation.

There has been a long line of research on learning visual attention [18, 3, 1, 24, 2, 37, 35]. Their objective is to learn the attention mechanism that can adaptively focus on salient part of an image or video for various computer vision tasks, such as object recognition [18, 1, 2], object tracking [3], caption generation [37], image generation [35], etc. Our work is an extension of this idea to semantic segmentation by transfer learning.

3 Algorithm Overview

Figure 1: Overall architecture of the proposed algorithm. Given a feature extracted from the encoder, the attention model estimates adaptive spatial saliency of each category associated with input image (Section 4.2). The outputs of attention model are subsequently fed into the decoder, which generates foreground segmentation mask of each focused region (Section 4.3). During training, we fix the encoder by pre-trained weights, and leverage the segmentation annotations from source domain to train both the decoder and the attention model, and image-level class labels in both domains to train the attention model. After training, semantic segmentation on the target domain is performed naturally by exploiting the decoder trained with source images and the attention model adapted to target domain (Section 5).

This paper tackles the weakly-supervised semantic segmentation problem in transfer learning perspective. Suppose that we have two sources of data, and , which are composed of and images, respectively. Note that a set of images in target domain, denoted by , only have image-level class labels while the other set of data in source domain, referred to as , have pixel-wise segmentation annotations. Our objective is to improve the weakly-supervised semantic segmentation on the target domain using the segmentation annotations available in the source domain. We assume that both target and source domains are composed of exclusive sets of categories. In this setting, there is no direct supervision (i.e., ground-truth segmentation labels) for the categories in the target domain, which makes our objective similar to a weakly-supervised semantic segmentation setting.

To transfer segmentation knowledge from source to target domain, we propose a novel encoder-decoder architecture with attention model. Figure 1 illustrates the overall architecture of the proposed algorithm. The network is composed of three parts: encoder, decoder and attention model between the encoder and the decoder. In this architecture, the input image is first transformed to a multi-dimensional feature vector by the encoder, and the attention model identifies salient region for each category associated with the image. The output of the attention model reveals location information of each category in a coarse feature map, where the dense and detailed foreground segmentation mask for each category is subsequently obtained by the decoder.

Training our network involves different mechanisms for source and target domain examples, since they are associated with heterogeneous annotations with different levels of supervision. We leverage the segmentation annotations from source domain to train both the decoder and the attention model with segmentation objective, while image-level class labels in both target and source domains are used to train the attention model under classification objective. The training is performed jointly for both objectives using examples from both domains.

The proposed architecture exhibits several advantages to capture transferrable segmentation knowledge across domains. Employing the decoupled encoder-decoder architecture [13] makes it possible to share the information for shape generation among different categories. The attention model provides not only predictions for localization but also category-specific information that enables us to adapt the decoder trained in source domain to target domain. The combination of two components makes information for segmentation transferable across different categories, and provides useful segmentation prior that is missing in weakly annotated images in target domain.

4 Architecture

This section describes the architecture of the proposed algorithm, including the attention model and the decoder.

4.1 Preliminaries

We first describe notations and general configurations of the proposed model. Our network is composed of four parts, and , which are neural networks corresponding to encoder, attention model, classifier and decoder, respectively. Our objective is to train all components using the examples from both domains.

Let denotes a training image from either source or target domain. We assume that the image is associated with a set of class labels , which is given by either ground-truth (in training) or prediction (in testing). Given an input image , the network first extracts a feature descriptor as


where is the model parameter for the encoder, and and denote the number of hidden units in each channel and the number of channels, respectively. We employ VGG-16 layer net [34] pre-trained on ImageNet [6] as our encoder , and the feature descriptor is obtained from the last convolutional layer to retain spatial information in input image. The extracted feature and associated labels are then used to generate attention, which is discussed in the following subsection.

4.2 Attention model

(a) Input image
Figure 2: Examples of learned attentions. (a) Input image, (b), (c) and (d) represent attention weights obtained by Eq. (3). The proposed attention model adaptively focuses on different areas in an image depending on input labels.

Given a feature descriptor extracted from the encoder and its associated class labels , the objective of our attention model is to learn a set of positive weight vectors defined over a 2D space, where each element of represents the relevance of each feature location to category. Our attention model can be formally given by


where denotes a one-hot label vector for the category, and represents unnormalized attention weights. To encourage the model to pay attention to only a part of the image, we normalize to using a softmax function as suggested in [37].

To generate category-specific attention using our attention model , we employ multiplicative interactions [23] between feature and label vector. It learns a set of gating parameters represented by a 3-way tensor to model correlation between feature and label vectors. For scalability issue, we reduce the number of parameters by the factorization technique proposed in [23], and our model can be written as


where denotes element-wise multiplication and is bias. Note that the weights are given by and , where and denote the size of label vector and the number of factors, respectively. We observe that using multiplicative interaction generally gives better results than additive ones (e.g., concatenation), because it is capable of capturing high-order dependency between the feature and the label.

To apply the attention to our transfer-learning scenario, the model should be trainable in both target and source domains. Since examples from the target domain are associated with only image-level class labels, we put additional layers on top of the attention model, and optimize both and under classification objective. To this end, we extract features based on the category-specific attention by aggregating features over the spatial region as follows:


Intuitively, represents a category-specific feature defined over all the channels in the feature map.

Using the weak annotations associated with both target and source domain images, our attention model is trained to minimize the classification loss as follows:


where denote the loss between ground-truth and predicted label vector , respectively, which are both defined for single label and aggregated over for each image. We employed softmax loss function to measure classification loss .

The optimization of Eq. (6) may involve potential overfitting issue, since the ground-truth class label vector is given as an input to attention model as well. In practice, we observe that our model can avoid this issue by effectively eliminating the direct link from attention to label prediction and constructing intermediate representation based on the original feature .

Figure 2 illustrates the learned attention weights for individual classes. We observe that the attention model adapts spatial saliency effectively depending on its input labels.

4.3 Decoder

The attention model described in the previous section generates a set of adaptive saliency maps for each category , which provides useful information for localization. Given these attentions, the next step of our algorithm is to reconstruct dense foreground segmentation mask for each attended category by the decoder. However, the direct application of attentions weights to segmentation may be problematic, since the activations tend to be sparse due to the softmax operation in Eq. (3) and may lose information encoded in the feature map useful for shape generation.

To resolve this issue and reconstruct useful information for segmentation, we feed the additional inputs to the decoder using attention and the original feature . Rather then directly using the attention, we exploit the intermediate representation obtained from Eq. (5). It represents relevance of each channel out of the feature maps with respect to category. Then we aggregate spatial activations in each channel of the feature using as coefficients, which is given by


where represents densified attention in the same size with and serves as inputs to the decoder. As shown in Figure 3, densified attention maps preserve more details of the object shape compared to the original attention (). See Appendix A for more comprehensive analysis of Eq. (7).

Given densified attention as input, the decoder is trained to minimize the segmentation loss by the following objective function


where denotes a binary segmentation mask of image for category, and denotes pixel-wise loss function between ground-truth and predicted segmentation mask. Similar to classification, we employ a softmax loss function for . Since training requires ground-truth segmentation annotations, the objective function is only optimized with images in source domain.

We employ recently proposed deconvolution network [26] for our decoder architecture . Given an input to the decoder , it generates a segmentation mask in the same size with input image by multiple successive operations of unpooling, deconvolution and rectification. Pooling switches are shared between pooling and unpooling layers, which is appropriate to recover accurate object boundary. We refer to [26] for more details.

Note that we train our decoder to generate foreground segmentation of each attention . By decoupling classification, which is a domain specific task, from the decoding [13], we can capture category-independent information for shape generation and apply the architecture to any unseen categories. Since all weights in the decoder are shared between different categories, it potentially encourages the decoder to capture common shape information that can be generally applicable to multiple categories.

(a) Input image
(b) Attention
(c) Densified attention
Figure 3: Examples of attention () and densified attention ().

5 Training and Inference

Combining Eq. (6) and (8), the overall objective function is given by


where controls balance between classification and segmentation. During training, we optimize Eq. (9) using examples from both domains. Note that it allows joint optimization of attention model for both classification and segmentation. Although our attention model is generally good even trained with only class labels (see Figure 3), training attention based only on classification objective sometimes lead to noisy predictions due to missing supervision for localization. By jointly training with segmentation objective, we can regularize it to avoid finding noisy solution for target domain categories. After training, we remove the classification layers since it is required only in training to learn attentions for the data from target domain categories.

For inference of target domain images with the trained model, we first apply a separate classifier to identify a set of labels associated with the image. Then for each identified label , we iteratively construct attention weights and obtain foreground segmentation mask from the decoder output. Given foreground probability maps from all labels , the final segmentation label is obtained by taking the maximum probability in channel direction.

6 Experiments

This section describes detailed information about implementation and experiment, and provides results in a challenging benchmark dataset.

Method bkg aero bike bird boat bottle bus car cat chair cow table dog horse mbk person plant sheep sofa train tv mean
EM-Adapt [27] 67.2 29.2 17.6 28.6 22.2 29.6 47.0 44.0 44.2 14.6 35.1 24.9 41.0 34.8 41.6 32.1 24.8 37.4 24.0 38.1 31.6 33.8
CCNN [28] 68.5 25.5 18.0 25.4 20.2 36.3 46.8 47.1 48.0 15.8 37.9 21.0 44.5 34.5 46.2 40.7 30.4 36.3 22.2 38.8 36.9 35.3
MIL+seg [31] 79.6 50.2 21.6 40.9 34.9 40.5 45.9 51.5 60.6 12.6 51.2 11.6 56.8 52.9 44.8 42.7 31.2 55.4 21.5 38.8 36.9 42.0
DecoupledNet [13] 86.5 69.9 33.6 58.5 42.4 50.4 68.8 63.2 67.5 11.5 61.8 20.0 61.2 66.7 60.1 50.8 30.2 67.9 33.9 59.2 51.0 53.1
EM-Adapt [27] - - - - - - - - - - - - - - - - - - - - - 47.6
TransferNet 85.3 68.5 26.4 69.8 36.7 49.1 68.4 55.8 77.3 6.2 75.2 14.3 69.8 71.5 61.1 31.9 25.5 74.6 33.8 49.6 43.7 52.1
TransferNet-GT 85.2 70.6 25.3 61.7 42.2 38.9 67.5 53.9 73.3 20.6 81.5 26.9 69.6 73.2 66.6 36.7 26.9 82.9 42.2 54.4 39.3 54.3
DecoupledNet 79.2 13.1 7.7 38.4 14.3 15.0 14.7 46.0 60.5 3.7 28.0 1.7 54.0 37.5 24.0 9.2 4.5 46.2 3.4 18.7 13.0 25.4
BaselineNet 81.4 30.6 9.2 41.8 27.0 32.9 46.9 44.7 61.2 7.0 59.2 4.7 55.2 55.5 22.7 32.0 17.5 65.7 15.8 33.6 18.1 36.3
Table 1: Evaluation results on PASCAL VOC 2012 validation set.
Method bkg aero bike bird boat bottle bus car cat chair cow table dog horse mbk person plant sheep sofa train tv mean
FCN-8s [21] 91.2 76.8 34.2 68.9 49.4 60.3 75.3 74.7 77.6 21.4 62.5 46.8 71.8 63.9 76.5 73.9 45.2 72.4 37.4 70.9 55.1 62.2
CRF-RNN [38] 93.1 90.4 55.3 88.7 68.4 69.8 88.3 82.4 85.1 32.6 78.5 64.4 79.6 81.9 86.4 81.8 58.6 82.4 53.5 77.4 70.1 74.7
DeepLab-CRF [4] 93.1 84.4 54.5 81.5 63.6 65.9 85.1 79.1 83.4 30.7 74.1 59.8 79.0 76.1 83.2 80.8 59.7 82.2 50.4 73.1 63.7 71.6
DeconvNet [26] 93.1 89.9 39.3 79.7 63.9 68.2 87.4 81.2 86.1 28.5 77.0 62.0 79.0 80.3 83.6 80.2 58.8 83.4 54.3 80.7 65.0 72.5
EM-Adapt [27] 76.3 37.1 21.9 41.6 26.1 38.5 50.8 44.9 48.9 16.7 40.8 29.4 47.1 45.8 54.8 28.2 30.0 44.0 29.2 34.3 46.0 39.6
CCNN [28] 70.1 24.2 19.9 26.3 18.6 38.1 51.7 42.9 48.2 15.6 37.2 18.3 43.0 38.2 52.2 40.0 33.8 36.0 21.6 33.4 38.3 35.6
MIL+seg [31] 78.7 48.0 21.2 31.1 28.4 35.1 51.4 55.5 52.8 7.8 56.2 19.9 53.8 50.3 40.0 38.6 27.8 51.8 24.7 33.3 46.3 40.6
TransferNet 85.7 70.1 27.8 73.7 37.3 44.8 71.4 53.8 73.0 6.7 62.9 12.4 68.4 73.7 65.9 27.9 23.5 72.3 38.9 45.9 39.2 51.2
Table 2: Evaluation results on PASCAL VOC 2012 test set.

6.1 Implementation Details


We employ PASCAL VOC 2012 [7] as target domain and Microsoft COCO (MS-COCO) [20] as source domain, which have 20 and 80 labeled semantic categories, respectively. To simulate the transfer learning scenario, we remove all training images containing 20 PASCAL categories from MS-COCO dataset, and use only 17,443 images from 60 categories (with no overlap with the PASCAL categories) to construct the source domain data. We train our model using image-level class labels in PASCAL VOC dataset and segmentation annotations in MS-COCO dataset, respectively, and evaluate the performance on PASCAL VOC 2012 benchmark images.


We initialize the encoder by fine-tuning the pre-trained CNN from ImageNet [6] to perform multi-class classification on combined datasets of PASCAL VOC and MS-COCO. The weights in the attention model and classification layers ( and , respectively) are pre-trained by optimizing Eq. (6). Then we optimize both decoder, attention model and classification layers jointly using the objective function in Eq. (9), while the weights in the decoder () are initialized with zero-mean Gaussians. We fix the weights in the encoder () during training.


We implement the proposed algorithm based on Caffe [14] library. We employ Adam optimization [16] to train our network with learning rate 0.0005 and default hyper-parameter values proposed in [16]. The size of mini-batch is set to 64. We trained our models using a NVIDIA Titan X GPU. Training our model takes 4 hours for the pre-training of attention model including classification layers, and 10 hours for the joint training of all other parts, respectively.


We exploit an additional classifier trained on PASCAL VOC dataset to predict class labels on target domain images. The predicted class labels are used to generate segmentations as described in Section 5. Optionally, we employ post processing based on fully-connected CRF [17]. In this case, we apply the CRF on foreground/background probability maps for each class label independently, and obtain combined segmentations by taking pixel-wise maximums of foreground probabilities across labels.

6.2 Comparison to Other Methods

This section presents evaluation results of our algorithm with the competitors on PASCAL VOC 2012 benchmark dataset. We follow comp6 evaluation protocol, and scores are measured by computing Intersection over Union (IoU) between ground truth and predicted segmentations.

Table 1 summarizes the evaluation results on PASCAL VOC 2012 validation dataset. We compared the proposed algorithm with state-of-the-art weakly- and semi-supervised algorithms111Strictly speaking, our method is not directly comparable to both approaches since we use auxiliary examples. Note that we do not use ground-truth segmentation annotations for the categories used in evaluation, since the examples are from different categories. . Our method is denoted by TransferNet, and TransferNet-GT indicates our method with ground-truth class labels for segmentation inference, which serves as the upper-bound performance of our method since it assumes classification is perfect. The proposed algorithm outperforms all weakly-supervised semantic segmentation techniques with substantial margins, although it does not employ any ground-truth segmentations for categories used in evaluation. The performance of the proposed algorithm is comparable to semi-supervised semantic segmentation methods, which exploits a small number of ground-truth segmentations in addition to weakly-annotated images for training. The results suggest that segmentation annotations from different categories can be used to make up the missing supervision in weakly-annotated images; the proposed encoder-decoder architecture based on attention model successfully captures transferable segmentation knowledge from the exclusive segmentation annotations and uses it as prior for segmentation in unseen categories.

Table 2 summarizes our results on PASCAL VOC 2012 test dataset. Our algorithm exhibits superior performance to weakly-supervised approaches, but there are still large performance gaps with respect to fully-supervised approaches. It shows that there is domain-specific segmentation knowledge which cannot be made up by annotations form different categories.

The qualitative results of the proposed algorithm are presented in Figure 5. Our algorithm often produces accurate segmentations in the target domain by transferring the decoder trained with source domain examples, although it is not successful in capturing some category-specific fine details in some examples. The missing details can be recovered through post-processing based on CRF. Since the attention model in the target domain may not be perfect due to missing supervisions, our algorithm sometimes produces noisy predictions as illustrated in Figure 5(b). See Apendix C for more qualitative comparisons.

6.3 Comparison to Baselines

To better understand the benefits from the attention model in our transfer learning scenario, we compare the proposed algorithm with two baseline algorithms, which are denoted by DecoupledNet and BaselineNet. (See Appendix B for detailed configurations of the baselines and proposed algorithm.)

DecoupledNet directly applies the architecture proposed in [13] to our transfer learning task by training the decoder in MS-COCO and applying it to PASCAL VOC for inference. The model employs the same decoupled decoder architecture to ours, but has a direct connection between encoder and decoder without attention mechanism. The result in Table 1 shows that the model trained on source domain fails to adapt to target domain categories. It is mainly because the decoder cannot interpret the features from unseen categories in target domain. Our model can mitigate the issue since the attention model provides coherent representations to decoder across domains.

Although the above baseline shows the benefits of the attention model in our architecture, the advantage of attention estimation from the intermediate layer may be still ambiguous. To this end, we employ another FCN [21]-style baseline denoted by BaselineNet, which uses class score map as input to the decoder. It can be considered as a specific case of our method that the attention is extracted from the final layer of the classification network (). The performance of BaselineNet is better than DecoupledNet since the class score map is a general representation to the decoder across different categories. However, the performance is considerably worse than the proposed method as shown in Table 1 and Figure 5. We observe that the class score map is sparse and in row-resolution, while densified attention map in our model contains richer information for segmentation.

The comparisons to the baseline algorithms show that transferring segmentation knowledge from different categories is a very challenging task. The straightforward extensions of existing architectures have difficulty in generalizing the knowledge across different categories. In contrast, our model effectively transfers segmentation knowledge by learning general features through attention mechanism.

Figure 4: Performance of the proposed algorithm with varying number of annotations in the source domain.

6.4 Impact of Number of Annotations in the Source Domain

To see the impact of number of annotations in the source domain, we conduct additional experiments by varying the number of annotations in the source domain (MS-COCO). To this end, we randomly construct subsets of training data by varying their sizes in ratios (50%, 25%, 10%, 5% and 1%) and average the performance in each size with 3 subsets. The results are illustrated in Figure 4. In general, more annotations in the source domain improve the segmentation quality on the target domain. Interestingly, the performance of the proposed algorithm is still better than other weakly-supervised methods even with a very small fraction of annotations. It suggests that exploiting even small number of segmentations from other categories can effectively reduce the gap between the approaches based on strong and weak supervisions.

(a) Examples that our method produces accurate segmentation.
(b) Examples that our method produces inaccurate segmentation due to misclassification (top) or inaccurate attention (bottom).
Figure 5: Examples of semantic segmentation on PASCAL VOC 2012 validation images. The attentions (3rd column) are extracted from the model trained using Eq. (9), and aggregated over all categories for visualization. (a) Our methods based on attention model (TransferNet and TransferNet+CRF) produce accurate segmentation result even without CRF by transferring learned segmentation knowledge from source domain. Our results tend to be denser and more accurate than the results from BaselineNet, which generates segmentation from class score map. (b) Our algorithm may produce inaccurate segmentation when the input labels are wrong due to misclassification (top) or attention output is noisy (bottom).

7 Conclusion

We propose a novel approach for weakly-supervised semantic segmentation, which exploits extra segmentation annotations in different categories to improve segmentation on the dataset with missing supervisions. The proposed encoder-decoder architecture with attention model is appropriate to capture transferable segmentation knowledge across categories. The results on challenging benchmark dataset suggest that the gap between approaches based on strong and weak supervision can be reduced by transfer learning. We believe that scaling up the proposed algorithm to a large number of categories would be one of interesting future research direction (e.g., semantic segmentation on 7.6K categories in ImageNet dataset using segmentation annotations from 20 PASCAL VOC categories).


  • [1] B. Alexe, N. Heess, Y. W. Teh, and V. Ferrari. Searching for objects driven by context. In NIPS. 2012.
  • [2] J. Ba, V. Mnih, and K. Kavukcuoglu. Multiple object recognition with visual attention. In ICRL, 2015.
  • [3] L. Bazzani, N. de Freitas, H. Larochelle, V. Murino, and J.-A. Ting. Learning attentional policies for object tracking and recognition in video with deep networks. In ICML, 2011.
  • [4] L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille. Semantic image segmentation with deep convolutional nets and fully connected CRFs. In ICLR, 2015.
  • [5] J. Dai, K. He, and J. Sun. BoxSup: Exploiting bounding boxes to supervise convolutional networks for semantic segmentation. arXiv preprint arXiv:1503.01640, 2015.
  • [6] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. Imagenet: A large-scale hierarchical image database. In CVPR, 2009.
  • [7] M. Everingham, L. Van Gool, C. K. Williams, J. Winn, and A. Zisserman. The pascal visual object classes (voc) challenge. IJCV, 88(2):303–338, 2010.
  • [8] Y. Ganin and V. Lempitsky. Unsupervised domain adaptation by backpropagation. In ICML, 2015.
  • [9] R. Gopalan, R. Li, and R. Chellappa. Domain adaptation for object recognition: An unsupervised approach. In ICCV, 2011.
  • [10] B. Hariharan, P. Arbeláez, R. Girshick, and J. Malik. Simultaneous detection and segmentation. In ECCV, 2014.
  • [11] B. Hariharan, P. Arbeláez, R. Girshick, and J. Malik. Hypercolumns for object segmentation and fine-grained localization. In CVPR, 2015.
  • [12] J. Hoffman, S. Guadarrama, E. S. Tzeng, R. Hu, J. Donahue, R. Girshick, T. Darrell, and K. Saenko. Lsda: Large scale detection through adaptation. In NIPS. 2014.
  • [13] S. Hong, H. Noh, and B. Han. Decoupled deep neural network for semi-supervised semantic segmentation. In NIPS, 2015.
  • [14] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell. Caffe: Convolutional architecture for fast feature embedding. In MM, pages 675–678. ACM, 2014.
  • [15] A. Khosla, T. Zhou, T. Malisiewicz, A. Efros, and A. Torralba. Undoing the damage of dataset bias. In ECCV, 2012.
  • [16] D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. In ICRL, 2015.
  • [17] P. Krähenbühl and V. Koltun. Efficient inference in fully connected crfs with gaussian edge potentials. In NIPS, 2011.
  • [18] H. Larochelle and G. E. Hinton. Learning to combine foveal glimpses with a third-order boltzmann machine. In NIPS. 2010.
  • [19] J. J. Lim, R. Salakhutdinov, and A. Torralba. Transfer learning by borrowing examples for multiclass object detection. In NIPS, 2011.
  • [20] T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick. Microsoft coco: Common objects in context. In ECCV. 2014.
  • [21] J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In CVPR, 2015.
  • [22] D. Martin, C. Fowlkes, D. Tal, and J. Malik. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In ICCV, 2001.
  • [23] R. Memisevic. Learning to relate images. TPAMI, 2013.
  • [24] V. Mnih, N. Heess, A. Graves, and K. Kavukcuoglu. Recurrent models of visual attention. In NIPS, 2014.
  • [25] M. Mostajabi, P. Yadollahpour, and G. Shakhnarovich. Feedforward semantic segmentation with zoom-out features. CVPR, 2015.
  • [26] H. Noh, S. Hong, and B. Han. Learning deconvolution network for semantic segmentation. In ICCV, 2015.
  • [27] G. Papandreou, L.-C. Chen, K. Murphy, and A. L. Yuille. Weakly-and semi-supervised learning of a DCNN for semantic image segmentation. arXiv preprint arXiv:1502.02734, 2015.
  • [28] D. Pathak, P. Krähenbühl, and T. Darrell. Constrained convolutional neural networks for weakly supervised segmentation. In ICCV, 2015.
  • [29] D. Pathak, E. Shelhamer, J. Long, and T. Darrell. Fully convolutional multi-class multiple instance learning. arXiv preprint arXiv:1412.7144, 2014.
  • [30] N. Patricia and B. Caputo. Learning to learn, from transfer learning to domain adaptation: A unifying perspective. In CVPR, 2014.
  • [31] P. O. Pinheiro and R. Collobert. From image-level to pixel-level labeling with convolutional networks. In CVPR, 2015.
  • [32] B. Russell, A. Torralba, K. Murphy, and W. T. Freeman. Labelme: a database and web-based tool for image annotation. IJCV, 2010.
  • [33] K. Saenko, B. Kulis, M. Fritz, and T. Darrell. Adapting visual category models to new domains. In ECCV, 2010.
  • [34] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In ICLR, 2015.
  • [35] Y. Tang, N. Srivastava, and R. R. Salakhutdinov. Learning generative models with visual attention. In NIPS. 2014.
  • [36] T. Tommasi, F. Orabona, and B. Caputo. Learning categories from few examples with multi model knowledge transfer. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 36:928–941, 2014.
  • [37] K. Xu, J. Ba, R. Kiros, K. Cho, A. C. Courville, R. Salakhutdinov, R. S. Zemel, and Y. Bengio. Show, attend and tell: Neural image caption generation with visual attention. In ICML, 2015.
  • [38] S. Zheng, S. Jayasumana, B. Romera-Paredes, V. Vineet, Z. Su, D. Du, C. Huang, and P. Torr. Conditional random fields as recurrent neural networks. In ICCV, 2015.


Appendix A Analysis of Densified Attention

This section describes more comprehensive analysis of the densified attention discussed in Section 4.3. By plugging Eq. (5) into Eq. (7), we obtain the densified attention, which is given by


where is the Gram matrix of ; each element in denoted by represents similarity between and pixels in the feature map. Eq. (10) means that the densified attention is given by the weighted linear combination of rows in the Gram matrix, where the weights are given by the attention . The densified attention reveals more detailed shape of the objects than the sparse attention weight by highlighting not only attended pixels but also visually correlated areas.

Figure 6 visualizes the densified attention obtained by selecting a row of the Gram matrix in Eq. (10). The row is given by the one-hot attention vector , which represents a single pixel attention in the feature map. We observe that it successfully generates dense activation maps based only on extremely sparse attentions using the correlation of pixels. It also suggests that using the densified attention as an input to the decoder is more useful to generate accurate and dense segmentation mask than using the original sparse attention.

Appendix B Comparisons to Baseline Architectures

Figure 7 illustrates the detailed configurations of the algorithms used in Section 6.3—the proposed algorithm and two baselines denoted by BaselineNet and DecoupledNet. The baseline algorithms are straightforward extensions of existing CNNs for semantic segmentation, which are designed for transfer learning without attention mechanism. They share the same encoder and decoder architectures with the proposed model but have different approaches to construct the input to the decoder .

BaselineNet (Figure 7(b)) is an extension of FCN [21]-style architecture for our transfer learning scenario. Given an input image, it generates a set of 2D class-specific score maps retaining spatial information based on fully-convolutional network [21]. Then for each class presented in the image, it extracts a two-dimensional () score map of the selected class, and converts it to the input to the decoder through a fully-connected layer. Based on this architecture, transfer learning can be achieved by training the decoder in source domain, and fine-tuning the classification network with target domain categories.

DecoupledNet (Figure 7(c)) is an extension of the decoupled encoder-decoder architecture proposed in [13] for transfer learning scenario. Given an input image, it first identifies categories presented in the image using the outputs of the classification network, and subsequently generates foreground segmentation of each identified category by the decoder. To this end, it computes gradient of class score with respect to the feature map by back-propagation, and constructs input to the decoder by combining the feature and gradient maps using a few feed-forward layers. In this architecture, transfer learning can be achieved by the same way to the BaselineNet; training decoder in source domain, and fine-tuning the classification network with target domain categories.

Table 1 presents comparisons of the proposed algorithm to the baseline architectures. DecoupledNet directly exploits feature and gradient maps to construct input to the decoder. Since the both representations are domain-specific and changed by fine-tuning the classification network in different domains, the decoder trained on the source domain is difficult to be generalized to unseen categories in the target domain. Our architecture alleviates this problem using attention, where the attention model provides coherent representation to the decoder across domains. Compared to BaselineNet, the proposed architecture is more effective to reconstruct crucial information required for segmentation since it employs the attention as well as the original features; it is more appropriate to accomplish more accurate segmentation.

The poor performance of the baseline architectures suggests that transferring segmentation knowledge across domain is a very challenging task, and naive extensions of the existing architectures may not be able to handle this challenge effectively. The proposed architecture based on attention mechanism is appropriate to transfer the decoder across domains and obtain accurate segmentation.

Figure 6: Visualizations of selected rows in the Gram matrix. (Top row) input image. (Bottom row) visualizations of rows in the Gram matrix which are selected from the pixel location marked on the image. Each row is reshaped to original spatial shape of the feature map for effective visualization.
Figure 7: Detailed configurations of the architectures presented in Section 6.3.

Appendix C Additional Results

Figure 8 presents qualitative results of the state-of-the-art weakly-supervised semantic segmentation techniques [27, 28] including our algorithm on PASCAL VOC 2012 validation images. The compared algorithms adopt post-processing based on fully-connected CRF [17] to refine segmentation results. Nevertheless, the segmentation of other methods is not successful frequently since the output predictions of the CNN are often too noisy and inaccurate to capture precise object shapes. Our approach tends to find more accurate object boundaries even without CRF by exploiting the decoder trained with segmentation annotations in different categories, and the results exhibit distinguishing performance compared to existing weakly-supervised approaches.

Input Image         Ground-truth        EM-Adapt [27]           CCNN [28]           TransferNet        TransferNet+CRF

Figure 8: Comparisons of semantic segmentation results on PASCAL VOC 2012 validation images. The proposed algorithm tends to find more accurate object boundary compared to other weakly-supervised approaches even without CRF.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
Add comment
Loading ...
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test description