Rethinking Convolutional Semantic Segmentation Learning
Deep convolutional semantic segmentation (DCSS) learning doesn’t converge to an optimal local minimum with random parameters initializations; a pre-trained model on the same domain becomes necessary to achieve convergence.In this work, we propose a joint cooperative end-to-end learning method for DCSS. It addresses many drawbacks with existing deep semantic segmentation learning; the proposed approach simultaneously learn both segmentation and classification; taking away the essential need of the pre-trained model for learning convergence. We present an improved inception based architecture with partial attention gating (PAG) over encoder information. The PAG also adds to achieve faster convergence and better accuracy for segmentation task. We will show the effectiveness of this learning on a diabetic retinopathy classification and segmentation dataset.
s O m \IfBooleanTF#1∥#3∥#2∥#3#2∥_L_2(Ω)
Semantic segmentation task using deep learning is formulated as a pixel-wise classification problem [15, 10, 16, 2, 4]. Generally, it’s an encoder-decoder structure, one encoder to encode the features to lower dimensional features map and one decoder to remap the features map into several probabilistic maps (number of classes) of same width and height as input using bilinear interpolation or transposed convolutions [5, 6]. End-to-end learning of the encoder-decoder model best achieved using a pre-trained classification model for the encoder on the same domain. However, for a simple problem with low spatial resolutions input size and fewer number of classes pre-trained model isn’t necessary, though it helps speed up the learning process. For complex problems with many numbers of classes and higher resolutions, learning step doesn’t converge to the best local optimum without a pre-trained model; leading to inferior validation mean IOU(mean pixel intersection over union). For problems with completely unrelated domain such as medical images, satellite images etc. segmentation becomes two unrelated step; first, define and train a classification model and then use the same for the encoder to train the encoder-decoder model.
How do we simplify the segmentation learning for any domains and complexity? Can we formulate the two-step processes into one step; at the same time learning both steps? We can formulate this as high-level multi-tasks learning concept to train the segmentation system. The idea of pure multi-tasks learning deviates from the main principle behind the simplification to learn one task from the feedback of another task. Here in our case segmentation is dependent on the pre-trained model, achieved using the classification step, which is completely independent of segmentation. In general, all tasks of a usual multi-tasks learning model can be learned independently; but here segmentation can’t be learned independently. Another aspect of this attempt is that the classification step should be trainable on loose labels (Section II-C) of the dataset.
Addressing the above concerns, in this work, we define a fully end-to-end learning system, that learns classification and segmentation jointly. We will show the model performance on a diabetic retinopathy (DR) classification and features segmentation dataset.
End-to-end joint segmentation and classification learning: In contrast to the existing semantic segmentation models, the proposed learning system doesn’t require a pre-trained model; it learns from the feedback of the classification node. Advantages of a joint segmentation learning system?
a. The scarcity of labeled segmentation data; this method can achieve better accuracy with the fewer number of segmentation ground-truths.
b. Eliminating the need for pre-trained model requirements for convergence.
c. Training complex model for any domains with larger input size.
d. Learning a classification model as by-product alongside.
Improved Residual Inception We have proposed a new improved residual inception block.
Partial Attention: A partial attention mechanism to use the mid-level encoder layer features for decoder layers.
Training with partial random labels: Learning method that can learn segmentation labeling while cooperatively training a classifier agent on partial random labels.
I-B Related Work
In this section, we will quickly go through some of the early and recent works on semantic segmentation using deep learning. Several recent works using encoder-decoder structure focus on improving the mean IOU for semantic segmentation benchmarks such as PASACAL-VOC, MS-COCO and biomedical image datasets. One of the first success using the convolutional net for semantic segmentation was achieved in FCN; where fully connected layers were replaced using unit strided convolution and upsampling layers initialized with simple bilinear interpolation were used to reconstruct the feature map. In DeconvNet, a multilayer deconvolution and unpooling network were used to predict the segmentation masks. Segnet decoder uses encoder max-pooling indices to upsample low-resolution feature maps without learning the upsampling layers. Also, they have used convolution layer for each upsampling layer to get dense feature map. Recently, object detection and semantic segmentation were jointly learned in Mask R-CNN, they have used FCN to predict segmentation mask for each region of interest. Also several other recent methods such as PSPNet, RefineNet, ENet, Sharpmask etc. demonstrated the advantages of encoder-decoder model using convolution and transposed convolution for the task of semantic segmentation. In general, the predicted segmentation mask is coarse and requires post-processing to remove outliers. DenseCRF is one effective post-processing layer for semantic segmentation [17, 18, 19], it refines the segmentation masks exploiting the pixel-level pairwise closeness.
Ii-a Cooperative Learning
We formulate the segmentation learning as a cooperative learning problem, where two agents operate on two overlapping domains and interact with each other to achieve their respective goals. The terms overlapping domains implies that the datasets for classification and segmentation are sampled from the same distribution; it’s not necessary to have an identical sample for both agents. One agent learns to generate segmentation masks for images with the help of the agent responsible for classification. Even though the classifier agent is capable of self-learning, the interaction with the segmentation agent results in improved generalizability. Each of the agents shares a chunk of parameters but uses separate optimizers to learn their parameters. The shared parameters of the agents are learned by both of the agents at a different rate; this is where the agents share knowledge and learn cooperatively. Learning rate for classifier agent is higher than that of segmentation agent as it possesses class specific discriminative knowledge about the domain. Outputs of the two agents Eq 1 for some input is calculated using the shared parameters ; denotes the output from the shared model part, denotes the output of the classifier agent with private parameters and denotes the output of the segmentation agent with private parameters .
Ii-B Network Architecture
Residual Inception module We have used a modified version of the base inception block proposed in , where an additional convolution was used before max-pooling to ensure rich feature representation. We have added a residual connection from the block input to output as shown in Fig 2. The concatenated features were convolved with another convolution to reduce aliasing and added with convolved input features; both operations have the same number of output filters. Apart from that, we have used a convolution factorization version Fig 3 of the same module for the classifier node. The idea of convolution factorization was proposed in , it reduces the number of parameters without effecting the generalization performance.
Partial attention We use partial attention over the selected encoder feature maps. Feeding the encoder information with upsampling boost the reconstructed feature map representation power. The attention mechanism here only focuses on the encoder feature maps with spatial resolutions equal or larger than that of the upsampling outputs. Max-pooling to reduce spatial resolution of encoder feature maps to the upsampling output size; max-pooling ensures maximally activated features. Fig 4 shows an overview of the partial attention mechanism used, where are the last feature maps (before reducing the feature map width and height for each block) of each convolution block of the encoder. The switch is on if the corresponding ; is a max-pooling operation for all with stride and identity operation for .
Cooperative learning model follows the compositional design principles stated in ; strided convolution is used to down-sample the encoder spatial resolution. To achieve richer feature representation, residual inception block as shown in Fig 2 is used for the encoder.
The encoder last layer output is used as the input of both the classifier and the segmentation agents. Classifier agent private node is designed using the residual factorized inception block as shown in Fig 3, followed by a global pool layer and a fully connected layer with a number of classes for the classifier. For each spatial dimension level of the encoder and the classifier private node, multiples cascaded residual inception blocks were used. Fig 5 shows the complete design of the joint segmentation and classification learning.
The decoder is designed using consecutive transposed convolutions with stride and kernel size followed by a partial attention block to incorporate encoder information for recovering object information. The partial attention block attend over few selective encoder feature maps with a spatial dimension larger or equal than that of the upsampled decoder output. Encoder feature maps learn input object information at multiple abstractions; the attention mechanism enables faster convergence of decoder parameters with improved generalization performance.
Ii-C Training on partial random labels
The cooperative learning model can also be used when the number of labeled samples for the classifier agent is small. The classifier agent trained on partial random labels doesn’t downgrade the segmentation agent performance. When we use around 20% of labeled samples for classifier and around 80% samples with random labels accuracy of classifier agent are lowered without effecting the performance of the segmentation agent. This observation indicates the ability of both of the agents to learn cooperatively. Also, this helps towards the goal of training on a dataset with the very low number of labeled images; especially relevant on medical image domains.
Iii Experiment Setup
Dataset: We train the classification agent on the Eyepacs diabetic retinopathy classification dataset  with 5 classes. The 5 classes of this dataset indicate DR severity level; level 0 for no DR, level 1 for mild DR, level 2 for moderate DR, level 3 for severe DR and 4 for proliferative DR. For training 25K images were used and validation performance was calculated on a set of 10k images. For testing, we have used 55K test set provided by eyepacs. The test set was labeled internally by ophthalmologists. Each image was resized to .
The segmentation agent is trained on a dataset prepared with the help of ophthalmologists; We selected a subset of around 5K images from Eyepacs dataset and annotated retinal pathological features with mask-based annotation. This dataset contains the annotation of around 5 pathological features (hard exudates, soft exudates, hemorrhage, microaneurysms, drusen) with one additional mask annotation for retinal artifacts. The set is split into training and validation set; 3.5K images for training and 1.5K images were used for validation.
Preprocessing: For both the classifier and the segmentation, is used as the input size. All input images were resized into and a randomly cropped patch of size is used as input. Extensive data augmentation is used such as random flip left/right, up/down and random padding was used for both classification and segmentation. Segmentation masks were also processed with same augmentation its corresponding image. Additionally, for classification additional augmentation such as for as changing the image pixel values randomly using hue, contrast and saturation were used. Also, each image is standardized by it’s mean and dividing its standard deviation.
Training procedure: Batch normalization is used to reduce covariate shift and achieve faster convergence of both agents. Also, we have used Nesterov momentum optimizer with polynomial learning rate policy. For the classifier agent, the learning rate is 30 times than that of the segmentation agent. The batch size used for the classifier agent is 32 times than that of the segmentation agent. The use of higher batch size is to facilitate faster learning and information transfer from the classifier agent to the segmentation agent. We use lower learning rate for the segmentation agent to make sure that the information it passes to the classifier agent is minimal, but aid to generalization.
|Original Image||Ground truth||Unet_inception||JointSeg[This work]|
Iii-a Results Analysis
We have evaluated out model performance on the eyepacs test set of 55k images. We have also trained VGG-19 and inceptionv3 to compare classification agent performance with this proposed approach. Also for the segmentation agent performance comparisons we have trained FCN-8s and UNet on the segmentation set. Here we use the improved inception block for UNet training; which is an improved version of the original UNet design.
Table I shows the mean-IOU comparisons with the proposed approach and the existing algorithms. A significant performance gain is obtained with join segmentation learning. Here fcn-8s and Unet models parameters were initialized with a pre-trained model; trained on eyepacs classification set.
|JointSeg [This work]||66.1|
If we randomly initialize the FCN-8s and UNet model parameters; results become non-informatory as shown in Table II; both FCN-8s and UNet could not achieve convergence to an optimal local minimum. From Table I and Table II we can understand the limitations and the importance of pre-trained model for traditional segmentation learning. But the joint learning achieves convergence to an optimal local minimum with random initializations.
|JointSeg [This work]||66.1|
Table III shows the performance of the classifier agent with standard deep learning classifier model.
|Method||Top-1 (%)||Top-2 (%)|
|JointSeg [This work]||85.1||88.1|
Table IV shows the classifier agent performance on a binary classification scenario. For a DR screening system to understand whether a person has DR, a binary label is enough. For this test, we combine all images with DR level higher than or equal to 1 as abnormal and DR level 0 as normal. The agent outperforms a normal human ophthalmologist for binary classification.
|Method||Binary Accuracy (%)|
|JointSeg [This work]||95.5|
Figure 6 shows the qualitative performance of the segmentation agent with respect to the ground truths and modified UNet design. Joint Segmentation performs better than improved UNet design. Considering the complexity of this problem to identify very small lesions; the proposed architecture was able to learn meaningful features without a pre-trained model.
We have presented an end-to-end deep convolutional semantic segmentation learning method that achieves faster convergence and better accuracy with random parameters initializations. Also, the network learns both classification and segmentation on the same domain. We also showed the performance of our methods on biomedical image processing dataset eyepacs to detect and classify diabetic retinopathy. One of the important aspects of this work is that the classifier agent can be trained on lose/ partial random labels without degrading segmentation agent performance; this helps towards solving the label scarcity problems of medical image domain.
We would like to thank the data team leads RajaRajalakshmi Kodhandapani and Ophthalmologists of Artelus; who have helped to prepare the segmentation dataset. Also, I would like to thank colleagues Anurag Sahay and Lalit Pant for their valuable advice during this work.
-  J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In CVPR, 2015
-  M. A. Islam, M. Rochan, N. D. Bruce, and Y. Wang. Gated feedback refinement network for dense image labeling. In CVPR, 2017
-  G. Lin, A. Milan, C. Shen, and I. Reid. Refinenet: Multipath refinement networks with identity mappings for high-resolution semantic segmentation. arXiv:1611.06612, 2016
-  C. Peng, X. Zhang, G. Yu, G. Luo, and J. Sun. Large kernel mattersâimprove semantic segmentation by global convolutional network.arXiv:1703.02719, 2017.
-  M. D. Zeiler, G. W. Taylor, and R. Fergus. Adaptive deconvolutional networks for mid and high level feature learning. In ICCV, 2011
-  H. Noh, S. Hong, and B. Han. Learning deconvolution network for semantic segmentation. In ICCV, 2015.
-  Haloi, M. Traffic sign classification using deep inception based convolutional networks. arXiv preprint arXiv:1511.02992, 2015.
-  Haloi, M. Deep Learning: Generalization Requires Deep Compositional Feature Space Design. arXiv preprint arXiv:1706.01983, 2017.
-  Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 2818-2826), 2016.
-  H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia. Pyramid scene parsing network. arXiv:1612.01105, 2016.
-  A. Paszke, A. Chaurasia, S. Kim, and E. Culurciello. Enet: A deep neural network architecture for real-time semantic segmentation. arXiv preprint arXiv:1606.02147, 2016.
-  He, K., Gkioxari, G., DollÃ¡r, P., Girshick, R. (2017). Mask r-cnn. arXiv preprint arXiv:1703.06870.
-  Pinheiro, P. O., Lin, T. Y., Collobert, R., DollÃ¡r, P. (2016). Learning to refine object segments. In European Conference on Computer Vision (pp. 75-91). Springer International Publishing.
-  P. Krahenbuhl and V. Koltun. Efficient inference in fully connected crfs with gaussian edge potentials. In NIPS, 2011.
-  V. Badrinarayanan, A. Kendall, and R. Cipolla. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. arXiv:1511.00561, 2015.
-  Z. Wu, C. Shen, and A. van den Hengel. Wider or deeper: Revisiting the resnet model for visual recognition. arXiv:1611.10080, 2016.
-  L.C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille. Semantic image segmentation with deep convolutional nets and fully connected crfs. In ICLR, 2015
-  L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L.Yuille. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. arXiv:1606.00915, 2016.
-  Zheng, S., Jayasumana, S., Romera-Paredes, B., Vineet, V., Su, Z., Du, D., Torr, P. H. Conditional random fields as recurrent neural networks. In Proceedings of the IEEE International Conference on Computer Vision (pp. 1529-1537), 2015.
-  Haloi, M. Improved microaneurysm detection using deep neural networks. arXiv preprint arXiv:1505.04424, 2015.
-  Ronneberger, O., Fischer, P., Brox, T. (2015, October). U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention (pp. 234-241). Springer, Cham.
-  S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv:1502.03167, 2015.
-  Simonyan, K., Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
-  https://github.com/openagi/tefla
-  Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., Ghemawat, S. Tensorflow: Large-scale machine learning on heterogeneous distributed systems. arXiv preprint arXiv:1603.0446, 2016.
-  O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, et al. Imagenet large scale visual recognition challenge. arXiv:1409.0575, 2014.
-  https://www.kaggle.com/c/diabetic-retinopathy-detection
-  B. Hariharan, P. Arbelaez, L. D. Bourdev, S. Maji, and J. Malik. Semantic contours from inverse detectors. In ICCV, 2011.