Multi-Path Feedback Recurrent Neural Networks for Scene Parsing

Multi-Path Feedback Recurrent Neural Networks for Scene Parsing


In this paper, we consider the scene parsing problem and propose a novel Multi-Path Feedback recurrent neural network (MPF-RNN) for parsing scene images. MPF-RNN can enhance the capability of RNNs in modeling long-range context information at multiple levels and better distinguish pixels that are easy to confuse. Different from feedforward CNNs and RNNs with only single feedback, MPF-RNN propagates the contextual features learned at top layer through multiple weighted recurrent connections to learn bottom features. For better training MPF-RNN, we propose a new strategy that considers accumulative loss at multiple recurrent steps to improve performance of the MPF-RNN on parsing small objects. With these two novel components, MPF-RNN has achieved significant improvement over strong baselines (VGG16 and Res101) on five challenging scene parsing benchmarks, including traditional SiftFlow, Barcelona, CamVid, Stanford Background as well as the recently released large-scale ADE20K.


Scene parsing has drawn increasing research interestdue to its wide applications in many attractive areas like autonomous vehicles, robot navigation and virtual reality. However, it remains a challenging problem since it requires solving segmentation, classification and detection simultaneously.

Recently, convolutional neural networks (CNNs) have been widely used for learning image representations and applied for scene parsing. However, CNNs can only capture high-level context information learned at top layers that have large receptive fields (RFs). The bottom layers are not exposed to valuable context information when they learn features. In addition, several recent works have demonstrated that even the top layers in a very deep model e.g. VGG16 [18] have limited RFs and receive limited context information [10] in fact. Therefore, CNNs usually encounter difficulties in distinguishing pixels that are easy to confuse locally and high-level context information is necessary. For example, in Figure 1, without the information of global context, the “field” pixels and “building” pixels are categorized incorrectly.

To address this challenge, we propose a novel multi-path feedback recurrent neural network (MPF-RNN). The overall architecture of our proposed MPF-RNN is illustrated in Figure ?. It has following two appealing characteristics. Firstly, MPF-RNN establishes recurrent connections to propagate downwards the outputs of the top layer to multiple bottom layers so as to fully exploit the context in the training process of different layers. Benefited from the enhanced context modeling capability, MPF-RNN gains a strong discriminative capability. Note that compared with previous RNN-based methods [13] which use only one feedback connection from the output layer to the input layer and the layer-wise self-feedback connections, respectively, MPF-RNN is superior by using a better architecture, i.e. explicitly incorporating context information into the training process of multiple hidden layers, which learns concepts with different abstractness. Secondly, MPF-RNN effectively fuses the output features across different time steps for classification or a more concrete parsing purpose. We empirically demonstrate that such multi-step fusion greatly boosts the final performance of MPF-RNN.

To verify the effectiveness of MPF-RNN, we have conducted extensive experiments over five popular and challenging scene parsing datasets, including SiftFlow [11], Barcelona [23], CamVid [2], Stanford Background [6] and recently released large-scale ADE20K [31] and demonstrated that MPF-RNN is capable of greatly enhancing the discriminative power of per-pixel feature representations.

Figure 1: Illustration on importance of context information for distinguishing pixels. Without context, the field pixels in the first image are misclassified as desert. Similarly, the street pixels in the second image are misclassified as building. MPF-RNN can better model context information and thus produces more smooth and semantically meaningful parsing results. Best viewed in color.
Figure 1: Illustration on importance of context information for distinguishing pixels. Without context, the field pixels in the first image are misclassified as desert. Similarly, the street pixels in the second image are misclassified as building. MPF-RNN can better model context information and thus produces more smooth and semantically meaningful parsing results. Best viewed in color.

2Related Work

Image Context Modeling One type of context modeling approaches for scene parsing is to use the probability graphical models (PGM) (e.g. CRF) to improve parsing results. In [3], CNN features are combined with a fully connected CRF to get more accurate parsing results. Compared to MPF-RNN, such methods usually suffer from intense calculation in inference due to their used time consuming mean field inference, which hinders their application in real-time scenario. Farabet et al. [5] encoded context information through surrounding contextual windows from multi-scale images. [20] proposed a recursive neural network to learn a mapping from visual features to the semantic space for pixel classification In [16], a parsing tree was used to propagate global context information. In [10], the global context feature is obtained by pooling the last layer’s output via global average pooling and then concatenated with local feature maps to learn a per-pixel classifier. Our method is different from them in both the context modeling scheme and the network architecture and achieves better performance.

Recurrent Neural Networks RNN has been employed to model long-range context in images. For instance, [13] built one recurrent connection from the output to the input layer, and [9] introduced layer-wise self-recurrent connections. Compared with those methods, the proposed MPF-RNN models the context by allowing multiple forms of recurrent connections. In addition, MPF-RNN combines the output features at multiple time steps for pixel classification. [21] utilized a parallel multi-dimensional long short-term memory for fast volumetric segmentation. However, its performance was relatively inferior. [17] were based on similar motivations that used RNNs to refine the learned features from a CNN by modeling the contextual dependencies along multiple spatial directions. In comparison, MPF-RNN incorporates the context information into the feature learning process of CNNs.

3The MPF-RNN Model

3.1Multi-Path Feedback

Throughout the paper, we use the following notations. For conciseness, we only consider one training sample. We denote a training sample as where is the raw image and is the ground truth parsing map of the image with as the ground truth category at the location in which is the number of categories. Since MPF-RNN is built upon CNNs by constructing recurrent connections from the top layer to multiple hidden layers, we therefore firstly consider a CNN composed of layers. Each layer outputs a feature map and we denote it as . Here and represent the input and final output of CNN. denotes the parameter of filters or weights to be learned in the -th layer. Using above notations, the outputs of an -layered CNN at each layer can be written as

where performs linear transformations on and is a composite of multiple specific functions including the activation function, pooling and softmax. Here the bias term is absorbed into .

MPF-RNN chooses layers out of layers and constructs recurrent connections from the top layer to each selected layer. Let denote the set of selected layers and let index the layers. By introducing the recurrent connections, each layer in takes both the output of its previous layer and the output of the top layer at the last time step as inputs. With denoting the index of time steps, Eqn. can be rewritten as

where and denote the output of the -th layer and the transformation matrix from the output of the -th layer to the hidden layer at time step , respectively. Note that is time-invariant, which means parameters in weight layers are shared across different time steps to reduce the model’s memory consumption. is time-variant so as to learn time step specific transformations for absorbing the context information from top layers. The output of top layers at different time steps conveys context information at different scales thus should have different transformations. The advantages of such a choice are verified in our experiments.

Following [10], function normalizes the input as where is a learnable scaler and denotes the norm of . As verified in [10], normalizing two feature maps output at different layers before performing combination is beneficial for convergence during training. One reason is that those two feature maps generally have different magnitude scales and they may slow down the convergence when their magnitudes are not balanced.

Now we proceed to explain the intuitions behind constructing multiple feedback connections. Since the RF of each layer in a deep neural network increases along with the depth, bottom and middle layers have relatively small RF. This limits the amount of context information that is perceptible when learning features in lower layers. Although higher layers might have larger RF and encode longer-range context, the context captured by higher layers cannot explicitly influence the states of the units in subsequent layers without top-down connections. Moreover, according to recent works [10], the effective RF might be much smaller than its theoretical value. For example, in a VGG16 model, although the theoretical RF size for the top layer fc7 is equal to , the effective RF size for the top layer fc7 is only about 1/4 of the theoretical RF. Due to the inadequate context information, layers in a deep model might not be able to learn context-aware features with strong discriminative power to distinguish local confusing pixels. Figure 1 shows examples of local confusion labeling by a CNN. To make layers in CNN incorporate context information, we therefore propose to propagate the output of the top layer to multiple hidden layers. Modulated with context, the deep model is context-aware when hierarchically learning concepts with different abstract levels. As a result, the model captures context dependencies to distinguish confusing local pixels. Actually, as the time step increases, the states of every unit in recurrent layers would have increasingly larger RF sizes and incorporate richer context.

We would also like to explain the reasons why we only build recurrent connections from the last layer: First, the last layer has the largest receptive field among all the layers whose output feature contains the richest contexts. Thus putting other layers in recurrent connections will introduce redundant context information and may hurt performance. Secondly, including more layers significantly increases computation cost. Only applying on the last layer gives a good trade-off between performance and efficiency.

Figure 2: The features learned with a small number of time steps have stronger capability of depicting small objects compared to the features learned with a large number of time steps. Left: Input image. Middle: Output of fc7 at t=1. Right: Output of fc7 at t=3. The plant in the bottom-left coner of the left image is captured in the middle image but ignored in the right image.
Figure 2: The features learned with a small number of time steps have stronger capability of depicting small objects compared to the features learned with a large number of time steps. Left: Input image. Middle: Output of fc7 at . Right: Output of fc7 at . The plant in the bottom-left coner of the left image is captured in the middle image but ignored in the right image.

3.2Multi-Step Fusion Loss

As shown in Figure ?, MPF-RNN is trained through a back propagation through time (BPTT) method which is equivalent to unfolding the recurrent network to a deep feedforward network according to the time steps by sharing parameters of weight layers across different time steps. The conventional objective function is minimizing the weighted sum of cross-entropy losses between the output probability of the unfolded network and the binary per-pixel ground truth label.

There are two disadvantages with this objective function. First, features for small objects in an image might be inconspicuous in the higher-level feature map due to a stack of convolution and pooling operations, which hurts the parsing performance since scene images have many small objects. Secondly, since the depth of an unfolded feedforward model with many time steps is large, training bottom layers in early time steps may suffer from a vanishing gradient problem. To handle the above two disadvantages, we propose a new objective function with respect to the output feature maps from multiple time steps. Formally, the objective function we propose is given as

Here denotes the number of time steps, denotes the combined feature of the top layer’s output at each time step, and denotes the logarithmic prediction probability produced by a classifier which maps from input to per-pixel category label . The specific formulation of depends on the chosen classifier for predicting per-pixel label, such as MLP, SVM and logistic regression classifier, etc.

In our method, the classifier contains a deconvolution layer which resizes the discriminative representation to be the same size as the input image for producing dense predictions, followed by a softmax loss layer to produce per-pixel categorical prediction probabilities. denotes the parameters in the deconvolution layer. Note that denotes the category of the pixel at location . balances the importance of feature maps at time step and is the class weight vector. We further discuss them in the experiments.

The benefits of using Eqn. are two-fold. On one hand, it improves MPF-RNN’s discriminability for small objects. Because the output features at later time steps have increasingly larger receptive field (RF) which receives wider-range context information, they may fail to distinguish small objects. In contrast, output features at earlier time steps have smaller RF and are more discriminative in capturing small objects. Therefore the combination of features in the last time step with those of ealier time steps helps the model identify small objects. On the other hand, for deep models with hundreds or thousands layers and large time steps, it may avoid the gradient vanishing problem in bottom layers at early time steps by giving shorter paths from them to the loss layer. Note that the multi-step fusion used in our method is different from the feature fusion method used in FCN [12]. While FCN combines feature maps from hidden layers in a CNN, we combine the outputs at different time steps in an RNN model, which retains stronger context modeling capabilities.


To better understand MPF-RNN, we compare it with several existing RNN models. Conventional RNN is trained with a sequence of inputs by computing the following sequences: and ,
where are the hidden layer and output layer at time step , respectively, while are weights matrices between the input and hidden layers, and among the hidden units themselves, respectively. is the output matrix between hidden and output layers. are activation functions. Compared with the conventional RNNs, MPF-RNN has multiple recurrent connections from the output layer to hidden layers (as indicated by Eqn. (Equation 2)) and moreover, MPF-RNN combines the outputs of multiple time steps as the final output (see Eqn. ).

Compared with those two recently proposed RNN-based models [13], which have one recurrent connection from the output layer to the input layer and layer-wise self-recurrent connections, respectively, our model can be deemed as a generalization by allowing more general recurrent connections. Specifically, when (recall is the index set of selected layers with which the output layer has feedback connections), there will be a recurrent connection from the output layer to the input layer in our model, as in [13]; when , there will be self-recurrent connections, as in [9]. By utilizing context information in different layers, our model has a stronger capability to learn discriminative features for each pixel. We also highlight the main difference of network architecture between MPF-RNN and RCN [8]. RCN is still essentially a feed-forward network which utilizes features in higher layers to aggregate features in bottom layer for facial keypoint localization problem, while MPF-RNN is a novel RNN architecture for better modeling the long-range context information in scene parsing problem.

The way to build multiple recurrent connections from the top layer to subsequent layers is a reminiscent of fully recurrent nets (FRN) [25] which is an MLP with each non-input unit receiving connections from all the other units. Different from FRN, we employ convolution/deconvolution layers to model the recurrent connections in MPF-RNN so as to preserve the spatial dependencies of 2D images. Besides, we do not build recurrent connections from the output layer to every subsequent layer, since the neighboring layers contain redundant information and the “fully recurrent” way is prone to over-fitting in scene parsing tasks. In addition, we perform multi-step fusion to improve the final performance, while FRN only uses the output of the final layer. To the best of our knowledge, we are among the first to apply multiple convolutional recurrent connections for solving scene parsing problems.


4.1Experiment Settings and Implementation Details

Evaluation Metrics

Adopted by most previous works as evaluation metrics, the per-pixel accuracy (PA) and the average per-class accuracy (CA) are used. PA is defined as the percentage of all correctly classified pixels while CA is the average of all category-wise accuracies.

Baseline Models

Following [3], we use a variant of the ImageNet pre-trained VGG16 network as the baseline model and fine-tune it on four scene parsing datasets, including SiftFlow, Barcelona, CamVid and Stanford Background. Here, fully connected (FC) layers (fc6 and fc7) are replaced by convolution layers. To accelerate the dense labeling prediction, the kernel size of fc6 is reduced to and the number of channels at FC layers is reduced to 1,024. This model is referred to as “VGG16-baseline” in the following experiments. Note that the VGG16-baseline model is equivalent to the MPF-RNN model with time step equal to 1. For simplicity, in the following text, we denote conv5 as conv5_1, conv4 as conv4_1 and conv3 as conv3_1.

Multi-Path Feedback

A practical problem with using MPF-RNN is how to choose the layer set , to which the recurrent connections from the top layer are constructed. There are two thumb rules. First, the layers at very bottom should not be chosen since they generally learn basic and simple patterns like edges, circles and dots, in which global context is not helpful. Secondly, it should be avoided to choose too many neighboring layers so that abundant information may be reduced in the features learned from neighboring layers. Following the above two rules, we conduct experiments with various on the validation set of SiftFlow and choose the one with best validation performance. The comparative results are shown in Table 1, from which we can see that choosing gives the best performance.

Multi-Step Fusion

There is a trade-off between the computation efficiency and model complexity when choosing the number of time steps for unfolding MPF-RNN. A deeper network can learn more complex features and give better performance. On the other hand, a large model is time consuming to train and test, which turns out to be problematic when applied to real-time scenarios, e.g. automatic driving. Table 2 compares the performance of MPF-RNN with different time steps on SiftFlow, from which we observe that among other time steps, achieves the best performance with fast speed. Therefore, we choose as the default time step for all datasets. Besides, we set and throughout our experiments. As an alternative, we can also see as learnable parameters and train them jointly with the whole model end-to-end. However, we do not explore this method in the paper.


The hyperparameters introduced by MPF-RNN, including and are fine-tuned on the validation set of SiftFlow as introduced above and then fixed for other datasets where MPF-RNN uses VGG16 network. Our experiments verified that such values of hyperparameters are optimum in all datasets.

Loss Re-Weighting

Since in scene parsing tasks, the class distribution is extremely unbalanced, it is common to re-weight different classes during training to attend rare classes [5]. In our model, we adopt the reweighting strategy by [17] because of its simplicity and effectiveness. Briefly, the weight for class is defined as where is the frequency of class and is a dataset-dependent scalar, which is defined according to 85%/15% frequent/rare classes rule.

Fine-Tuning Strategy

Our model is fine-tuned over target datasets using the stochastic gradient descent algorithm with momentum. For models using VGG16 network, settings of hyper-parameters including learning rate, weight decay and momentum follow [10]. The reported results are based on the model trained in 40 epochs. Data augmentation is used to reduce the risk of over-fitting and improve the generalization performance of deep neural network models. To make a fair comparison with other state-of-the-art methods, we only adopt the common random horizontal flipping and cropping during training.

Computational Efficiency

On a NVIDIA Titan X GPU, the training of MPF-RNN (the model in Table 3) on SiftFlow dataset finishes in about 6 hours and the testing time for an image with the resolution of 256256 is 0.06s.

Table 1: Comparative study of effects of different recurrent connections on final performance of MPF-RNN over SiftFlow dataset. The time step of MPF-RNN is fixed as 2 in all experiments in this table. The best results are shown in bold.
Recurrent connections PA(%) CA(%)
VGG16-baseline 84.7 51.5
{fc7} 85.4 55.1
{fc6, fc7} 85.9 55.5
{conv5, fc6, fc7} 86.2 55.8
{conv4, conv5, fc6, fc7} 86.4 56.3
{conv3, conv4, conv5, fc6, fc7}


We test MPF-RNN on five challenging scene parsing benchmarks, including SiftFlow [11], Barcelona [23], CamVid [2], Stanford Background [6] and ADE20K. We report the quantitative results here and more qualitative results of MPF-RNN are given in the Supplementary Material.


The SiftFlow dataset [11] consists of 2,400/200 color images with 33 semantic labels for training and testing.

Table 2: Comparative study of effects of different recurrent time steps and multi-step fusion (MSF) on final performance of MPF-RNN over SiftFlow dataset. herein.
(lr)2-3 (lr)4-5 Time steps PA(%) CA(%) PA(%) CA(%)
VGG16-baseline 84.7 51.5 N.A. N.A.
T = 2 86.4 56.3 86.0 55.5
T = 3
T = 4 86.8 55.9 84.7 54.8

Model Analysis

We analyze the MPF-RNN by investigating the effects of its two important components separately, i.e. multi-path feedback and multi-step fusion. Table 1 lists the performance of MPF-RNN, as well as the baseline models when different recurrent connections are used. The number of time steps is set as 2 for all MPF-RNN models and combination weights are and . Compared with the baseline model which has achieved 84.7%/51.5% PA/CA on this dataset, adding recurrent connections to multiple subsequent layers significantly improves the performance in terms of PA and CA. Only adding one recurrent connection to the layer fc7 can increase the performance to 85.4%/55.1%, which proves the benefit of global context modulating to the performance. Continuing adding recurrent connections to conv5 and conv4 consistently improves the performance. The reason for the continuous improvement is the sufficient utilization of the context information in learning context-aware features in different hidden layers. Based on above experimental results, it is verified that multi-path feedback is beneficial for boosting the performance of scene parsing systems. We also note that the performance tends to slow down its increasing (although still much higher than the baseline model) when adding too many recurrent connections (). Such a phenomenon implies the overfitting due to the increased number of parameters.

Table 3: Comparison with the state-of-the-art methods on SiftFlow.
Methods PA(%) CA(%)
Tighe et al. 79.2 39.2
Sharma et al. 79.6 33.6
Singh et al. 79.2 33.8
Yang et al. 79.8 48.7
Liang et al. 84.3 41.0
Long et al. 85.2 51.7
Liu et al. 86.8 52.0
Shuai et al. 85.3 55.7
MPF-RNN(ours) 86.9 56.5

Table 2 shows the effects of different time steps and multi-step fusion on the performance. We conduct experiments when by setting . Concluded from Table 2, the PA/CA consistently improves from (VGG16-baseline) to . Such improvement is attributed to the more complex discriminative features learned using larger time steps, which is proportional to the depth of the feedforward deep model after MPF-RNN is unfolded. The improvement is also consistent with observations in [18] that deeper and larger models have stronger feature representation capabilities compared with shallow ones. When , the performance is worse than that when due to the overfitting problem.

We also conduct experiments when only the top layer’s feature in the last time step is used as the input to the classifier. Specifically, we set and and keep the other conditions unchanged. It is observed from Table 2 that under any time steps, the PA/CA are worse than those when multi-step fusion is applied. There are two reasons for the inferior performance. Updating bottom layers parameters is insufficient since it takes many hidden layers to propagate the error message from the output layer to bottom layers. In addition, as illustrated in Figure 2, the output layer at the last time step might ignore small objects, which could otherwise be complemented by using multi-step fusion. Above experiment demonstrates the effectiveness of multi-step fusion. Note that although the conclusions made in model analysis are based on the results on SiftFlow, we have verified through experiments that they also hold on other three datasets where MPF-RNN uses VGG16 network.

Comparison with State-of-the-Art

The comparison results of MPF-RNN with other state-of-the-art methods are shown in Table 3, from which we can see that MPF-RNN achieves the best performance against all the compared methods. Specifically, our method significantly outperforms Pinheiro et al. [13] and Liang et al. [9] by increasing the PA/CA by 9.2%/26.7% and 2.6%/15.5%, respectively. Note that [13] and [9] are all RNN-based methods. The remarkable improvement of MPF-RNN over them is attributed to MPF-RNN’s powerful context modeling strategy by constructing multiple recurrent connections in various forms. Compared with [10], MPF-RNN achieves much better CA due to its superior capability of distinguishing small objects.


To further verify scalability of the MPF-RNN to large-scale dataset and deeper networks, we conduct experiments using ResNet101 [7] on recently released ADE20K dataset [31] which also serves as the dataset of scene parsing challenge in ILSVRC161. Containing 20K/2K/3K fully annotated scene-centric train/val/test images with 150 classes, ADE20K has more diverse scenes and richer annotations compared to other scene parsing dataset, which make it a challenging benchmark. Following [4], all convolution layers in original ResNet101 [7] after conv34 are replaced with dilated convolution layers to compute dense scores at stride of 8 pixels, followed by a deconvolution layer to reach original image resolution. This model is used as a strong baseline and referred to as ResNet101-Baseline. Since the test set is not available till submission of the paper, we use the val set to test the performance of MPF-RNN. In order to fine-tune hyperparameters, we randomly extract 2K images from train set as our validation data and retrain our model using whole train set after fixing hyperparameters. Through validation, we set and the value of and are the same as those in the last section. MPF-RNN and ResNet101-Baseline are trained for 20 epochs by fine-tuning ResNet1012 on ADE20K with the same solver configurations of [4]. Data augmentation only include random horizontal flipping and cropping. We do not use loss re-weighting in this dataset.

Table 4: Comparison results of MPF-RNN on ADE20K val set. The results of FCN, SegNet and DilatedNet are referred from the reported number in
Methods PA(%) mIOU(%)
FCN 71.32 29.39
SegNet 71.0 21.64
DilatedNet 73.55 32.31
Res101-Baseline 73.71 32.65
MPF-RNN (ours) 76.49 34.63

In Table 4, we compare the performance of MPF-RNN and baseline models. Following the evaluation metric in ILSVRC16 scene parsing challenge, we compare the PA and mean IOU (the mean value of intersection-over-union between the predicted and ground-truth pixels across all classes) between different models. It is observed from Table 4 that Res101-Baseline has achieved better performance compared to other models based on VGG16 due to its superior deep architecture. By using MPF-RNN, MPF-RNN significantly surpasses this strong baseline model by 2.78%/1.98 in terms of PA/mIOU, again demonstrating the capability of MPF-RNN to improve the scene parsing performance of deep networks.

Results on Other Datasets

We further evaluate MPF-RNN on Barcelona, Stanford Background and Camvid datasets. MPF-RNN achieves the best performance on these three datasets: (i) Stanford Background: (PA) and (CA), outperforming state-of-the-art with (PA) and (CA) absolutely; (ii) Barcelona: (PA) and (CA) higher than state-of-the-art with around (PA) and (CA) absolutely; (iii) Camvid: (PA) and (CA), outperforming state-of-the-art with (PA) and (CA) absolutely. Due to space limits, more numbers and details of experimental results on these three datasets are deferred to the supplementary material.


We proposed a novel Multi-Path Feedback recurrent neural network (MPF-RNN) for better solving scene parsing problems. In MPF-RNN, multiple recurrent connections are constructed from the output layer to hidden layers. To learn features discriminative for small objects, output features at each time step are combined for per-pixel prediction in MPF-RNN. Experimental results over five challenging scene-parsing datasets, including SiftFlow, ADE20K, Barcelona, Stanford Background, and CamVid clearly demonstrated the advantages of MPF-RNN for scene parsing.

Acknowledgement Xiaojie Jin is immensely grateful to Bing Shuai at NTU and Wei Liu at UNC - Chapel Hill for valuable discussions. The work of Jiashi Feng was partially supported by National University of Singapore startup grant R-263-000-C08-133 and Ministry of Education of Singapore AcRF Tier One grant R-263-000-C21-112.



  1. 2015.
    Badrinarayanan, V.; Kendall, A.; and Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation.
  2. 2008.
    Brostow, G. J.; Shotton, J.; Fauqueur, J.; and Cipolla, R. Segmentation and recognition using structure from motion point clouds.
  3. 2015.
    Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; and Yuille, A. L. Semantic image segmentation with deep convolutional nets and fully connected crfs.
  4. 2016.
    Chen, L.; Papandreou, G.; Kokkinos, I.; Murphy, K.; and Yuille, A. L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs.
  5. 2013.
    Farabet, C.; Couprie, C.; Najman, L.; and LeCun, Y. Learning hierarchical features for scene labeling.
  6. 2009.
    Gould, S.; Fulton, R.; and Koller, D. Decomposing a scene into geometric and semantically consistent regions.
  7. 2015.
    He, K.; Zhang, X.; Ren, S.; and Sun, J. Deep residual learning for image recognition.
  8. 2015.
    Honari, S.; Yosinski, J.; Vincent, P.; and Pal, C. Recombinator networks: Learning coarse-to-fine feature aggregation.
  9. 2015.
    Liang, M.; Hu, X.; and Zhang, B. Convolutional neural networks with intra-layer recurrent connections for scene labeling.
  10. 2015.
    Liu, W.; Rabinovich, A.; and Berg, A. C. Parsenet: Looking wider to see better.
  11. 2009.
    Liu, C.; Yuen, J.; and Torralba, A. Nonparametric scene parsing: Label transfer via dense scene alignment.
  12. 2015.
    Long, J.; Shelhamer, E.; and Darrell, T. Fully convolutional networks for semantic segmentation.
  13. 2013.
    Pinheiro, P. H., and Collobert, R. Recurrent convolutional neural networks for scene parsing.
  14. 2014.
    Roy, A., and Todorovic, S. Scene labeling using beam search under mutex constraints.
  15. 2015.
    Schwing, A. G., and Urtasun, R. Fully connected deep structured networks.
  16. 2014.
    Sharma, A.; Tuzel, O.; and Liu, M.-Y. Recursive context propagation network for semantic scene labeling.
  17. 2015.
    Shuai, B.; Zuo, Z.; Wang, G.; and Wang, B. Dag-recurrent neural networks for scene labeling.
  18. 2014.
    Simonyan, K., and Zisserman, A. Very deep convolutional networks for large-scale image recognition.
  19. 2013.
    Singh, G., and Kosecka, J. Nonparametric scene parsing with adaptive feature relevance and semantic context.
  20. 2011.
    Socher, R.; Lin, C. C.; Manning, C.; and Ng, A. Y. Parsing natural scenes and natural language with recursive neural networks.
  21. 2015.
    Stollenga, M. F.; Byeon, W.; Liwicki, M.; and Schmidhuber, J. Parallel multi-dimensional lstm, with application to fast biomedical volumetric image segmentation.
  22. 2014.
    Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; and Rabinovich, A. Going deeper with convolutions.
  23. 2010.
    Tighe, J., and Lazebnik, S. Superparsing: scalable nonparametric image parsing with superpixels.
  24. 2015.
    Visin, F.; Kastner, K.; Courville, A. C.; Bengio, Y.; Matteucci, M.; and Cho, K. Reseg: A recurrent neural network for object segmentation.
  25. 1989.
    Williams, R. J., and Zipser, D. A learning algorithm for continually running fully recurrent neural networks.
  26. 2014.
    Yang, J.; Price, B.; Cohen, S.; and Yang, M.-H. Context driven scene parsing with attention to rare classes.
  27. 2015.
    Yu, F., and Koltun, V. Multi-scale context aggregation by dilated convolutions.
  28. 2012.
    Zhang, Y., and Chen, T. Efficient inference for fully-connected crfs with stationarity.
  29. 2015.
    Zheng, S.; Jayasumana, S.; Romera-Paredes, B.; Vineet, V.; Su, Z.; Du, D.; Huang, C.; and Torr, P. H. Conditional random fields as recurrent neural networks.
  30. 2014.
    Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; and Torralba, A. Object detectors emerge in deep scene cnns.
  31. 2016.
    Zhou, B.; Zhao, H.; Puig, X.; Fidler, S.; Barriuso, A.; and Torralba, A. Semantic understanding of scenes through the ade20k dataset.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
Add comment
Loading ...
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test description