Learning to Select Pre-trained Deep Representations with Bayesian Evidence Framework

Learning to Select Pre-trained Deep Representations with
Bayesian Evidence Framework

Yong-Deok Kim     Taewoong Jang     Bohyung Han     Seungjin Choi
Software R&D Center, Device Solutions, Samsung Electronics, Korea
Stradvision Inc., Korea
Department of Computer Science and Engineering, POSTECH, Korea
yd.mlg.kim@samsung.com   taewoong.jang@stradvision.com   {bhhan,seungjin}@postech.ac.kr
This work was done when Y. Kim and T. Jang were with POSTECH.
Abstract

We propose a Bayesian evidence framework to facilitate transfer learning from pre-trained deep convolutional neural networks (CNNs). Our framework is formulated on top of a least squares SVM (LS-SVM) classifier, which is simple and fast in both training and testing, and achieves competitive performance in practice. The regularization parameters in LS-SVM is estimated automatically without grid search and cross-validation by maximizing evidence, which is a useful measure to select the best performing CNN out of multiple candidates for transfer learning; the evidence is optimized efficiently by employing Aitken’s delta-squared process, which accelerates convergence of fixed point update. The proposed Bayesian evidence framework also provides a good solution to identify the best ensemble of heterogeneous CNNs through a greedy algorithm. Our Bayesian evidence framework for transfer learning is tested on 12 visual recognition datasets and illustrates the state-of-the-art performance consistently in terms of prediction accuracy and modeling efficiency.

1 Introduction

Image representations from deep CNN models trained for specific image classification tasks turn out to be powerful even for general purposes [2, 6, 7, 21, 23] and useful for transfer learning or domain adaptation. Therefore, CNNs trained on specific problems or datasets are often fine-tuned to facilitate training for new tasks or domains [2, 6, 13, 15, 16, 36], and an even simpler approach—application of off-the-shelf classification algorithms such as SVM to the representations from deep CNNs [7]—is getting more attractive in many computer vision problems. However, fine-tuning of an entire deep network still requires a lot of efforts and resources, and SVM-based methods also involve time consuming grid search and cross validation to identify good regularization parameters. In addition, when multiple pre-trained deep CNN models are available, it is unclear which pre-trained models are appropriate for target tasks and which classifiers would maximize accuracy and efficiency. Unfortunately, most existing techniques for transfer learning or domain adaptation are limited to empirical analysis or ad-hoc application specific approaches.

Figure 1: We address a problem to select the best CNN out of multiple candidates as shown in this figure. Additionally, our algorithm is capable of identifying the best ensemble of multiple CNNs to further improve performance.

We propose a simple but effective algorithm for transfer learning from pre-trained deep CNNs based on Bayesian least squares SVM (LS-SVM), which is formulated with Bayesian evidence framework [18, 29] and LS-SVM [26]. This approach automatically determines regularization parameters in a principled way, and shows comparable performance to the standard SVMs based on hinge loss or squared hinge loss. More importantly, Bayesian LS-SVM provides an effective solution to select the best CNN out of multiple candidates and identify a good ensemble of heterogeneous CNNs for performance improvement. Figure 1 illustrates our approach. We also propose a fast Bayesian LS-SVM, which maximizes the evidence more efficiently based on Aitken’s delta-squared process [1].

One may argue against the use of LS-SVM for classification because the least squares loss function in LS-SVM tends to penalize well-classified examples. However, least squares loss is often used for training multilayer perceptron [4] and shows comparable performance to SVMs [28, 37]. In addition, Bayesian LS-SVM provides a technically sound formulation with outstanding performance in terms of speed and accuracy for transfer learning with deep representations. We also propose a fast Bayesian LS-SVM, which maximizes the evidence more efficiently based on Aitken�s delta-squared process [1]. Considering simplicity and accuracy, we claim that our fast Bayesian LS-SVM is a reasonable choice for transfer learning with deep learning representation in visual recognition problems. Based on this approach, we achieved promising results compared to the state-of-the-art techniques on 12 visual recognition tasks.

The rest of this paper is organized as follows. Section 2 describes examples of transfer learning or domain adaptation based on pre-trained CNNs for visual recognition problems. Then, we discuss Bayesian evidence framework applicable to the same problem in Section 3 and its acceleration technique using Aitken’s delta-squared process in Section 4. The performance of our algorithm in various applications is demonstrated in Section 5.

2 Related Work

Since AlexNet [17] demonstrated impressive performance in the ImageNet large scale visual recognition challenge (LSVRC) 2012, a few deep CNNs with different architectures, e.g., VGG [25] and GoogLeNet [27], have been proposed in the subsequent events. Instead of training deep CNNs from scratch, some people have attempted to refine pre-trained networks for new tasks or datasets by updating the weights of all neurons or have adopted the intermediate outputs of existing deep networks as generic visual feature descriptors. These strategies can be interpreted as transfer learning or domain adaptation.

Refining a pre-trained CNN is called fine-tuning, where the architecture of the network may be preserved while weights are updated based on new training data. Fine-tuning is generally useful to improve performance [2, 6, 13, 36] but requires careful implementation to avoid overfitting. The second approach regards the pre-trained CNNs as feature extraction machines and combines the deep representations with the off-the-shelf classifiers such as linear SVM [7, 34], logistic regression [7, 34], and multi-layer neural network [21]. The techniques in this category have been successful in many visual recognition tasks [2, 23, 24].

When combining a classification algorithm with image representations from pre-trained deep CNNs, we often face a critical issue. Although several deep CNN models trained on large scale image repositories are publicly available, there is no principled way to select a CNN out of multiple candidates and find the best ensemble of multiple CNNs for performance optimization. Existing algorithms typically rely on ad-hoc methods for model selection and fail to provide clear evidence for superior performance [2].

3 Bayesian LS-SVM for Model Selection

This section discusses a Bayesian evidence framework to select the best CNN model(s) in the presence of transferable multiple candidates and identify a reasonable regularization parameter for LS-SVM classifier automatically.

3.1 Problem Definition and Formulation

Suppose that we have a set of pre-trained deep CNN models denoted by . Our goal is to identify the best performing deep CNN model among the networks for transfer learning. A naïve approach is to perform fine tuning of network for target task, which requires substantial efforts for training. Another option is to replace some of fully connected layers in a CNN with an off-the-shelf classifier such as SVM and check the performance of target task through parameter tuning for each network, which would also be computationally expensive.

We adopt a Bayesian evidence framework based on LS-SVM to achieve the goal in a principled way, where the evidence of each network is maximized iteratively and the maximum evidences are used to select a reasonable model. During the evidence maximization procedure, the regularization parameter of LS-SVM is identified automatically without time consuming grid search and cross-validation. In addition, the Bayesian evidence framework is also applied to the construction of an ensemble of multiple CNNs to accomplish further performance improvement.

3.2 Ls-Svm

We deal with multi-label or multi-class classification problem, where the number of categories is . Let be a training set, where is a feature vector and is a binary variable that is set to 1 if label is given to and 0 otherwise. Then, for each class , we minimize a least squares loss with regularization penalty as follows:

(1)

where and . The optimal solution of the problem in (1) is given by

(2)

where is the eigen-decomposition of and is an identity matrix. This regularized least squares approach has clear benefit that it requires only one eigen-decomposition of to obtain the solution in (2) for all combinations of and .

3.3 Bayesian Evidence Framework

The optimization of the regularized least squares formulation presented in (1) is equivalent to the maximization of the posterior with fixed hyperparamters and denoted by , where . The posterior can be decomposed into two terms by Bayesian theorem as

(3)

where corresponds to Gaussian observation noise model given by

(4)

and denotes a zero-mean isotropic Gaussian prior as

(5)

Note that we dropped superscript for notational simplicity from the equations in this subsection.

In the Bayesian evidence framework [18, 29], the evidence, also known as marginal likelihood, is a function of hyperparameters and as

(6)

Under the probabilistic model assumptions corresponding to (4) and (5), the log evidence is given by

(7)

where the precision matrix and mean vector of the posterior are given respectively by

The log evidence is maximized by repeatedly alternating the following fixed point update rules

(8)

which involves the derivation of as

(9)

where are eigenvalues of . Note that and should be re-estimated after each update of and .

Another pair of update rules of and are derived by an expectation-maximization (EM) technique as

(10)
(11)

but these procedures are substantially slower than the fixed point update rules in (8).

Through the optimization procedures described above, we determine the regularization parameter . Although the estimated parameters are not optimal, they may still be reasonable solutions since they are obtained by maximizing marginal likelihood in (6).

3.4 Model Selection using Evidence

The evidence computed in the previous subsection is for a single class, and the overall evidence for entire classes, denoted by , is obtained by the summation of the evidences from individual classes, which is given by

(12)

We compute the overall evidence corresponding to each deep CNN model, and choose the model with the maximum evidence for transfer learning. We expect that the selected model performs best among all candidates, which will be verified in our experiment.

In addition, when an ensemble of deep CNNs needs to be constructed for a target task, our approach selects a subset of good pre-trained CNNs in a greedy manner. Specifically, we add a network with the largest evidence in each stage and test whether the augmented network improves the evidence or not. The network is accepted if the evidence increases, or rejected otherwise. After the last candidate is tested, we obtain the final network combination and its associated model learned with the concatenated feature descriptors from accepted networks.

4 Fast Bayesian LS-SVM

Bayesian evidence framework discussed in Section 3 is useful to identify a good CNN for transfer learning and a reasonable regularization parameter. To make this framework even more practical, we present a faster algorithm to accomplish the same goal and a new theory that guarantees the converges of the algorithm.

4.1 Reformulation of Evidence

We are going to reduce to a function with only one parameter that directly corresponds to the regularization parameter . To this end, we re-write by using the eigen-decomposition as

(13)

where is the -th diagonal element in and denotes the -th element in . Then, we re-parameterize into as

(14)

The derivative of with respect to is given by

and we obtain the following equation by setting this derivative to zero,

(15)

Finally, we obtain a one-dimensional function of the log evidence by plugging (15) into (4.1), which is given by

(16)

Figure 2 illustrates the curvature of this log evidence function with respect to .

Figure 2: Plot of the log evidence with respect to . Note that is neither convex nor concave.

4.2 New Fixed-point Update Rule

We now derive a new fixed point update rule and present the sufficient condition for the existence of a fixed point. The stationary points in (4.1) with respect to satisfy

(17)

and we update the fixed-point by maximizing (4.1) as

(18)

As illustrated in Figure 2, in (4.1) is neither convex nor concave as illustrated in the supplementary file. However, we can show the sufficient condition of the existence of the fixed point using the following theorem.

Theorem 1.

Denote the update rule in (18) by . If is a binary variable and is an normalized nonnegative vector, then has a fixed point.

Proof.

We first show that is asymptotically linear in as

Since is binary and is normalized and nonnegative, we can derive the following two relations,

(19)
(20)

where . From (19) and (20), it is shown that .

Obviously, and there exists a such that . The intermediate value theorem implies the existence of such that , where as illustrated in Figure 3. ∎

Figure 3: Aitken’s delta-squared process. The fixed point update function is approximated by green dashed line and its intersection with becomes the next update point.

The fixed point is unique if is concave. Although it is always concave according to our observation, we have no proof yet and leave it as a future work

4.3 Speed Up Algorithm

  Input: and .
  Output: Optimal solutions .
  Initialize                  // e.g.,
   eigen-decomposition
   diag
  repeat
     
      UPDATE
      UPDATE
     
     if  or  then
        
     end if
  until     // e.g.,
  
Algorithm 1 Fast Bayesian Least Squares
  
  
  
  
  return
Algorithm 2 = UPDATE

We accelerate the fixed point update rule in (18) by using Aitken’s delta-squared process [1]. Figure 3 illustrates the Aitken’s delta-squared process. Let’s focus on the two points and , and line going through these two points. The equation of this line is

(21)

where and are replaced by and , respectively. The idea behind Aitken’s method is to approximate fixed point using the intersection of the line in (21) with line , which is given by

(22)

Our fast Bayesian learning algorithm for the regularized least squares problem in (1) is summarized in Algorithm 1. In our algorithm, we first compute the eigen-decomposition of . This is the most time consuming part but needs to be performed only once since the result can be reused for every label in . After that, we obtain the regularization parameter through an iterative procedure.

When we apply the Aitken’s delta-squared process, we have two potential failure cases as in Figure 4(a) and 4(b). The first case often arises if the initial is far from the fixed point , and the second case occurs when the approximating line in (21) is parallel to . Fortunately, these failures rarely happen in practice and can be handled easily by skipping the procedure in (22) and updating with .

Figure 4: Two failure cases of Aitken’s delta-squared process. (left) The first case arises if initial is far from the fixed point , which results in . (right) The second case occurs when approximating line (dashed green) is parallel to , where .

Figure 5: Comparison between Aitken’s delta-squared process, fixed point update rules, and EM update rules on PASCAL VOC 2012 dataset (class = aeroplane). Aitken’s delta-squared process significantly faster than other methods.

Figure 5 demonstrates the relative convergence rates of three different techniques—Aitken’s delta-squared process in Algorithm 1, fixed point update rules in (8), and EM update method, where the Aitken’s delta-squared process is significantly faster than others for convergence.

5 Experiments

We present the details of our experiment setting and the performance of our algorithm compared to the state-of-the-art techniques in 12 visual recognition benchmark datasets.

5.1 Datasets and Image Representation

The benchmark datasets involve various visual recognition tasks such as object recognition, photo annotation, scene recognition, fine grained recognition, visual attribute detection, and action recognition. Table 1 presents the characteristics of the datasets. In our experiment, we followed the given train and test split and evaluation measure of each dataset. For the datasets with bounding box annotations such as CUB200-2011, UIUC object attribute, Human attribute, and Stanford 40 actions, we enlarged the bounding boxes by 150% to consider neighborhood context as suggested in [23, 2].

Dataset Task Box Measure
PASCAL VOC 2007 [8] object recognition 5011 4952 20 1.5 mean AP
PASCAL VOC 2012 [9] object recognition 5717 5823 20 1.5 mean AP
Caltech 101 [12] object recognition 3060 6086 102 1 mean Acc.
Caltech 256 [14] object recognition 15420 15187 257 1 mean Acc.
ImageCLEF 2011 [20] photo annotation 8000 10000 99 11.9 mean AP
MIT Indoor Scene   [22] scene recognition 5360 1340 67 1 mean Acc.
SUN 397 Scene [32] scene recognition 19850 19850 397 1 mean Acc.
CUB 200-2011 [30] fine-grained recognition 5994 5794 200 1 mean Acc.
Oxford Flowers   [19] fine-grained recognition 2040 6149 200 1 mean Acc.
UIUC object attributes [11] attribute detection 6340 8999 64 7.1 mean AUC
Human attributes [5] attribute detection 4013 4022 9 1.8 mean AP
Stanford 40 actions [33] action recognition 4000 5532 40 1 mean AP
Table 1: Characteristics of the 12 datasets. : number of training data, : number of test data, : number of classes, : average number of labels per image, AP: average precision, Acc.: accuracy, AUC: area under the ROC curve.

For deep learning representations, we selected 4 pre-trained CNNs from the Caffe Model Zoo: GoogLeNet [31], VGG19 [25], and AlexNet [7] trained on ImageNet, and GoogLeNet trained on Places  [31]. As generic image representations, we used the 4096 dimensional activations of the first fully connected layer in VGG19 and AlexNet and the 1024 dimensional vector obtained from the global average pooling layer located right before the final softmax layer in GoogLeNet.

Our implementation is in Matlab2011a, and all experiments were conducted on a quad-core Intel(R) core(TM) i7-3820 @ 3.60GHz processor.

5.2 Bayesian LS-SVM vs. SVM

We first compare the performance of our Bayesian LS-SVM with the standard SVM when they are applied to deep CNN features for visual recognition problems. We used only a single image scale in this experiment. LIBLINEAR [10] package is used for SVM training and the regularization parameters are selected by grid search with cross validations.

Table 2 presents the complete results of our experiment. Bayesian LS-SVM is competitive to SVM in terms of prediction accuracy even with significantly reduced training time. Training SVM is getting slower than Bayesian LS-SVM as the number of classes increases so it is particularly slow in Caltech 256 and SUN 397 datasets.

LS-SVM SVM LS-SVM SVM
Bayesian
CV (5 folds)
CV (5 folds)
Bayesian
CV (5 folds)
CV (5 folds)
CNN Best Acc. Evidence Time Acc. Time Best Acc. Time Best Acc. Evidence Time Acc. Time Best Acc. Time
PASCAL VOC 2007 [8] SUN-397 [32]
85.3 85.2 46.9 1.1 85.2 8.4 85.0 84.7 122.4 48.1 47.0 12.8 3.1 48.1 36.5 54.2 54.2 8739.6
74.1 73.8 38.6 1.0 74.0 8.1 74.1 73.9 144.3 61.1 60.1 13.2 2.9 61.1 34.4 63.3 63.3 8589.4
85.9 85.8 48.0 41.9 85.8 172.2 85.9 85.8 257.5 55.0 53.7 12.9 57.4 54.9 419.8 57.1 57.1 20254.0
75.2 75.0 32.5 41.7 75.0 160.4 75.3 75.2 211.1 45.4 44.9 12.7 50.8 45.4 419.0 48.6 48.6 10781.8
PASCAL VOC 2012 [9] CUB-200 [30]
84.4 84.3 51.3 1.2 84.3 8.6 83.9 83.7 140.8 65.2 64.3 15.6 1.3 64.1 11.0 67.6 56.5 1201.9
73.2 72.9 40.6 1.1 73.1 8.4 73.2 73.1 170.7 16.4 13.6 14.9 1.5 15.0 11.1 16.8 11.1 1664.6
85.2 85.1 52.9 42.7 85.2 161.5 85.6 85.4 295.9 69.2 68.6 15.8 44.1 61.5 259.2 71.1 59.4 2776.2
74.1 73.9 34.3 42.7 74.0 161.8 74.4 74.3 160.7 59.0 58.5 15.5 45.3 46.6 257.9 61.4 51.6 1645.5
Caltech 101 [12] Oxford Flowers [19]
90.6 90.0 37.8 1.0 89.6 6.0 91.4 85.1 325.0 85.5 84.7 21.8 0.9 82.0 5.5 87.4 72.0 198.8
57.0 54.3 30.6 0.9 55.1 5.9 57.2 41.8 390.3 55.6 51.7 19.4 0.9 51.8 5.5 57.1 32.8 234.7
92.2 92.1 40.9 31.5 88.8 142.7 92.2 86.8 729.4 87.5 87.1 22.5 26.9 82.1 142.2 87.6 73.4 520.9
89.3 89.2 37.3 32.0 83.4 146.9 90.0 83.5 595.3 87.6 87.6 22.9 27.3 81.8 146.7 88.3 77.1 271.3
Caltech 256 [14] UIUC Attributes [11]
77.8 77.2 59.9 2.3 77.8 21.8 81.2 81.2 4060.4 91.5 90.3 13.5 1.4 90.9 8.0 91.3 90.6 605.5
44.9 42.6 55.9 2.2 44.9 21.2 48.6 48.6 4991.8 87.8 86.6 10.5 1.3 87.1 7.4 88.0 87.6 726.0
82.0 81.1 62.3 52.5 81.7 339.7 82.7 82.7 9653.1 92.5 91.1 14.4 43.8 92.0 186.3 92.2 91.7 1285.4
69.7 68.9 58.6 52.9 69.7 336.9 72.3 72.3 5348.6 91.4 89.9 12.9 44.1 91.0 191.2 90.8 90.5 683.7
ImageCLEF [20] Human Attributes [5]
49.1 48.9 20.5 1.5 48.8 37.0 47.7 47.4 1218.6 76.0 75.8 -74.8 1.0 75.8 5.0 74.2 74.1 70.6
47.5 47.1 20.8 1.4 47.1 36.9 47.1 46.7 1410.5 58.7 58.4 -103.1 1.0 58.0 4.8 56.9 56.5 85.5
50.7 50.3 21.3 45.9 50.4 248.5 50.4 50.1 2531.2 75.4 75.1 -76.0 40.3 75.2 124.2 73.1 72.8 131.9
44.8 44.6 18.7 46.1 44.6 245.9 44.4 44.1 2140.0 71.9 71.3 -84.4 40.7 71.7 121.2 70.0 69.9 63.3
MIT Indoor [22] Stanford 40 Action [33]
66.7 66.0 30.1 1.2 66.7 5.8 69.4 69.2 400.9 70.2 69.8 100.4 1.0 69.6 11.6 69.8 69.6 211.7
80.0 79.9 35.2 1.1 80.0 5.8 81.1 80.4 402.5 48.3 47.6 86.5 1.1 47.9 11.4 48.2 47.7 246.2
73.2 73.1 31.1 42.6 73.2 186.8 74.7 74.7 895.5 75.4 75.2 109.3 41.1 75.1 142.9 75.8 75.3 418.7
62.0 61.1 28.6 42.2 60.5 187.4 63.1 63.1 460.9 58.0 57.7 89.6 41.5 57.5 156.5 57.4 57.1 206.8
Table 2: Bayesian LS-SVM versus SVM. Without time consuming cross validation procedure, Bayesian LS-SVM achieves prediction accuracy competitive to SVM. In addition, Bayesian LS-SVM selects the proper CNN for each task by using the evidence (see bold-faced numbers). Best accuracy in LS-SVM and SVM denotes the maximum achievable accuracy in test dataset using all available learned models. Note that the selected model by Bayesian evidence framework or cross validation may not be the best one in testing. The following sets of regularization parameters are tested for cross validation in LS-SVM and SVM, respectively: and . (: GoogLeNet-ImageNet, : GoogLeNet-Place, : VGG19, and : AlexNet)

Another notable observation in Table 2 is that the order of prediction accuracy is highly correlated to the evidence. This means that the selected model by Bayesian LS-SVM produces reliable testing accuracy and a proper deep learning image representation is obtained without time consuming grid search and cross validation. Note that cross validations in LS-SVM and SVM play the same role, but are less reliable and slower than our Bayesian evidence framework. The capability to select the appropriate CNN model and the corresponding regularization parameter is one of the most important properties of our algorithm.

5.3 Comparison with Other Methods

We now show that our Bayesian LS-SVM identifies a combination of multiple CNNs to improve accuracy without grid search and cross validation. For each task, we select a subset of 4 pre-trained CNNs in a greedy manner; we add CNNs to our selection, one by one, until the evidence does not increase. Our algorithm is compared with DeCAF [7], Zeiler [34], INRIA [21], KTH-S [23], KTH-FT [2], VGG [25], Zhang [35, 36], and TUBFI [3]. In addition, our ensembles identified by greedy evidence maximization are compared with the oracle combinations—the ones with the highest accuracy in test set found by exhaustive search—and the best combinations found by exhaustive evidence maximization.

Table 3 presents that our ensembles approach achieves the best performance in most of the 12 tasks. The identified ensembles by the greedy approach are consistent with the selections by exhaustive evidence maximization and even oracle selections111This option is practically impossible since it requires evaluation with test dataset using all available models for model selection. made by testing accuracy maximization. Note that our network selections are natural and reasonable; GoogLeNet-ImageNet and VGG19 are selected frequently while GoogLeNet-Place is preferred to GoogLeNet-ImageNet in MIT Indoor and SUN-397 since the datasets are constructed for scene recognition. It turns out that the proposed algorithm tends to choose the networks with higher accuracies in the target task even though it makes selections based only on the evidence in a greedy manner. An interesting observation is that our result is less consistent with the selections by oracle and exhaustive evidence maximization in Stanford 40 Actions dataset, where GoogLeNet-Place seems to provide complementary information even with its low accuracy and is helpful to improve recognition performance. It is probably because actions are frequently performed at typical places, e.g., a fair portion of images in brushing teeth class are taken from bathrooms.

Method VOC07 VOC12 CAL101 CAL256 CLEF MIT SUN Birds Flowers UIUC Human Action
DeCAF - - 86.9 - - - 38.0 65.0 - - - -
Zeiler - 79.0 86.5 74.2 - - - - - - - -
INRIA 77.7 82.8 - - - - - - - - - -
KTH-S 71.8 - - - - 64.9 49.6 62.8 90.5 90.6 73.8 58.9
KTH-FT 80.7 - - 71.3 56.0 67.1 91.3 91.5 74.6 66.4
VGG 89.7 89.3 92.7 86.2 - - - - - - - -
Zhang - - - - - - - 76.4 - - 79.0 -
TUBFI - - - - 44.3 - - - - - - -
87.5 86.2 90.5 77.7 50.3 71.3 48.3 64.7 88.1 91.1 78.4 71.0
75.7 74.9 53.8 42.1 48.1 80.8 59.8 14.9 57.8 87.3 59.7 48.4
88.4 87.8 93.3 83.3 52.4 77.8 56.1 69.9 91.5 91.8 79.1 77.0
75.0 73.9 88.3 69.7 52.3 77.5 42.4 60.7 86.7 89.9 71.3 57.7
Oracle
(exhaustive) 90.0 89.4 95.3 86.1 55.7 84.9 67.5 77.3 94.7 92.0 80.8 78.6
Max evid.
(exhaustive) 90.0 89.4 95.3 86.1 55.5 84.7 67.5 77.3 94.5 92.0 80.8 78.6
Ours
(greedy) 90.0 89.4 95.3 86.1 55.5 84.7 67.5 77.3 94.5 92.0 80.8 77.8
Table 3: Comparison to existing methods in the 12 benchmark datasets. The best ensembles identified by maximizing evidence through exhaustive search mostly coincide with the oracle combinations—the ones with the highest accuracy in test set, which is also found by exhaustive search. The ensembles identified by our greedy search are very similar to the ones by these exhaustive search methods, and our algorithm consequently performs best in many tested datasets. We used three scales as done in [25], where we simply averaged the prediction scores from three scales.

6 Conclusion

We described a simple and efficient technique to transfer deep CNN models pre-trained on specific image classification tasks to another tasks. Our approach is based on Bayesian LS-SVM, which combines Bayesian evidence framework and SVM with a least squares loss. In addition, we presented a faster fixed point update rule for evidence maximization through Aitken’s delta-squared process. Our fast Bayesian LS-SVM demonstrated competitive results compared to the standard SVM by selecting a deep CNN model in 12 popular visual recognition problems. We also achieved the state-of-the-art performance by identifying a good ensemble of the candidate models through our Bayesian LS-SVM framework.

Acknowledgements

This work was partly supported by Institute for Information & Communications Technology Promotion (IITP) grant funded by the Korea government (MSIP) [B0101-16-0307; Basic Software Research in Human-level Lifelong Machine Learning (Machine Learning Center), B0101-16-0552; Development of Predictive Visual Intelligence Technology (DeepView)], and National Research Foundation (NRF) of Korea [NRF-2013R1A2A2A01067464].

References

  • [1] A. C. Aitken. On Bernoulli’s numerical solution of algebraic equations. Proceedings of the Royal Society of Edinburgh, 46:289–305, 1927.
  • [2] H. Azizpour, A. S. Razavian, J. Sulivan, A. Maki, and S. Carlsson. From generic to specific deep representations for visual recognition. In CVPR Workshops, 2015.
  • [3] A. Binder, W. Samek, M. Kloft, C. Müller, K.-R. Müller, and M. Kawanabe. The joint submission of the TU Berlin and Fraunhofer FIRST (TUBFI) to the ImageCLEF2011 photo annotation task. 2011.
  • [4] C. M. Bishop. Neural Networks for Pattern Recognition. Clarendon press Oxford, 1995.
  • [5] L. D. Bourdev, S. Maji, and J. Malik. Describing people: A poselet-based approach to attribute classification. In ICCV, 2011.
  • [6] K. Chatfield, K. Simonyan, A. Vedaldi, and a. Zisserman. Return of the devil in the details: Delving deep into convolutional nets. In BMVC, 2014.
  • [7] J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, n. Zhang, E. Tzeng, and T. Darrell. DeCAF: A deep convolutional activation feature for generic visual recognition. In ICML, 2014.
  • [8] M. Everingham, L. V. Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The PASCAL Visual Object Classes Challenge 2007 (VOC 2007) Results, 2007.
  • [9] M. Everingham, L. V. Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The PASCAL Visual Object Classes Challenge 2012 (VOC 2012) Results, 2012.
  • [10] R. E. Fan, K. W. Chang, C. J. Hsieh, X. R. Wang, and C. J. Lin. LIBLINEAR: A library for large linear classification. JMLR, 9:1871–1874, 2008.
  • [11] A. Farhadi, I. Endres, D. Hoiem, and D. A. Forsyth. Describing objects by their attributes. In CVPR, 2009.
  • [12] L. Fei-Fei, R. Fergus, and P. Perona. Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. CVIU, 106(1):59–70, 2007.
  • [13] R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object detection and semantic segmentation. In CVPR, pages 580–587. IEEE, 2014.
  • [14] G. Griffin, A. Holub, and P. Perona. Caltech-256 object category dataset. Technical report, California Institute of Technology, 2007.
  • [15] B. H. H. Nam. Learning multi-domain convolutional neural networks for visual tracking. In CVPR, 2016.
  • [16] B. H. H. Noh, S. Hong. Learning deconvolution net- work for semantic segmentation. In ICCV, 2015.
  • [17] A. Krizhevsky, I. Sutskever, and G. E. Hinton. ImageNet classification wit deep convolutional neural networks. In NIPS, volume 25, 2012.
  • [18] D. J. C. MacKay. Bayesian interpolation. Neural Computation, 4(3):415–447, 1992.
  • [19] M.-E. Nilsback and A. Zisserman. Automated flower classification over a large number of classes. In Proceedings of the Indian Conference on Computer Vision, Graphics and Image Processing, 2008.
  • [20] S. Nowak, K. Nagel, and J. Liebetrau. The CLEF 2011 photo annotation and concept-based retrieval tasks. In CLEF Workshop Notebook Paper, 2011.
  • [21] M. Oquab, L. Bottou, I. Laptev, and J. Sivic. Learning and transferring mid-level image representations using convolutional neural networks. In CVPR, 2014.
  • [22] A. Quattoni and A. Torrabla. Recognizing indoor scenes. In CVPR, 2009.
  • [23] A. S. Razavian, H. Azizpour, J. Sullivan, and S. Carlsson. CNN features off-the-shelf: An astounding baseline for recognition. In CVPR Workshops, 2014.
  • [24] P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun. OverFeat: Integrated recognition, localization and detection using convolutional networks. In ICLR, 2014.
  • [25] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In ICLR, 2015.
  • [26] J. A. K. Suykens and J. Vandewalle. Least squares support vector machine classifiers. Neural Processing Letters, 9(3):293–300, 1999.
  • [27] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, D. A. S. Reed, D. Erhan, V. Vanhoucke, and A. Rabinovich. Going deeper with convolutions. In CVPR, 2015.
  • [28] T. Van Gestel, J. A. K. S. B. Baesems, S. Viaene, J. Vanthienen, G. Dedene, B. De Moor, and J. Vandewalle. Benchmarking least squares support vector machines classifiers. Machine Learning, 54(1):5–32, 2004.
  • [29] T. Van Gestel, J. A. K. Suykens, G. Lanckrie, A. Lambrechts, B. D. Moor, and J. Vandewalle. Bayesian framework for least-squares support vector machine classifiers, gaussian processes, and kernel fisher discriminant analysis. Neural Computation, 14(5):1115–1147, 2002.
  • [30] C. Wah, S. Branson, P. Welinder, P. Perona, and S. Belongie. The Caltech-UCSD Birds-200-2011 dataset. Technical report, California Institute of Technology, 2011.
  • [31] Z. Wu, Y. Zhang, F. Yu, and J. Xiao. A GPU implementation of GoogLeNet. Technical report, Princeton University, 2014.
  • [32] J. Xiao, J. Hays, K. A. Ehinger, A. Oliva, and A. Torrabla. SUN database: Large-scale scene recognition from abbey to zoo. In CVPR, 2010.
  • [33] B. Yao, X. Jiang, A. Khosla, A. L. Lin, L. J. Guibas, and L. Fei-Fei. Action recognition by learning bases of action attributes and parts. In ICCV, 2011.
  • [34] M. D. Zeiler and R. Fergus. Visualizing and understanding convolutional networks. In ECCV, 2014.
  • [35] N. Zhang, , M. Paluri, M. Ranzato, T. Darrell, and L. Bourdev. PANDA: Pose aligned networks for deep attribute modeling. In CVPR, 2014.
  • [36] N. Zhang, J. Donahue, R. Girshick, and T. Darrell. Part-based R-CNNs for fine-grained category detection. In ECCV, 2014.
  • [37] P. Zhang and J. Peng. SVM vs regularized least squares classification. In ICPR, 2004.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
""
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
   
Add comment
Cancel
Loading ...
4622
This is a comment super asjknd jkasnjk adsnkj
Upvote
Downvote
""
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters
Submit
Cancel

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test
Test description