FAME: Face Association through Model Evolution

Face Association through Model Evolution

Eren Golge
Bilkent University
06800 Ankara/Turkey
   Pinar Duygulu
Bilkent University
06800 Ankara/Turkey

We attack the problem of learning face models for public faces from weakly-labelled images collected from web through querying a name. The data is very noisy even after face detection, with several irrelevant faces corresponding to other people. We propose a novel method, Face Association through Model Evolution (FAME), that is able to prune the data in an iterative way, for the face models associated to a name to evolve. The idea is based on capturing discriminativeness and representativeness of each instance and eliminating the outliers. The final models are used to classify faces on novel datasets with possibly different characteristics. On benchmark datasets, our results are comparable to or better than state-of-the-art studies for the task of face identification.

1 Introduction

To label faces of friends in social networks or celebrities and politicians in news, automatic methods are indispensable to manage large number of face images piling up on the web. On the other hand, unlike their counterparts in controlled datasets, faces on the web inherit all type of challenges naturally, resulting in the traditional methods incapable to recognise.

Recent availability of real-world face datasets [6, 23] accelerated the works in web-scale face verification, that is given a pair of faces deciding their identity. On the other hand, identification, that is finding the identity of a face, is still relatively less studied for the real-world faces. The requirement for a considerable amount of faces to be labeled is the main bottleneck for scalability in identification. Continuous inclusion of new individuals, and new instances for each individual should also be considered for web-scale identification task.

In this study, we challenge the identification of faces for famous people. The famous people tend to change their make-up, hair style/colour, and accessories more often compared to regular people, resulting in large number of varieties in face images. Moreover, they are likely to appear with others in photographs, causing faces of irrelevant people to be retrieved.

We propose a new method, FAME, that utilises the noisy results obtained through a name query to construct models in identifying famous people. Our models evolve through consecutive iterations to associate the query name with the correct set of faces. These models are then used to label faces on novel datasets. FAME removes the outlier faces in constructing models, while retaining the diversity as much as possible. Details of FAME will follow the review of recent work on relevant domains.

2 Related Work

Naming faces using weakly-labeled data: The work of Berg et al. is one of the first attempts in labelling large number of faces from weakly-labeled web images [6, 5] with the ”Labeled Faces in the Wild” (LFW) dataset introduced. It is assumed that in an image at most one face can correspond to a name, and names are used as constraints in clustering faces. Appearances of faces are modelled through Gaussian mixture model with one mixture per name. In [6], k-PCA is used to reduce the dimensionality of the data and LDA is used for projection. Initial discriminant space learned from faces with a single associated name is used for clustering through a modified k-means. Better discriminants are then learned to re-cluster. In [5] face name associations are captured through an EM based approach.

For aligning names and faces in an (a)symmetric way, Pham et al. [32] cluster the faces using a hierarchical agglomerative clustering method. They use the constraint that faces in an image cannot be in the same cluster. They then use an EM based approach for aligning names and faces based on probability of reoccurrences. They use a 3D morphable model for face representation. They introduce the picturedness and namedness: the probability of a person being in the picture based on textual info, and being in the text based on visual info.

Ideally, there should be a single cluster per person. However, these methods are likely to produce clusters with several people mixed in, and multiple clusters for the same person.

In [29, 30], Ozkan and Duygulu consider the problem as retrieving faces for a single query name, and then pruning the set from the irrelevant faces. A similarity graph is constructed where the nodes are faces, and edges are the similarity between faces. With the assumption that the most similar subset of faces will correspond to the queried name, the densest component in the graph is sought using a greedy method. In [15], the method of [29, 30] is improved by introducing the constraint for each image to contain a single instance of the queried person and replacing the threshold in constructing the binary graphs with assigning non-zero weights to k nearest neighbours. The authors further generalised the graph based method for for multi-person naming, as well as null assignments. They propose a min-cost max-flow based approach to optimise face name assignments under unique matching constraints.

In [18] face-name association problem is tackled as a multiple instance learning problem over pairs of bags. Detected faces in an image is put into a bag, and names detected in the caption are put into the corresponding set of labels. A pair of bags is labeled as positive if they share at least one label, and negative otherwise. The results are reported on Labelled Yahoo! News dataset which is obtained through manually annotating and extending LFW dataset. In [16], it is shown that the performance of graph-based and generative approaches for text-based face retrieval and face-name association tasks can be improved with the incorporation of logistic discriminant based metric learning (LDML) [17].

Kumar et al. [23] introduced attribute and smile classifiers for verifying the identity of faces. For describable aspects of visual appearance, binary attribute classifiers are trained with the help of AMT. Moreover, simile classifiers are trained to recognise the similarity of faces to specific reference people. Pub-Fig, dataset of public figures on the web, is presented alternative to LFW with larger number of individuals each having more instances.

Recently, PubFig83, a subset of PubFig dataset with near-duplicates eliminated and individuals with large number of instances are selected, is provided for face identification task [33]. Inspired from biological systems, Pinto et al. [33] consider V1-like features and introduce both single- and multi-layer feature extraction architecture followed by LinearSVM classifier.

[27] define the open-universe face identification problem as identifying faces with one of the labeled categories in a dataset including distractor faces that do not belong to any of the labels. In [2], the authors combine PubFig83, as being the set of labeled individuals, and LFW, as being the set of distractors. On this set, they evaluate a set of identification methods including nearest neighbour, SVM, sparse representation based classification (SRC) and its variants, as well as linearly approximated SRC that they proposed in [27].

Other recent work include [37] where Fisher vectors on densely sampled SIFT features are utilised. Large margin dimensionality reduction is used to reduce high dimensionality.

Harvesting web for concept learning: Recently, there have been many studies on harvesting web for re-ranking of search results and building qualified training sets [14, 7, 4, 13, 24, 35, 8]. In [7] visual features and surrounding the text are used for collecting animal images from web, and visual exemplars are obtained through clustering text. Relevant clusters are required to be identified manually, as well as irrelevant images in clusters. In [24], OPTIMOL framework is presented to incrementally learn object categories from web search results. Given a set of seed images a non parametric latent topic model is applied to categorise collected web images. The model is iteratively updated with the newly categorised images. To prevent over specialised results, a set of cache images with high diversity are retained at each iteration. In [35] after the removal of abstract images from the search results collected through text and image search, text and metadata are used to re-rank the images. A visual classifier is trained by sampling from the top ranked images as positives and random images from other categories as negatives. Recently, NEIL [8] is proposed to learn object and scene categories, as well as common sense knowledge using web search results. Discovering representative and discriminative instances: Our method is also related to the recently emerged studies in discovering discriminative patches.  [25, 22, 38, 11, 10, 20, 12, 21]. In [38], discriminative patches in images are discovered through an iterative method which alternates between clustering and training discriminative classifiers. Li et al. [25] solves same problem with multiple instance learning. [11] and [21] apply the idea to scene images for learning discriminative properties by embracing the unsupervised exemplar models. Moreover [10] enhances the unsupervised learning schema by more robust alternation of Mean-Shift clusteringalgorithm. Disciminative patch ideas is also applied to video domain by [20].

3 Our approach

An important caveat in learning models from weakly-labelled data is the impurity of the collection. To be useful, spurious instances should be eliminated before generating models for each category. In this study, we present a new approach for learning better models through iteratively pruning the data (see Figure1). First, we benefit from large number of global negatives representing the rest of the world against the class of interest. Next, among the candidate in-class examples we try to separate the most confident instances from the others. These two successive steps are repeated to eliminate outlier instances iteratively. To consider intra-class variability, we use a representation that results in large dimensional feature vectors to make each class linearly separable even when the data include some level of variation. The model evolution and representation will be detailed in the following.

Figure 1: Overview of the proposed method.

3.1 Model Evolution

We propose a method that allows the models to evolve through eliminating the outlier instances with successive linear classifiers. First, we learn a hyperplane that separates the initial set of candidate class instances from the large set of global negatives. Global negative set is curated by the instances of other classes and the random face images collected from Web. Then, we select some fraction of the class instances that are distant from the separating hyperplane. We use these instances as the discriminative seed set, since they are confidently classified against the rest of the world. We consider the rest of the class data as possible negatives. We then learn another model that try to capture in-class dissimilarities between discriminative examples and possible negatives. At the final step, we combine the confidence scores of the first and the second models. By combining the two scores, that respectively correspond to the confidence of being different from the rest of the world, and in-class affinity of the instance, we get a measure of instance saliency. Over these confidence scores we detect instances with the lowest scores as the outliers for that iteration. These steps are iterated multiple times up to a desired level of pruning. The representation that we use (see Section3.2) might cause computational burden with complicated learning models. Therefore, we leverage simple linear regression (LR) models with L1 norm regularisation performing sparse feature selection as the learning evolves. Sparsity makes categories more distinct and captures category related commonalities.

Algorithm 1 summarises our data elimination procedure. refers to the examples collected for a class and refers to the the vast numbers of global negatives. Each vector is a dimensional representation of a single face image. At each iteration , the first LR model learns a hyperplane between the candidate set of class instances and global negatives . Then the current is divided into two subsets: instances in that are farthest from the hyperplane are kept as the candidate positive set () and the rest is considered as the negative set () for the next model. is the set of salient instances for the class and is the set of possible spurious instances. The second LR model uses as positive and as the negative set to learn best possible hyperplane separating them. For each instance in , by aggregating the confidence values of both models, we eliminate instances with the lowest scores as the outliers. At the next iterations, we run all the steps again and end up with a clean set of class instances .

This iterative procedure continues until it satisfies a stopping condition which is refined by ’s training accuracy as the measure of present data quality. As we incrementally remove poor instances, we expect to have better separation against the negative instances therefore ’s accuracy increases. However, if the accuracy saturates or degrade then we stop the algorithm. Alternatively, when we have very large number of class instances, we can divide data into two independent subset and apply the iterative elimination to both as we measure the quality of one set’s over the other set’s at each iteration . It is similar to co-training approach and more robust to over-fitting, albeit it requires very large number of instances for convincing results.

1 In the real code we use vectorized implementation whereas we write down iterative pseudo-code for the favour of simplicity.
Input: , , ,
2 while do
3        )
4 end while
Algorithm 1 FAME

3.2 Representation

To represent face images we learn two distinct set of filters by an unsupervised method similar to [9] (see Figure2 ). First set is learned from the raw-pixel random patches extracted from grey-scale images. The second set is learned from LBP [1] encoded images. First set of learned filters are receptive to edge- and corner-like structural points and the second set is more sensitive to textural commonalities of the LBP histogram statistics.LBP encoded images are invariant to illumination since the intensity relations between pixels are considered instead of exact pixel values. We use rotation invariant LBP encoding [26] that gives binary codes for each pixel. Then, we convert these binary codes into corresponding integer values. A Gaussian filter is used to smooth out the heavy-tailed locations.

Figure 2: Random set of filters learned from (a) whitened raw image pixels, and (b) LBP encoded images. (c) Outlier filters of raw-image filters. (d) LBP encoding of a given RGB image. We might observe eye or mount shaped filters from the raw image filters and more textural information from the LBP encoded filters. Outlier filters are very cluttered and observe low number of activations mostly from background patches.

The pipeline in order to learn filters from both raw-pixel and LBP images is as follows. First we extract a set of randomly sampled patches in the size of predefined receptive field. Then contrast normalisation is applied to each patch (for only raw-image filters) and patches are whitened to reduce the correlations among dimensions. These patches are clustered using k-means into groups. We perform thresholding to centroids with box-plot statistics over the activations counts to remove the outlier centroids that are supposedly not representative for the face images but background clutters. After the learning phase, centroid activations are collected from receptive fields with small striding. We applied spatial average pooling onto five different grids including a grid at the center of the image additional to 4 equal-sized quadrants since face images includes important spatial regularities at the center. We use triangular activation function to map each receptive field to learned centroids. This yields a dimensional representation for each face. However, since we use two different set of filters, at the end, each image presented by dimensions. Thresholding of centroid activations provides a implicit removal of outlier patches as well as the salient set of centroids. We use those outlier centroids to eliminate patches at the feature extraction step by assuming the patches assigned to outlier centroids are not relevant thus avoiding them in pooling.

4 Experiments

4.1 Datasets

Images are collected using Bing to train models. Then, two recent benchmark datasets, FAN-large [28] and PubFig83[33], are used for testing.Bing collection:For a given name, 500 images are gathered using Bing image search 111\(https://www.bing.com/\). Categories are chosen as the people having more than 50 annotated face images in FAN-large or PubFig83 datasets. In total, 226691 images are collected corresponding to 365 name categories in FAN-large, and 83 name categories in PubFig83. Additional 2500 face images for queries ”female face”, ”male face”, ”face images” are collected to construct the global negatives. Face detector of [40] is used for detecting faces. Only the most confident detection is selected from each image to be put into the initial pool of faces associated with the name (on the average 450 faces per category). Other detections are added to global negatives.Test collection:We use two sets from FAN-large face dataset [28]: EASY and ALL. EASY subset includes faces larger than 60x70 pixels. ALL includes all names without any size constraint. We use 138 names from EASY, and 365 from ALL subsets, with 23952 and 199295 images respectively. On the average there are 541 images for each name. We also use PubFig83[33] dataset, which is the subset of well-known PugFig dataset with 83 different celebrities having at least 100 images. PubFig83 is more convenient set for face identification problem with near-duplicate images and the ones that are no longer available at Internet are removed[3]. We shaped a controlled test environment by using PubFig83+LFW [3]: extending PubFig83 with some distract images from LFW [19] not belonging to any of the selected categories (distractors are six percent of correct instances).We use these distract images to extend our global negatives. For the controlled experiment, we select name categories with more than 270 images and mixed them with random set of distract images.Then we apply full stack of FAME with 5-fold cross-validation.

4.2 Implementation Details

The dataset is expanded with horizontally flipped images. Before learning filters from raw-pixel images, each grey-level face image is resized to 60 pixels height and LBP images resized to 120 pixels height. LBP encoding has been done by 16 different filter orientation and at radius 2. We sample random patches from images and apply contrast normalization to only raw-pixel patches. Then, we perform ZCA whitening transform and set to 0.5. We use receptive field of 6x6 regions with 1 stride and learn 2400 centroids for both raw-pixel images and LBP encoded images. Hence, we conclude 2 (raw-pixel + LBP) x 5 (pooling grids) x 2400 (centroids) dimensional feature representation of each image. For instance to centroid distances we used Euclidean Distance. We detect the outliers by a threshold at the 99% upper whisker of the centroid activations. Our implementation of feature learning framework aggregated upon the code furnished by [9]. For iterative elimination, we train L1 norm Logistic Regression model with Gauss-Seidel algorithm [36] and final classification is done with Linear SVM through grafting algorithm [31] that learns sparse set of important features incrementally by using gradient information. At each FAME iteration we eliminate five images. We stop when there is no improvement on the first model accuracy. If the classifier saturates so quickly, iteration continues until 10% of the instances are pruned. If we encounter memory constraints due to large number of global negatives, at each iteration we sample a different set of negative instances, to provide slightly different linear boundaries that are able to detect different spurious instances.

4.3 Evaluations

We conduct controlled experiments over PubFig83+LFW. We select classes with at least 270 instances and inject 10% (27 instances) noise instances. There are six classes conforming that criterion. Noisy images are randomly chosen from global negatives consisting of ”distract” set of PubFig83+LFW and FAN-large faces that we collected. As a result, we have 297x6 training instances. We apply FAME to this data while applying cross-validation at each iteration step, between these six classes.

Figure3 helps to visualise the model evolution in FAME. As shown on the left, at each iteration dataset is divided into candidate positives and possible negatives: candidate positives are selected as the most representative instances of the class and true outliers are found among the possible negatives. As shown on the right, FAME is able to learn models from noisy weakly label sets, while eliminating the outliers at successive iterations for a variety of people.

Figure 3: (Top:) Some of the instances selected for , and for iterations . (Bottom:) Samples for final model faces and outliers found in the first two iterations.

As Figure4-(a) shows with the increasing number of iterations, more outliers are eliminated. Although some correct instances are also eliminated, the ratio is very low compared to the spurious instances. Moreover, our observations show that the eliminated positive examples are usually not in good quality and therefore their elimination from the final model is not harmful but rather helpful as supported with the results in Figure4-(b). As seen in Figure4-(c) , we can achieve up to 75.2 on FAN-Large (EASY) and 79.8 on PubFig83 by removing one outlier at each iteration: we prefer to eliminate five outliers for the efficiency.

Figure 4: Evaluations on PubFig83 dataset. (a) Total number of correct versus false outlier detections until FAME finds all the outliers for all the classes. We stop FAME for the saturated classes before the end of the plot. (b) Cross-validation and M1 accuracies as the algorithm proceeds. This shows the salient correlation between cross-validation classifier and the M1 models, without M1 models incurring over-fitting. (c) Effect of number of outliers removed at each iteration.
Feature Accuracy LBP filters 60.7 raw-pixel filters 71.6 LBP+raw-pixel filters 79.3 Num. Centroids 1500 2000 2400 Accuracy(%) 84.9 88.60 90.75
Table 1: (Left:) This table compares the performances obtained with different features on PubFig83 dataset with the models learned from web. As the figure suggests, even LBP filters are not competitive with raw-pixel filters, its textural information is subsidiary to raw-pixel filters with increasing performance. (Right:) Accuracy versus number of centroids .

We compare FAME with baseline method that learns models from the raw collection gathered through querying the name without any pruning. As seen in Table2 with one vs all L1 norm Linear SVM model on the raw data, the performance is very low on all datasets. Note that, on the datasets FAN-Large EASY and ALL, as well as PubFig83, we learn the models from web images and tested them on these novel datasets for the same categories. We also divided the collected Bing images into two subsets to test the effect of training and testing on the same type of dataset. FAME leads encouraging results even the model is susceptible to domain shifting problem, with a significant improvement over baseline.

The most similar data handling approach to ours is the method of Singh et al. [39], although there are important differences. First, [39] clusters the data to capture intra-cluster variance and uncover the representative instances. However, it requires to decide the optimal cluster number in advance and divides the problem into multiple homologous pieces which need to be solved separately. This increase the complexity of the proposed system.

Second difference lies in the philosophy. They aim to discover representative and discriminative set of instances whereas we aim to prune spurious ones. Hence, they need to keep all vast negative instances on memory but we can sample different subsets of global negatives and find corresponding outlier instances. It provides faster and easier way of data pruning. They divide each class into two sets and apply their scheme by interchanging data after each iteration like in the case of co-training learning procedure. Nevertheless, co-training demands large number of instances for reliable results. In our methodology, we prefer to use all the class data at once in our particular scheme. We evaluate the method of Sing et al. on the same datasets, and show that FAME is superior to their method (see Table2). We use the released code by Singh et al. [39] with up-limit settings that our resources allow.

To test the effectiveness of the proposed linear regression based model learning, we also compare our results by using only the model (FAME-M1) and using SVM for classification (FAME-SVM). As shown in Table2, all FAME models outperforms the baseline method as well as the method of [39] with a large improvement with the proposed LR model.

- Bing FAN-Large (EASY) FAN-Large (ALL) PubFig83
Baseline 62.5 56.5 52.7 52.8
Singh et al. [39] 74.7 65.9 62.3 71.4
FAME-M1 78.6 68.3 60.2 71.7
FAME-SVM 81.4 73.1 65.4 76.8
FAME-LR 83.7 74.3 67.1 79.3
Table 2: Accuracies (%) on FAN-Large [28] (EASY and ALL), PubFig83 and on the held-out set of our Bing data collection. There are three alternative FAME implementations. FAME-M1 uses only the model M1 which removes instances regarding global negatives. FAME-SVM uses SVM in training and FAME-LR is the proposed method using linear regression.

Finally, we compare the performance of FAME on the benchmark PubFig83 dataset with the other state-of-the-art studies on face identification. In this case, unlike the previous experiments where we learned the models from noisy images, in order to make a fair comparison we learned the models from the same dataset. As seen in Table3 FAME achieves the best accuracy in this setting. Referring back to Table2 even with the domain adaptation setting where the model is learned from the noisy web images our results are comparable to the most recent studies on face identification that train and test on the same dataset. Note that, the method of Pinto et al. [34] is similar to our classification pipeline but we prefer to learn the filters in an unsupervised way with the method of Coastes et al. [9]..

Method Pinto et al. [33] (S) Pinto et al.[33](M) face.com [33] Becker et al. [3]
Accuracy 75.6 87.1 82.1 85.9 90.75
Table 3: Accuracies (%) of face identification methods on PubFig83. [33] proposes single layer (S) and multi-layer (M) architectures. face.com API is also experienced in [33]. Note that, here FAME is learned from the same dataset.

5 Summary and future work

We propose a novel method to prune the web images collected for a query to learn models to be used for classification on novel datasets. We rely on large number of negative instances in selecting a set of good instances which are then used to learn models to eliminate the bad ones. The proposed method outperforms the baseline and is comparable to state-of-the-art methods even within the difficulties of domain adaptation. Although the proposed method is tested for identification of faces, it is a general method that could be used for other domains as we aim to attack as our future work.


  • [1] Timo Ahonen, Abdenour Hadid, and Matti Pietikainen. Face description with local binary patterns: Application to face recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(12):2037–2041, 2006.
  • [2] Brian C Becker and Enrique G Ortiz. Evaluating open-universe face identification on the web. In Computer Vision and Pattern Recognition Workshops (CVPRW), 2013.
  • [3] Brian C Becker and Enrique G Ortiz. Evaluating open-universe face identification on the web. In Computer Vision and Pattern Recognition Workshops (CVPRW), 2013 IEEE Conference on, pages 904–911. IEEE, 2013.
  • [4] Tamara L Berg and Alexander C Berg. Finding iconic images. In Computer Vision and Pattern Recognition Workshops, 2009.
  • [5] Tamara L. Berg, Alexander C. Berg, Jaety Edwards, and David A. Forsyth. Who’s in the picture?
  • [6] Tamara L. Berg, Alexander C. Berg, Jaety Edwards, Michael Maire, Ryan White, Yee Whye Teh, Erik Learned-Miller, and David A. Forsyth. Names and faces in the news.
  • [7] Tamara L. Berg and David A. Forsyth. Animals on the web.
  • [8] Xinlei Chen, Abhinav Shrivastava, and Abhinav Gupta. Neil: Extracting visual knowledge from web data. In International Conference on Computer Vision (ICCV), 2013.
  • [9] Adam Coates, Andrew Y Ng, and Honglak Lee. An analysis of single-layer networks in unsupervised feature learning. In International Conference on Artificial Intelligence and Statistics, pages 215–223, 2011.
  • [10] Carl Doersch, Abhinav Gupta, and Alexei A Efros. Mid-level visual element discovery as discriminative mode seeking. In Advances in Neural Information Processing Systems, pages 494–502, 2013.
  • [11] Carl Doersch, Saurabh Singh, Abhinav Gupta, Josef Sivic, and Alexei A Efros. What makes paris look like paris? ACM Transactions on Graphics (TOG), 31(4):101, 2012.
  • [12] Ian Endres, Kevin J Shih, Johnston Jiaa, and Derek Hoiem. Learning collections of part models for object recognition. In Computer Vision and Pattern Recognition (CVPR), 2013 IEEE Conference on, pages 939–946. IEEE, 2013.
  • [13] Jianping Fan, Yi Shen, Ning Zhou, and Yuli Gao. Harvesting large-scale weakly-tagged image databases from the web. In Computer Vision and Pattern Recognition (CVPR), 2010.
  • [14] Robert Fergus, Li Fei-Fei, Pietro Perona, and Andrew Zisserman. Learning object categories from google’s image search. In International Conference on Computer Vision (ICCV), 2005.
  • [15] Matthieu Guillaumin, Thomas Mensink, Jakob Verbeek, and Cordelia Schmid. Automatic Face Naming with Caption-based Supervision. In Computer Vision and Pattern Recognition (CVPR).
  • [16] Matthieu Guillaumin, Thomas Mensink, Jakob Verbeek, and Cordelia Schmid. Face recognition from caption-based supervision. International Journal of Computer Vision, 96(1):64–82, January 2012.
  • [17] Matthieu Guillaumin, Jakob Verbeek, and Cordelia Schmid. Is that you? Metric learning approaches for face identification. In International Conference on Computer Vision (ICCV 2009), 2009.
  • [18] Matthieu Guillaumin, Jakob Verbeek, and Cordelia Schmid. Multiple instance metric learning from automatically labeled bags of faces. In European Conference on Computer Vision (ECCV) , 2010.
  • [19] Gary B Huang, Manu Ramesh, Tamara Berg, and Erik Learned-Miller. Labeled faces in the wild: A database for studying face recognition in unconstrained environments. Technical report, Technical Report 07-49, University of Massachusetts, Amherst, 2007.
  • [20] Arpit Jain, Abhinav Gupta, Mikel Rodriguez, and Larry S Davis. Representing videos using mid-level discriminative patches. In Computer Vision and Pattern Recognition (CVPR), 2013 IEEE Conference on, pages 2571–2578. IEEE, 2013.
  • [21] Mayank Juneja, Andrea Vedaldi, CV Jawahar, and Andrew Zisserman. Blocks that shout: Distinctive parts for scene classification. In Computer Vision and Pattern Recognition (CVPR), 2013 IEEE Conference on, pages 923–930. IEEE, 2013.
  • [22] Gunhee Kim and Antonio Torralba. Unsupervised detection of regions of interest using iterative link analysis. In NIPS, volume 1, pages 4–2, 2009.
  • [23] N. Kumar, A. C. Berg, P. N. Belhumeur, and S. K. Nayar. Attribute and Simile Classifiers for Face Verification. In International Conference on Computer Vision (ICCV), 2009.
  • [24] Li-Jia Li and Li Fei-Fei. Optimol: automatic online picture collection via incremental model learning. International journal of computer vision, 88(2):147–168, 2010.
  • [25] Quannan Li, Jiajun Wu, and Zhuowen Tu. Harvesting mid-level visual concepts from large-scale internet images. CVPR, 2013.
  • [26] Timo Ojala, Matti Pietikäinen, and Topi Mäenpää. Gray scale and rotation invariant texture classification with local binary patterns. In Computer Vision-ECCV 2000, pages 404–420. Springer, 2000.
  • [27] Enrique G. Ortiz and Brian C. Becker. Face recognition for web-scale datasets. Computer Vision and Image Understanding, 118:153–170, January 2014.
  • [28] Mert Özcan, Jie Luo, Vittorio Ferrari, and Barbara Caputo. A large-scale database of images and captions for automatic face naming. In BMVC, pages 1–11, 2011.
  • [29] Derya Ozkan and Pinar Duygulu. A graph based approach for naming faces in news photos.
  • [30] Derya Ozkan and Pinar Duygulu. Interesting faces: A graph based approach for finding people in news. Pattern Recognition, 43(5):1717–1735, May 2010.
  • [31] Simon Perkins, Kevin Lacker, and James Theiler. Grafting: Fast, incremental feature selection by gradient descent in function space. The Journal of Machine Learning Research, 3:1333–1356, 2003.
  • [32] P.T. Pham, M.F. Moens, and T. Tuytelaars. Cross-media alignment of names and faces. IEEE Transactions on Multimedia, 12(1):13–27, 2010.
  • [33] Nicolas Pinto, Zak Stone, Todd Zickler, and David Cox. Scaling up biologically-inspired computer vision: A case study in unconstrained face recognition on facebook. In Computer Vision and Pattern Recognition Workshops (CVPRW), 2011 IEEE Computer Society Conference on, pages 35–42. IEEE, 2011.
  • [34] Nicolas Pinto, Zak Stone, Todd Zickler, and David Cox. Scaling up biologically-inspired computer vision: A case study in unconstrained face recognition on facebook. In Computer Vision and Pattern Recognition Workshops (CVPRW), 2011.
  • [35] Florian Schroff, Antonio Criminisi, and Andrew Zisserman. Harvesting image databases from the web. IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(4):754–766, 2011.
  • [36] Shirish Krishnaj Shevade and S Sathiya Keerthi. A simple and efficient algorithm for gene selection using sparse logistic regression. Bioinformatics, 19(17):2246–2253, 2003.
  • [37] K. Simonyan, O. M. Parkhi, A. Vedaldi, and A. Zisserman. Fisher Vector Faces in the Wild. In British Machine Vision Conference (BMVC), 2013.
  • [38] Saurabh Singh, Abhinav Gupta, and Alexei A. Efros. Unsupervised discovery of mid-level discriminative patches. In European Conference on Computer Vision (ECCV), 2012.
  • [39] Saurabh Singh, Abhinav Gupta, and Alexei A Efros. Unsupervised discovery of mid-level discriminative patches. In European Conference Computer Vision (ECCV). 2012.
  • [40] Xiangxin Zhu and Deva Ramanan. Face detection, pose estimation, and landmark localization in the wild. In Computer Vision and Pattern Recognition (CVPR), 2012.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
Add comment
Loading ...
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test description