Similarity-based text recognition by Deeply Supervised Siamese Network

Similarity-based text recognition by
Deeply Supervised Siamese Network

Ehsan Hosseini-Asl
Electrical and Computer Engineering
University of Louisville
Louisville, KY 40292, USA
ehsan.hosseiniasl@louisville.edu
\AndAngshuman Guha
Captricity, Inc.
Oakland, CA 94612 USA
angshumang@captricity.com
The work was performed during an internship at Captricity, Inc.
Abstract

In this paper, we propose a new text recognition model based on measuring the visual similarity of text and predicting the content of unlabeled texts. First a Siamese convolutional network is trained with deep supervision on a labeled training dataset. This network projects texts into a similarity manifold. The Deeply Supervised Siamese network learns visual similarity of texts. Then a K-nearest neighbor classifier is used to predict unlabeled text based on similarity distance to labeled texts. The performance of the model is evaluated on three datasets of machine-print and hand-written text combined. We demonstrate that the model reduces the cost of human estimation by . The error of the system is less than . The proposed model outperform conventional Siamese network by finding visually-similar barely-readable and readable text, e.g. machine-printed, handwritten, due to deep supervision. The results also demonstrate that the predicted labels are sometimes better than human labels e.g. spelling correction.

Similarity-based text recognition by
Deeply Supervised Siamese Network

Ehsan Hosseini-Aslthanks: The work was performed during an internship at Captricity, Inc.
Electrical and Computer Engineering
University of Louisville
Louisville, KY 40292, USA
ehsan.hosseiniasl@louisville.edu
Angshuman Guha
Captricity, Inc.
Oakland, CA 94612 USA
angshumang@captricity.com

1 Introduction

Optical Character Recognition (OCR) is traditionally used to convert images of machine printed text into textual content. Intelligent Character Recognition (ICR) is used to do the same for images of handwritten text. State-of-the-art OCR engines can work well, but only for clean data and where the OCR engine can be adjusted to deal with a single font or a small set of fonts. State-of-the-art ICR engines are significantly worse.

For a real-life application of high-accuracy character recognition involving both machine print and handwriting, one has to develop one’s own OCR/ICR engine. This typically requires plenty of character-segmented data, as well as labeling at the character level. This is a very expensive proposition in most real-world situations, if not an impossible one. To avoid the character segmentation cost, Keeler et al. (1991) proposed learning character segmentation and recognition simultaneously from un-segmented data. This does not work well in practice beyond numeric characters and for large vocabularies. There has been some work at limited-vocabulary whole word recognition, see, for example,  Lavrenko et al. (2004). To avoid character segmentation, LeCun et al. (1998) proposed a graph transformer network with Viterbi search. These kinds of models cannot compete in performance (training time) with modern deep neural nets that afford efficient implementations, for example, using GPU. Bunke et al. (2004) handled the segmentation problem using HMM-based recognition using the Viterbi algorithm. The present authors have experienced (convolutional) neural nets consistently out-performing HMMs in at least two domains, both with large quantities of industrial data: online handwriting recognition and speech recognition. Recently there also has been work involving innovative models where characters are predicted for an input word of a fixed-size input image. For instance, one model might predict the first character, another model might predict the second character, and so on. See for example, Jaderberg et al. (2014).

Figure 1: Examples of similar text.

Our goal is for a practical system that organically incorporates human labeling with machine learning to achieve very low error rates () while minimizing the amount of necessary human labeling.

Siamese Network (SN) is a type of end-to-end metric learning approach, consisting of a neural network that learns a discriminative function. SNs are trained by learning a similarity metric between pairs of data. SN models are applied to signature verification Bromley et al. (1993), digit recognition Hadsell et al. (2006), face recognition Chopra et al. (2005), Speech feature extraction Chen & Salman (2011), and Speech keyword detection Grangier & Bengio (2007). In this paper, we propose and discuss a novel method of text recognition that does not require character-segmented data. by predicting the content of a text using similar labeled texts. We use a Siamese Convolutional Network to learn the similarity between text images in a low-dimensional Euclidean space. To account for similar machine-printed and handwritten text, the SN is regularized by supervision in hidden layers Weston et al. (2012); Lee et al. (2014). Then a k-nearest neighbor algorithm is employed to predict the label based on most similar labeled texts. To account for unseen labels in test data, an interactive k-nearest neighbor algorithm with human annotation is employed for label prediction, and reducing the error. See Fig. 1 for examples of “similar” images i.e. with the same text content.

2 Proposed Model

In this section, we propose a model of text recognition based on learning similarity of images. In section 2.1, a Siamese network is proposed for learning similarity of text, and section 2.2 describes a text recognition framework.

2.1 Learning Text Similarity

To train a model to be able to learn the similarity between texts, a Siamese network is used as in Chopra et al. (2005); Hadsell et al. (2006). The Siamese network is trained to project the images into a feature space, where similar images are projected with short mutual Euclidean distance, and dissimilar images are projected with large mutual Euclidean distances. Training of the Siamese network is based on minimizing the contrastive loss of a pair of images,

(1)
Figure 2: Deeply Supervised Siamese Network (DSSN) for learning text similarity

where are the weights of the hidden layers and output layer of the Siamese network, is the label of paired images, i.e. 0 if similar and 1 if dissimilar, is the Euclidean distance between a pair of images, and is the desired Euclidean distance between a pair of dissimilar images. Siamese networks have shown promising results in learning similarity of the handwritten digits dataset, MNIST. However, in complicated cases of long text, capturing similarities between two texts is infeasible using a single loss function in the output layer of Siamese network. The performance of contrastive loss is dependent on feature extraction of the hidden layers, where it should capture the similarities in a hierarchical way, to enable the output layer to extract features which can clearly represent the similarities of long and complex text. In order to boost the performance of the Siamese network for learning similarity of long text, we used the method of deep supervision proposed in Lee et al. (2014), where several contrastive loss functions are used for hidden and output layers, to improve the discriminativeness of feature learning, as shown in Fig. 2. Therefore, the proposed method is called Deeply Supervised Siamese Network (DSSN) and it is trained using the combined contrastive loss,

(2)

where indicates the index for hidden layer, is the output layer. Eq. 2 indicates that the loss of each hidden layer is only the function of weights of that layer, i.e. . The DSSN network generates a Similarity Manifold, where similar texts are projected with short mutual Euclidean distances. The next section describes the text recognition model based on the Similarity Manifold. The ADADELTA method of gradient descent (Zeiler (2012)) is used to update the parameters of DSSN.

2.2 Text Recognition by Text Similarity

This section describes a text recognition framework to predict the label of text using the DSSN model developed in the previous section. The text recognition model is based on feature extraction of text using proposed DSSN network. We use a K-nearest neighbor algorithm to predict the label of text images in test data, based on similarity distance to the labeled text in train data. The predicted label is compared with human estimation as shown in Fig. 3(a).

(a) (b)
Figure 3: (a) Proposed Framework for text recognition based on text similarity. (b) Efficiency metric to measure the reduction in human estimation of label.

Our human-based model for text label prediction is based on voting of two humans on a text image. The proposed framework in Fig. 3 (a) is motivated by our goal of reducing the cost of human estimations while maintaining a low error rate. As shown in Fig. 3(a), the predicted label of DSSN-KNN is accompanied by a confidence value. We choose two parameters, and , such that the confidence value can be classified as highly confident, confident and not confident. If the model’s prediction confidence is high, we omit the required two human estimations. However, if the prediction is only confident, we validate the predicted label of DSSN-KNN with one human estimation. The parameters and are chosen by tuning the model’s performance on the training set (or one can use a validation set).

To measure the performance of DSSN-KNN in reducing the human estimation, we define an efficiency metric as shown in Fig. 3(b),

(3)

where is the total number of text samples, and are the number of confidently wrong and confidently correct predictions, and and are the number of high-confident wrong and high-confident correct predictions, respectively.

Note that the efficiency metric definition implicitly assumes a low rate of disagreement between two humans labeling the same image or between a human and the DSSN-KNN model. If this rate is (which is what we see in practice, see AC column in Table 4), the metric will overcount the reduction in the required number of human estimates by . In the case of disagreement, extra human estimates will be needed to resolve conflicts.

The DSSN-KNN model can be used in one of two modes: ROBOTIC and ASSISTIVE.

ROBOTIC mode is suggested by Fig. 3 – (i) for high confidence predictions, we skip human labeling, (ii) for medium confidence predictions, we ask for human confirmation and (iii) for low confidence predictions, we discard the prediction and ask for at least two human estimates.

ASSISTIVE mode is to ignore (high confidence threshold) – (i) for high and medium confidence predictions, we ask for human confirmation and (ii) for low confidence predictions, we discard the prediction and ask for at least two human estimates.

By definition, ASSISTIVE mode results in zero error from the DSSN-KNN model. But efficiency is lower because are folded into in the numerator of Fig. 3(b). On the other hand, ROBOTIC mode has higher efficiency at the cost of some DSSN-KNN errors unchecked by humans. We want this error to be under 0.5%.

Data: text
Output: label

1:Extract the feature of new text image from hidden layer of DSSN
2:Predict the label and confidence using k-nearest-neighbor algorithm
3:if mode=ROBOTIC then
4:     if  then
5:         Output=label
6:     else if  then
7:         verify the label with 1 human
8:         if DSSN-KNN and human disagree then
9:              get another human label
10:              if two humans agree then
11:                  Update KNN dictionary
12:                  Output=human label
13:              else
14:                  Output=label
15:              end if
16:         else
17:              Output=label
18:         end if
19:     else
20:         Label the text with 2 human
21:         Update KNN dictionary
22:         Output=human label
23:     end if
24:else // mode=ASSISTIVE
25:     if  then
26:         verify the label with 1 human
27:         if DSSN-KNN and human disagree then
28:              get another human label
29:              if two human agree then
30:                  Update KNN dictionary
31:                  Output=human label
32:              else
33:                  Output=label
34:              end if
35:         else
36:              Output=label
37:         end if
38:     else
39:         Label the text with 2 human
40:         Update KNN dictionary
41:         Output=human label
42:     end if
43:end if
Algorithm 1 Interactive Text Recognition by DSSN-KNN

3 Experiments

In this section, we design several experiments to evaluate the performance of our proposed model of text recognition. The DSSN-KNN model is pretrained on MNIST data and then fine-tuned on the datasets to minimize the loss function of Eq. 2. We select minibatch size of paired texts, containing similar pairs and dissimilar pairs, to train the Similarity manifold. We used Caffe,  Jia et al. (2014), and Theano,  Bastien et al. (2012), on Amazon EC2 g2.8xlarge instances with GPU GRID K520 for our experiments. First we apply some metrics to evaluate the performance of DSSN in learning the similarity manifold in section 3.1. Then the performance of DSSN-KNN is evaluated for text recognition of three hand-written text datasets in section 3.2.

3.1 Similarity Manifold Evaluation

In order to evaluate the performance of proposed DSSN for text recognition, we evaluated the trained similarity manifold for detecting similar and dissimilar texts. For this purpose, we implemented two separate experiments for non-numeric and numeric texts.

The non-numeric dataset contains 8 classes, where two major classes dominate in sample count. We found that most of the human-labeled ’blanks’ are actually not blank, and contain some text from the two major classes. This misclassified text in training data hurts the performance of DSSN.

(a) (b)
Figure 4: Similarity manifold visualization of machine-printed non-numeric texts in (a) hidden and (b) output layer, using t-SNE projection.

To investigate the distribution of text in the similarity manifold, the feature spaces of hidden layers and output layer are visualized in Fig. 4 and Fig. 5. Fig.4 shows the visualization of texts based on the 50- and 20-dimensional features extracted in ’conv2’ and ’ReLu’ layers, respectively. (Visualization of multidimensional data is done using a technique called t-SNE. See Van der Maaten & Hinton (2008).) It demonstrates that the three major classes are well-separated e.g. ’LEER, ”BECKLEY’ and ’Mountain Laurel’. Fig. 5 depicts the distribution of all texts in ’feat’ layer, where each region is expanded for better visualization. Accordingly, some boxes contain texts belonging to only one class, e.g. 2, 3, 5, 8, 9, 10, 11. The ’2014’ class is mixed with other classes of ’2018’ and ’2016’, as shown in boxes 1, 4, 6, 7, 13. The ’blank’ shreds in box 12 which are combined with ’2016’ texts are mis-labeled texts – reducing the clustering performance of the DSSN model.

Figure 5: Similarity manifold visualization of machine-printed numeric text using t-SNE projection.

In order to evaluate the similarity manifold, several random pairs of images are selected from the test set and feed-forwarded through the DSSN. Then, the Euclidean distance between the paired images is computed based on the output of ’feat’ layer. We choose a decision threshold, , such that is minimized over the training set. Here is the false positive rate (similar images predicted as dissimilar) and is the false negative rate (dissimilar images predicted as similar). We weigh more than because the former increases efficiency at the cost of accuracy while the latter does not hurt accuracy. Table 1 shows the results for the model initialized by MNIST data, and after fine-tuning on the training dataset.

DSSN FN FN Error
Pretrained by MNIST
After Fine-tuning
Table 1: Similarity prediction in Similarity manifold based on Euclidean Distance
Dataset Type Adjusted Rand Index
feat layer ip layer ReLu layer
non-numeric text
numeric text
Table 2: Text Clustering evaluation in Similarity manifold of different layers of DSSN in machine-printed texts

To further evaluate the similarity manifold, a clustering algorithm is applied on texts and the clustered texts are evaluated based on truth labels. For this test, we don’t need parallel networks of DSSN. We use the extracted features from hidden and output layers for clustering of the text. Several clustering algorithms were implemented: K-means, spectral clustering, DBSCAN and agglomerative clustering. To have a better evaluation of features in each layer, we applied clustering algorithms on the features of the ’ReLu’, ’ip’, and ’feat’ layers. The number of clusters for K-means and spectral clustering were set to 8. For DBSCAN and Agglomerative algorithms, the number of clusters was based on the similarity distance between text samples. The clustering performance is measured using Adjusted Rand Index (hub1985comparing, Rand (1971)). Table 2 shows the best clustering algorithm performance, which was agglomerative clustering on 3 layers of DSSN network.

3.2 Text Recognition Evaluation

In section 3.1, the similarity manifold learned by DSSN was evaluated for clustering and similarity prediction. This section focuses on performance of the proposed DSSN-KNN framework, as shown in Fig. 3 for text recognition. The trained DSSN model was tested on three difficult hand-written datasets. These datasets included hand-written and machine printed text with many variations of translation, scale and image patterns for each class. The number of texts and unique classes in each dataset are listed in Table 3.

Dataset#1 (Short text - Unit) Total data Train Data Test Data
No. of Images
No. of Labels
No. of Unseen Labels
No. of blank Images
Dataset#2 (Short text - Non-Numeric) Total data Train Data Test Data
No. of Images
No. of Labels
No. of Unseen Labels
No. of blank Images
Dataset#3 (Short text - Numeric and Non-Numeric) Total data Train Data Test Data
No. of Images
No. of Labels
No. of Unseen Labels
No. of blank Images
Table 3: Hand-written text image datasets

The text recognition performance of DSSN-KNN on the three datasets is listed in Table 5, where the reduction in human estimation is computed. The performance of DSSN-KNN is measured by Accuracy (AC), Accuracy of DSSN-KNN High-Confidence predicted labels (HCAC), Accuracy of medium-confident predicted labels validated by a human (HVAC), False Negative labels (FN), and High-Confidence False Negatives (HCFN). In order to select the confidence and high-confidence thresholds ( and ) for each dataset, we did a grid search over the two thresholds to minimize High Confidence False Negative (HCFN). The chosen thresholds for each dataset and the error values are shown in Table 4.

Dataset Method Model efficiency AC HCAC HVAC FN HCFN
Dataset#1 ROBOTIC SN
DSSN
ASSISTIVE SN -
DSSN -
Dataset#2 ROBOTIC SN
DSSN
ASSISTIVE SN -
DSSN -
Dataset#3 ROBOTIC SN
DSSN
ASSISTIVE SN -
DSSN -
Table 4: Text recognition performance on each dataset with respect to and to achieve HCFN
Dataset Type No. of labeled Images Human-less efficiency
ROBOTIC ASSISTIVE
Dataset #1 machine & hand
Dataset #2 machine & hand
Dataset #3 machine & hand
Table 5: Human-less estimation using proposed DSSN-KNN text recognition model.

Some of the text images where DSSN-KNN produces high confidence errors are shown in Fig. 6. It is evident that most of the example pairs are, in fact, mutually visually similar, and the ”errors” can be attributed to human errors in ground truth. Interestingly, DSSN-KNN sometimes predicts better-than-human labels, for example, spelling corrections.

Figure 6: Texts with HCFN error (where DSSN-KNN produce high confidence wrong prediction). The nearest neighbor text in Similarity Manifold chosen by KNN is shown.

4 Conclusion

In this paper, we proposed a new text recognition model based on visual similarity of text images. A Deeply Supervised Siamese Network is trained along with a K-nearest neighbor classifier, to predict labels of text images. The performance of the proposed model is evaluated for accuracy and reduction of human cost of labeling. The results show that the average value of human-less efficiency on successful field is: in ASSISTIVE mode with NO error, and in ROBOTIC mode with error. Observed errors are explainable. Predicted labels are sometimes better than human labels e.g. spell corrections. Some of the false negative errors we count are in whitespace and irrelevant punctuation (the ”real” error is lower than reported here).

References

  • Bastien et al. (2012) Bastien, Frédéric, Lamblin, Pascal, Pascanu, Razvan, Bergstra, James, Goodfellow, Ian J., Bergeron, Arnaud, Bouchard, Nicolas, and Bengio, Yoshua. Theano: new features and speed improvements. Deep Learning and Unsupervised Feature Learning NIPS 2012 Workshop, 2012.
  • Bromley et al. (1993) Bromley, Jane, Bentz, James W, Bottou, Léon, Guyon, Isabelle, LeCun, Yann, Moore, Cliff, Säckinger, Eduard, and Shah, Roopak. Signature verification using a “siamese” time delay neural network. International Journal of Pattern Recognition and Artificial Intelligence, 7(04):669–688, 1993.
  • Bunke et al. (2004) Bunke, Horst, Bengio, Samy, and Vinciarelli, Alessandro. Offline recognition of unconstrained handwritten texts using hmms and statistical language models. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 26(6):709–720, 2004.
  • Chen & Salman (2011) Chen, Ke and Salman, Ahmad. Extracting speaker-specific information with a regularized siamese deep network. In Advances in Neural Information Processing Systems, pp. 298–306, 2011.
  • Chopra et al. (2005) Chopra, Sumit, Hadsell, Raia, and LeCun, Yann. Learning a similarity metric discriminatively, with application to face verification. In Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE Computer Society Conference on, volume 1, pp. 539–546. IEEE, 2005.
  • Grangier & Bengio (2007) Grangier, David and Bengio, Samy. Learning the inter-frame distance for discriminative template-based keyword detection. In INTERSPEECH, pp. 902–905, 2007.
  • Hadsell et al. (2006) Hadsell, Raia, Chopra, Sumit, and LeCun, Yann. Dimensionality reduction by learning an invariant mapping. In Computer vision and pattern recognition, 2006 IEEE computer society conference on, volume 2, pp. 1735–1742. IEEE, 2006.
  • Jaderberg et al. (2014) Jaderberg, Max, Simonyan, Karen, Vedaldi, Andrea, and Zisserman, Andrew. Synthetic data and artificial neural networks for natural scene text recognition. arXiv preprint arXiv:1406.2227, 2014.
  • Jia et al. (2014) Jia, Yangqing, Shelhamer, Evan, Donahue, Jeff, Karayev, Sergey, Long, Jonathan, Girshick, Ross, Guadarrama, Sergio, and Darrell, Trevor. Caffe: Convolutional architecture for fast feature embedding. arXiv preprint arXiv:1408.5093, 2014.
  • Keeler et al. (1991) Keeler, James D, Rumelhart, David E, Leow, W. K., Lippmann, R. P., Moody, J. M., and Touretzky, D. S. Self-organizing integrated segmentation and recognition neural network. In Neural Information Processing Systems, volume 3, pp. 557–563, 1991.
  • Lavrenko et al. (2004) Lavrenko, Victor, Rath, Toni M, and Manmatha, R. Holistic word recognition for handwritten historical documents. In Document Image Analysis for Libraries, 2004. Proceedings. First International Workshop on, pp. 278–287. IEEE, 2004.
  • LeCun et al. (1998) LeCun, Yann, Bottou, Léon, Bengio, Yoshua, and Haffner, Patrick. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.
  • Lee et al. (2014) Lee, Chen-Yu, Xie, Saining, Gallagher, Patrick, Zhang, Zhengyou, and Tu, Zhuowen. Deeply-supervised nets. arXiv preprint arXiv:1409.5185, 2014.
  • Rand (1971) Rand, William M. Objective criteria for the evaluation of clustering methods. Journal of the American Statistical association, 66(336):846–850, 1971.
  • Van der Maaten & Hinton (2008) Van der Maaten, Laurens and Hinton, Geoffrey. Visualizing data using t-SNE. Journal of Machine Learning Research, 9(2579-2605):85, 2008.
  • Weston et al. (2012) Weston, Jason, Ratle, Frédéric, Mobahi, Hossein, and Collobert, Ronan. Deep learning via semi-supervised embedding. In Neural Networks: Tricks of the Trade, pp. 639–655. Springer, 2012.
  • Zeiler (2012) Zeiler, Matthew D. Adadelta: An adaptive learning rate method. arXiv preprint arXiv:1212.5701, 2012.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
""
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
   
Add comment
Cancel
Loading ...
198769
This is a comment super asjknd jkasnjk adsnkj
Upvote
Downvote
""
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters
Submit
Cancel

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test
Test description