Structural Data Recognition with Graph Model Boosting

Structural Data Recognition with Graph Model Boosting

Tomo Miyazaki and Shinichiro Omachi T. Miyazaki and S. Omachi are with the Graduate School of Engineering, Tohoku University, Japan.

This paper presents a novel method for structural data recognition using a large number of graph models. In general, prevalent methods for structural data recognition have two shortcomings: 1) Only a single model is used to capture structural variation. 2) Naive recognition methods are used, such as the nearest neighbor method. In this paper, we propose strengthening the recognition performance of these models as well as their ability to capture structural variation. The proposed method constructs a large number of graph models and trains decision trees using the models. This paper makes two main contributions. The first is a novel graph model that can quickly perform calculations, which allows us to construct several models in a feasible amount of time. The second contribution is a novel approach to structural data recognition: graph model boosting. Comprehensive structural variations can be captured with a large number of graph models constructed in a boosting framework, and a sophisticated classifier can be formed by aggregating the decision trees. Consequently, we can carry out structural data recognition with powerful recognition capability in the face of comprehensive structural variation. The experiments shows that the proposed method achieves impressive results and outperforms existing methods on datasets of IAM graph database repository.

1 Introduction

Structural data represented by graphs are a general and powerful representation of objects and concepts. A molecule of water can be represented as a graph with three vertices and an edge, where the vertices represent hydrogen and oxygen, and the relation is intuitively described by the edge. Structural data recognition is used in a wide range of applications; for example, in bioinfomatics, chemical compounds need to be recognized as active or inactive.

The difficulty of graph recognition arises out of a lack of mathematical analysis. Even measuring the distance between graphs requires various techniques. The problem of graph recognition has recently been actively studied [1, 2, 3, 4, 5]. Past research has led to notable progress in two aspects. First, graph models have been developed to capture structural variations. Second, the embedding of graphs into Euclidean space has been used to apply sophisticated recognition rules to the vector domain. However, both these aspects have drawbacks. The drawback of the former is that only naive recognition rules are applicable, such as the nearest neighbor (NN) or the k-nearest neighbor (k-NN) methods, which require a considerable amount of training data to achieve high performance. It is unusual for a sufficient amount of data to be available. The drawback of the second aspects above is the loss of structural variation in the embedding process.

Our aim in this paper is to overcome the drawbacks of the previous methods. Specifically, we develop a novel graph model and a boosting framework for structural data to strengthen recognition capability and capture structural variation. These are the two main contributions of this paper. We introduce each briefly below, and then describe them in greater detail in Sections 3 and 4.

The first contribution of this paper is a method to construct a novel graph model. Our model captures four types of structural variations, in vertex composition, edge composition, vertex attributes, and edge attributes. We need vertex correspondence among graphs for model construction, but it is challenging to obtain such correspondences. Most existing methods consider correspondence between graphs [6, 7, 8, 9]. We propose searching vertex correspondences among multiple graphs by assigning labels to vertices with median graphs [10]. The calculation of the median graph is relatively easy. This property helps us obtain the correspondences in a feasible amount of time. This is an important property, since we construct a large number of graph models in the proposed method. Moreover, we develop similarity between a graph and our graph model.

The second contribution of this paper is a novel approach to capture structural variation. In order to capture structural variation comprehensively, we construct a large number of models so that they can contain different structural variations and compensate one another. We then use a boosting method and the models to construct a classifier. Consequently, we can equip the classifier with comprehensive structural variation and a powerful recognition capability. We call this novel approach graph model boosting (GMB).

In experiments, we demonstrated structural data recognition using GMB on five graph datasets that were publicly available. We confirmed that accuracy of GMB notably increased as the boosting process continued. GMB also outperformed existing methods in the area by a large margin. The experimental results thus show the effectiveness of the GMB.

A preliminary version of the work reported here was first presented in a conference paper [11]. We consolidate and expand our previous description and results. Firstly, we provide additional technical details concerning the graph model and GMB. Our contributions are highlighted clearly in the Abstract and Introduction. Secondly, we carried out a wider survey about related work to clarify the significance of the proposed method. Lastly, additional experimental results are presented: time complexity to evaluate practicality, the impact of the parameters for robustness assessment to various datasets, and using other graph model to show the capability of GMB.

2 Related Work

Existing methods for graph recognition can be broadly categorized into three approaches: the one-vs-one approach, the model-based approach, and the embedding approach.

Methods in the one-vs-one approach attempt to classify graphs according to a criterion that can be measured for two graphs, such as graph isomorphism, subgraph isomorphism, and graph edit distance. Two graphs are graph isomorphic if the mapping of their vertices and edges is bijective. Subgraph isomorphism is the case where subgraphs of two graphs satisfy graph isomorphism. Isomorphism is determined by tree search [12] or backtracking [13]. However, it is difficult to determine subgraph isomorphism in case of noisy vertices and edges. Therefore, methods using graph edit distance have been developed. The graph edit distance is defined as the minimum sum of the costs of edit operations that transform the source graph into the target graph by substitution, deletion, and insertion [14, 15]. Since it is expensive to search every combination of edit operations, approximate edit operations are searched. The method proposed in [16] applies the A-star algorithm [17] and the one in [4] exploits Munkres’s algorithm [18]. The literature [19, 20] will help readers find more details related to graph matching and edit distance. The advantage of this approach is that the calculation is completed in polynomial time. However, the methods in this approach only adopt naive recognition rules, such as the NN method and the k-NN method. Furthermore, they only consider local structural variations because the edit distance is measured between two graphs; other graphs are ignored.

Methods in the model-based approach attempt to capture structural variation and classify graphs using a model. The median graph [10] is a model that captures global information concerning graphs. The median graph minimizes the total distance between training graphs. A random graph [21] is a model that specializes in capturing attribute variations at each vertex and edge. The random graph contains variables ranging from 0 to 1, which are associated with attribute values. The variables represent the probabilities of the attribute that the vertices take. However, numerous variables are required when attributes are continuous values. Improved models [22, 23, 24] based on random graph have been developed as well. There are three such models: First-order Gaussian graph, or FOGG [22], function-described graph, or FDG [23], and second-order random graphs, or SORGs [24]. The FOGG is a model designed to avoid increasing number of variables by replacing those of a random graph with parameters of a Gaussian distribution. FDG and SORG incorporate joint probabilities among the vertices and the edges to describe structural information. The difference between FDG and SORG is the numbers of vertices and edges at the calculation of joint probability. FDG uses pairs of vertices or edges, whereas multiple vertices and edges are used in SORG. Recently, models exploiting unsupervised learning methods have been developed [25]. Torsello and Hancock presented a framework to integrate tree graphs into one model by minimizing the minimum description length [26]. Furthermore, Torsello [27] expanded tree graphs to graphs and adopted a sampling strategy [28] to improve calculation efficiency. The EM algorithm has been applied to construct a model [29]. The methods in [25, 27, 29] concentrate on capturing variations in vertex and edge composition. For computational efficiency, a closure tree [30] has been developed. Each vertex of the tree contains information concerning its descendants, so that effective pruning can be carried out. The model-based approach can measure distance based on structural variation. However, these methods also adopt naive recognition rules, the same as in the one-vs-one approach.

Methods in the embedding approach attempt to apply sophisticated recognition rules that are widely used in the vector domain. The main obstacle to embedding graphs is the lack of a straightforward and unique transformation from a graph to a vector. For example, a graph with vertices can be transformed into adjacency matrices, since there are permutations of vertex labeling. Jain and Wysotzki embedded graphs into Euclidean space using the Schur–Hadamard inner product and performed k-means clustering by utilizing neural networks [31]. Bunke and Riesen [5] used the graph edit algorithm to embed graphs into a vector space and applied a support vector machine [32]. Some methods [33, 34, 3] involve applying AdaBoost to graphs. We categorize them as part of this approach because they use sophisticated recognition methods. Kudo et al. developed a decision stump that responds to whether a query graph includes a specific subgraph and constructs a classifier using stumps in AdaBoost [33]. Nowozin et al. developed a classifier based on LPBoost [34]. Zhang et al. [3] improved classifiers by incorporating an error-correcting coding matrix method [35] into boosting. Both [34] and [3] adopted the decision stump as a weak classifier. The difference between GMB and these methods [33, 34, 3] is the structural variation in the classifiers. Comprehensive structural variation is incorporated into the GMB, whereas local structural variation is used in [33, 34, 3] because of the subgraphs. The advantage of the embedding approach is that it can exploit powerful recognition methods. However, a disadvantage is that it ignores structural variation, since the graphs cannot be recovered from the vectors. In addition, embedding graphs with structural variation is a challenging task.

Summarizing the one-vs-one approach and the model-based approach requires a powerful recognition method. The methods in the embedding approach need to incorporate structural variations. The advantages of the model-based and the embedding approaches can complement each other to mitigate their disadvantages. Therefore, our strategy for overcoming the disadvantages is to integrate the model-based approach with a sophisticated recognition method. In this paper, we build several graph models and incorporate them into AdaBoost to construct a graph recognizer that can consider comprehensive structural variation.

3 Graph Model

3.1 Definition of Graph Model

We propose a graph model that captures four types of structural variation in graphs: vertex composition, edge composition, vertex attributes, and edge attributes. The graph model is defined as


where , , , and are the sets of vertices, edges, the probabilities of the vertices and edges, and the parameters of a probability density function of the vertices and edges for attributes, respectively. The compositions and attribute variations are captured by and , respectively. We use probability density function to calculate the probability that vertex takes attribute as . We give an example of how the model describes variation in vertex composition. Let and be elements of and , respectively. The vertex compositions and probabilities are as follows: at , at , at , at , at , at , and at . We can include attribute variations by multiplying .

We define likelihood function whereby a model generates an attributed graph . Let be a matching matrix between and , where is subject to and for any vertex in and in . If two vertices and are matched, , otherwise. Therefore, provides the correspondence between and . We use function to refer to the corresponding vertex . We calculate the likelihood of as


We call this graph model probabilistic attribute graph gneration model, PAGGM.

3.2 Model Construction

The construction of involves the use of training data and correspondences between and . Specifically, we calculate and from vertices and attributes that correspond using . We calculate and of the edges likewise.

Searching is a difficult problem. Attempts have been made to estimate by minimizing objective functions, such as minimum description length [25, 36] and entropy [37]. However, they are not applicable to our study because we search for over thousands of subsets of .

We propose a method to quickly calculate the correspondence. The key idea is to convert searching for into labeling vertices, where vertices in corresponding to the same vertex in share the same label. We exploit median graphs for labeling. The procedure for the assignment of labels is illustrated in Algorithm 1. is used for calculating correspondences between and a graph. Our proposed method can provide correspondences in feasible time because the calculation of the median graph takes time, where is the number of graphs in . We calculate median graph as


where is a function used to calculate the distance between graphs. We adopt an edit distance algorithm [16] as in this paper.



We create as many vertices and edges of PAGGM as the number of labels assigned by Algorithm 1. It is straightforward to organize the labels of edges in when the vertex labels of are available. Edges in have the same label if the connected vertices of the edges have the same label. Let be a function that refers to the labels of vertices; two edges and belong to the same labels if is equal to .

We calculate and in using the labels as


where is a function of proposition and takes if is true, and otherwise takes . We can accumulate the attributes of vertex and edge in as


Then, we calculate using and . However, we need to determine the type of the probability density function before calculating . There is a miscellany of types of attributes, such as continuous or discrete values, finite or infinite set, etc. A suitable function is determined by a type of attribute. In this paper, we adopt a normal density function and a discrete density function for continuous and discrete values, respectively.

3.3 Similarity

We describe similarity between graphs and the PAGGM. An intuitive choice for similarity is the logarithm of Eq. (2). This choice is favorable when every vertex and edge in matches that in . However, unmatched vertices and edges in occur often. In such cases, the logarithm of Eq. (2) is unsuitable because the unmatched vertices and edges in are not accounted for in Eq. (2). Therefore, we impose penalty on unmatched vertices and edges. We define similarity function of graph to a PAGGM as


where and represent sets of unmatched vertices and edges of , respectively. We experimentally set to in this paper.

It is difficult to calculate , which maximizes Eq. (2), when the number of vertices of the PAGGM is large. Unfortunately, the number of vertices is usually large. Therefore, we propose calculating with a set of median graphs calculated in Algorithm 1. The number of vertices of the median graph is small and, most importantly, the vertices and edges of the median graphs correspond to the PAGGM. Hence, the correspondence between and can be obtained through the median graphs. Specifically, we choose one of the median graphs and calculate a matching matrix between the median graph and . Then, we assign vertices in to according to the matching matrix. We repeat this assignment by choosing another median graph until all vertices in corresponding to the PAGGM or all median graphs have been chosen. Finally, we obtain matching matrix . The edit distance algorithm [16] is used to calculate the matching matrices between and the median graphs. We illustrate an example of the calculation of in Fig. 1.

\scalebox1[1] \includegraphicsMatching-Strategy-a \scalebox1[1]\includegraphicsMatching-Strategy-b-2
(a) graph (b) PAGGM
(c) Median graphs
(d) Process of assigning labels to the graph.
Fig. 1: Calculation of matching matrix between graph (a) and PAGGM (b). The colors represent labels. The correspondence develops from left to right in (c).

4 Graph Model Boosting

The aim of GMB is to capture comprehensive variation and construct a powerful classifier with a large number of graph models. Although the PAGGM can contain structural variations, the PAGGM is single and tends to capture major variations in the training graphs. Hence, minor but important variations are lost. Consequently, recognition errors occur on graphs in minor structures. To overcome this problem, we adopt a boosting algorithm to generate a large number of PAGGMs to complement one another. We provide an overview of the GMB in Fig. 2. In each boosting round, a PAGGM is constructed with a subset of weighted training data for each class. Note that the PAGGM can focus on minor graphs with large weights. Then, we form a decision tree that classifies a graph by recursively branching left or right down the tree according to the similarity between the graph and a PAGGM. Finally, we form a strong classifier by integrating the decision trees.



Fig. 2: Overview of graph model boosting.

4.1 Model extension for weighted training data

We extend the PAGGM for weighted training data. We begin by extending the median graph to a weighted median graph. Let be a set of weights for graphs in . We define the weighted median graph as


We replace the median graph in Algorithm 1 with the weighted median graph. Therefore, of is composed of . Subsequently, we extend the calculation of and to incorporate as


where is a function that refers to the position of in ascending order of . For example, given , . Let represent a duplication operator, such as . We calculate with . We replace and with and , respectively, when is given.

4.2 Boosting framework

We propose GMB as shown in Algorithm 2, where represents a set of class labels of . GMB is based on the AdaBoost algorithms [38, 39, 40], which are a suitable choice for model construction because they provide different subsets of training data. We construct PAGGMs using the subsets. In addition to the different subsets, the weight can diversify the PAGGMs. In the AdaBoost framework, the weights of error-prone graphs become larger than those of recognized graphs. Hence, we can focus on such error-prone graphs by constructing PAGGMs using weights.



We use decision trees [41] as weak classifiers . The trees are trained by searching branching rules composed of the PAGGM of a class and a threshold of similarity. The trees have different PAGGMs and branching rules because of randomness in the GMB. Subsequently, we use all trained trees to construct the strong classifier as


In recognition, evaluates test graphs using the trees and recognizes them by majority voting on the trees. Hence, the test graphs are comprehensively analyzed by measuring similarities to the PAGGMs and using more sophisticated recognition rules than NN and k-NN.

5 Experiments

We carried out the experiments below to verify the effectiveness of the proposed method. We measured recognition accuracy for PAGGM and GMB on five datasets.

5.1 Datasets

For the experiments, we used five datasets from the IAM graph database repository [42], which is widely used for graph-based pattern recognition and machine learning. We used the following datasets: Letter-LOW (LOW), Letter-MED (MED), Letter-HIGH (HIGH), GREC, and COIL-RAG (COIL).

The LOW, MED, and HIGH contained graphs representing distorted letters, which were the 15 capital letters of the Roman alphabet composed of straight lines only, such as A, E, F, H, I, K, L, M, N, T, U, V, W, X, Y, and Z. The terms “LOW,” “MED,” and “HIGH” represent the levels of distortion.

The GREC consisted of graphs representing symbols from architectural and electronic drawings. The symbols were converted into graphs by assigning vertices to corners, intersections, and circles on the symbols. The lines between symbols were edges. There were 22 classes in the GREC. As suggested in [43], we used graphs consisting of only straight lines.

The COIL was composed of 100 classes of graphs extracted from images of different objects of the COIL-100 database [44]. The images were segmented into regions according to color by the mean shift algorithm [45]. The vertices were assigned to the segmented regions, and adjacent regions were connected by edges. The attributes were color histograms.

We summarize the characteristics of the datasets in Table I. We calculated in the following probability density functions: two-dimensional normal distribution for attributes of (x,y), discrete distribution function for line type and RGB histogram, and one-dimensional normal density for the line boundary.

#train #test #class vertex attribute edge attribute ave(v) ave(e) max(v) max(e) #a class
LOW 750 750 15 (x,y) none 4.7 3.1 8 6 50
MED 750 750 15 (x,y) none 4.7 3.2 9 7 50
HIGH 750 750 15 (x,y) none 4.7 4.5 9 9 50
GREC 213 390 22 (x,y) Line type 11.6 11.9 24 26 15
COIL 2 400 1 000 100 RGB histogram Line boundary 3.0 3.0 11 13 24
TABLE I: Summary of characteristics of datasets used in the experiments. The Top row indicates dataset name, the number of training data items, test data, classes, attribute types of vertex and edge, average numbers of vertices and edges, maximum number of vertices and edges, and the number of graphs in a class.

5.2 Experiments with PAGGM

We carried out experiments with PAGGM. In order to evaluate PAGGM, we constructed one PAGGM for each class by using all training data. We then classified the test data by applying the NN method with similarity obtained by Eq. (12). For comparison, we calculated the median graphs [10] and classified test data by the edit distance algorithm [16]. The experimental results are shown in Table II.

Median 84.3 77.3 78.0 84.4 58.3
PAGGM 95.2 87.5 85.9 90.0 71.1
TABLE II: Recognition rates (%) of median graph and PAGGM using the NN method.

The results show that the PAGGM outperformed the median graph for every dataset. Both the PAGGM and the median graph were categorized into the model-based approach. The main difference was whether it contained structural variation. These results signify that recognition performance improved when we incorporated structural variation. The experimental results showed that the PAGGM successfully captured structural variation.

5.3 Experiments with Graph Model Boosting

We proved the effectiveness of GMB by showing the evolution of its recognition rate. The recognition rate fluctuated during the trials due to random sampling. We repeated the experiments 10 times and calculated the average recognition rate. In this experiment, we set the cardinality of subset to 5 in LOW, 35 in MED, 40 in HIGH, 6 in GREC, and 24 in COIL, where the number of boosting rounds was . Fig. 3 illustrates the evolutions. The evolutions showed that the average recognition rates increased steadily on every dataset. The initial average recognition rates were lower than the results with the single PAGGM, shown in Table II. At the beginning, GMB only captured the structural variation in subsets of the training data, resulting in poor results. However, the accuracy of GMB increased as the process continued and, finally, exceeded the results of the single PAGGM. This phenomenon signifies that comprehensive structural variations were successfully incorporated into the classifier of the GMB.



Fig. 3: Evolutions of average recognition rates of GMB in boosting rounds. The evolutions mark the highest values as: at 94 round on LOW, at 85 round on MED, at 99 round on HIGH, at 84 round on GREC, and at 83 round on COIL. Best viewed in color.

We carefully determined since random sampling is an important process. We repeated this experiment 10 times for each and calculated the maximum values of the averages recognition rates. Fig. 4 shows the results. The results show the robustness of GMB. In general, the effects were within a few points of one another on the datasets. GMB achieved good results on the datasets using various . Note that the results in LOW and GREC signified that performance can be maximized when the number of samples is small. The small number of samples facilitated the generation of different PAGGMs. Consequently, more comprehensive structural variations were incorporated into the GMB.



Fig. 4: The effects of the cardinarity of a subset . The y- and x-axes represent average recognition rate (%) and , respectively. The highest recognition rates were as follows: (a) at , (b) at , (c) at , (d) at , (e) at .

Graph matching is an NP-complete problem, and encounters a computation bottleneck in the GMB in Steps 2 and 4 in Algorithm 1. Although we adopted a method to calculate approximate matching [16], it took a considerable amount of time to perform over 100 rounds of boosting. Therefore, to handle this computation problem, we performed graph matching on all pairs prior to GMB. In the GMB, we simply referred to the results. The average processing times over 100 boosting rounds were 6.3 (s/round) in LOW, 12.9 (s/round) in MED, 8.6 (s/round) in HIGH, 13.5 (s/round) in GREC, and 631.4 (s/round) in COIL. Note that the number of classes and the cardinality of the subset in COIL were greater than in the other datasets. Consequently, the processing times were longer.

We compared GMB with existing graph recognition methods, categorized into the one-vs-one, the model-based, and the embedding approaches. We selected the following methods: Riesen’s method [4] from the one-vs-one approach, and Bunke’s method [5] from the embedding approach, He’s [30], Wong’s [21], Serratosa’s [23], and Sanfeliu’s [24] methods from the model-based approach. The comparisons on LOW, HIGH, and GREC are summarized in Table III. We referred to the results of [4, 30, 21, 23, 24] to [43] on the IAM database, whereas the original papers were not evaluated. Although the methods in the one-vs-one and the embedding approaches recorded high scores, the methods in the model-based approach obtained even higher scores. These results signify the importance of structural information. In the graph domain, sophisticated recognition rules need structural information to perform to their full capabilities. The PAGGM obtained high scores on every dataset. Furthermore, GMB outperformed the other methods. We successfully incorporated a large amount of structural information into the powerful recognition rule. These comparison results verified the effectiveness of GMB.

Method Approach LOW HIGH GREC
Riesen [4] One. 91.1 77.6 -
Bunke [5] Emb. 92.9 - -
Jiang [10] Model 96.4 74.5 80.5
He [30] Model 93.6 49.1 57.0
Wong [21] Model 93.9 80.1 85.8
Serratosa [23] Model 93.9 80.1 85.8
SanFeLiu [24] Model 94.0 80.9 91.2
Proposed (PAGGM) Model 95.2 85.9 90.0
Proposed (GMB) Boosting 98.7 88.9 96.7
TABLE III: Comparison of recognition rate with those of other methods.

Finally, to demonstrate the capability of GMB, we carried out further experiments whether GMB works well with other graph models. To this end, we used the median graph as graph models constructed in GMB. We show the results in Table IV. The recognition rates increased in all datasets. These results signify the capability of GMB.

Initial 91.3 77.8 67.8 84.8 76.3
Max 95.0 88.8 86.1 94.1 91.4
TABLE IV: Recognition rates (%) of GMB using median graph as model.

6 Conclusions

In this paper, we introduced PAGGM and GMB, a novel graph model and algorithm for graph recognition. PAGGM captures structural variations in composition and attribute of vertex and edge. We proposed an efficient algorithm for searching vertex correspondences among multiple graphs. The algorithm enables us to construct PAGGM in a feasible amount of time. GMB constructs a large number of PAGGMs to comprehensively capture the structural variation of graphs. Then, we formed a strong classifier using the constructed PAGGMs in the boosting framework. Consequently, the classifier was equipped with the requisite information concerning structural variation and a powerful recognition ability.

The experimental results showed that PAGGM successfully captured structural variation and GMB significantly enhances recognition performance. Furthermore, GMB outperformed the methods proposed in past work. Therefore, we successfully strengthened the recognition capability and the ability to deal with structural variation in graphs.

Structural data are a powerful representation of objects even in images. The proposed method can be applied to computer vision tasks where relation of visual features needs to be considered. Developing object recognition applications using the proposed method is planned for future work.


This research was supported by JSPS KAKENHI Grant Number 15H06009.


  • [1] D. White and R. C. Wilson, “Parts based generative models for graphs,” in International Conference on Pattern Recognition, 2008, pp. 1–4.
  • [2] L. Han, L. Rossi, A. Torsello, R. C. Wilson, and E. Hancock, “Information theoretic prototype selection for unattributed graphs,” in Structural, Syntactic, and Statistical Pattern Recognition, ser. Lecture Notes in Computer Science, 2012, vol. 7626, pp. 33–41.
  • [3] B. Zhang, G. Ye, Y. Wang, W. Wang, J. Xu, G. Herman, and J. Yang, “Multi-class graph boosting with subgraph sharing for object recognition,” in 20th International Conference on Pattern Recognition, 2010, pp. 1541–1544.
  • [4] K. Riesen and H. Bunke, “Approximate graph edit distance computation by means of bipartite graph matching,” Image and Vision Computing, vol. 27, no. 7, pp. 950–959, 2009.
  • [5] H. Bunke and K. Riesen, “Towards the unification of structural and statistical pattern recognition,” Pattern Recognition Letters, vol. 33, no. 7, pp. 811–825, 2012.
  • [6] M. Cho, J. Lee, and K. Lee, “Reweighted random walks for graph matching,” in Computer Vision - ECCV, ser. Lecture Notes in Computer Science, 2010, vol. 6315, pp. 492–505.
  • [7] M. Cho, K. Alahari, and J. Ponce, “Learning graphs to match,” in IEEE International Conference on Computer Vision, 2013, pp. 25–32.
  • [8] T. Cour, P. Srinivasan, and J. Shi, “Balanced graph matching,” in Advances in Neural Information Processing Systems 19, 2007, pp. 313–320.
  • [9] S. Gold and A. Rangarajan, “A graduated assignment algorithm for graph matching,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 18, no. 4, pp. 377–388, 1996.
  • [10] X. Jiang, A. Müunger, and H. Bunke, “On median graphs: Properties, algorithms, and applications,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 23, no. 10, pp. 1144–1151, 2001.
  • [11] T. Miyazaki and S. Omachi, “Graph model boosting for structural data recognition,” in Proceedings of the 23rd International Conference on Pattern Recognition, 2016, pp. 1708–1713.
  • [12] J. R. Ullmann, “An algorithm for subgraph isomorphism,” J. ACM, vol. 23, no. 1, pp. 31–42, 1976.
  • [13] D. E. Ghahraman, A. K. C. Wong, and T. Au, “Graph monomorphism algorithms,” IEEE Transactions on Systems, Man and Cybernetics, vol. 10, no. 4, pp. 189–196, 1980.
  • [14] M. A. Eshera and K. S. Fu, “A graph distance measure for image analysis,” IEEE Transactions on Systems, Man and Cybernetics, vol. SMC-14, no. 3, pp. 398–408, 1984.
  • [15] A. Sanfeliu and K. S. Fu, “A distance measure between attributed relational graphs for pattern recognition,” IEEE Transactions on Systems, Man and Cybernetics, vol. SMC-13, no. 3, pp. 353–362, 1983.
  • [16] M. Neuhaus, K. Riesen, and H. Bunke, “Fast suboptimal algorithms for the computation of graph edit distance,” in Proceedings of the International Conference on Structural, Syntactic, and Statistical Pattern Recognition, 2006, pp. 163–172.
  • [17] P. Hart, N. Nilsson, and B. Raphael, “A formal basis for the heuristic determination of minimum cost paths,” IEEE Transactions on Systems Science and Cybernetics, vol. 4, no. 2, pp. 100–107, 1968.
  • [18] J. Munkres, “Algorithms for the assignment and transportation problems,” Journal of the Society for Industrial and Applied Mathematics, vol. 5, no. 1, pp. 32–38, 1957.
  • [19] D. Conte, P. Foggia, C. Sansone, and M. Vento, “Thirty years of graph matching in pattern recognition,” International Journal of Pattern Recognition and Artificial Intelligence, vol. 18, no. 03, pp. 265–298, 2004.
  • [20] P. Bille, “A survey on tree edit distance and related problems,” Theor. Comput. Sci., vol. 337, no. 1-3, pp. 217–239, 2005.
  • [21] A. K. C. Wong and M. You, “Entropy and distance of random graphs with application to structural pattern recognition,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 7, no. 5, pp. 599–609, 1985.
  • [22] A. D. Bagdanov and M. Worring, “First order gaussian graphs for efficient structure classification,” Pattern Recognition, vol. 36, no. 6, pp. 1311–1324, 2003.
  • [23] F. Serratosa, R. Alquézar, and A. Sanfeliu, “Function-described graphs for modelling objects represented by sets of attributed graphs,” Pattern Recognition, vol. 36, no. 3, pp. 781–798, 2003.
  • [24] A. Sanfeliu, F. Serratosa, and R. Alquézar, “Second-order random graphs for modeling sets of attributed graphs and their application to object learning and recognition,” International Journal of Pattern Recognition and Artificial Intelligence, vol. 18, no. 03, pp. 375–396, 2004.
  • [25] A. Torsello and E. R. Hancock, “Learning shape-classes using a mixture of tree-unions,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 28, no. 6, pp. 954–967, 2006.
  • [26] J. Rissanen, “Modeling by shortest data description,” Automatica, vol. 14, no. 5, pp. 465–471, 1978.
  • [27] A. Torsello, “An importance sampling approach to learning structural representations of shape,” in IEEE Conference on Computer Vision and Pattern Recognition, 2008, pp. 1–7.
  • [28] J. M. Hammersley and D. C. Handscomb, Monte Carlo Methods.   Wiley, 1964.
  • [29] L. Han, R. Wilson, and E. Hancock, “A supergraph-based generative model,” in International Conference on Pattern Recognition, 2010, pp. 1566–1569.
  • [30] H. He and A. K. Singh, “Closure-tree: An index structure for graph queries,” in International Conference on Data Engineering, 2006, pp. 38–38.
  • [31] B. J. Jain and F. Wysotzki, “Central clustering of attributed graphs,” Machine Learning, vol. 56, no. 1-3, pp. 169–207, 2004.
  • [32] V. N. Vapnik, Statistical Learning Theory.   Wiley-Interscience, 1998.
  • [33] T. Kudo, E. Maeda, and Y. Matsumoto, “An application of boosting to graph classification,” in Advances in Neural Information Processing Systems 17, 2005, pp. 729–736.
  • [34] S. Nowozin, K. Tsuda, T. Uno, T. Kudo, and G. Bakir, “Weighted substructure mining for image analysis,” in IEEE Conference on Computer Vision and Pattern Recognition, 2007, pp. 1–8.
  • [35] T. G. Dietterich and G. Bakiri, “Solving multiclass learning problems via error-correcting output codes,” J. Artif. Int. Res., vol. 2, no. 1, pp. 263–286, 1995.
  • [36] L. Han, R. Wilson, and E. Hancock, “Generative graph prototypes from information theory,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 37, no. 10, pp. 2013–2027, 2015.
  • [37] H. Bunke, P. Foggia, C. Guidobaldi, and M. Vento, “Graph clustering using the weighted minimum common supergraph,” in Graph Based Representations in Pattern Recognition, ser. Lecture Notes in Computer Science, 2003, vol. 2726, pp. 235–246.
  • [38] Y. Freund and R. E. Schapire, “A decision-theoretic generalization of on-line learning and an application to boosting,” Journal of Computer and System Sciences, vol. 55, no. 1, pp. 119–139, 1997.
  • [39] T. Hastie, S. Rosset, J. Zhu, and H. Zou, “Multi-class adaboost,” Statistics and Its Interface, vol. 2, no. 3, pp. 349–360, 2009.
  • [40] E. L. Allwein, R. E. Schapire, and Y. Singer, “Reducing multiclass to binary: A unifying approach for margin classifiers,” J. Mach. Learn. Res., vol. 1, pp. 113–141, 2001.
  • [41] L. Breiman et al., Classification and regression trees.   Chapman & Hall, 1984.
  • [42] K. Riesen and H. Bunke, “IAM graph database repository for graph based pattern recognition and machine learning,” in Proceedings of the International Workshop on Structural, Syntactic, and Statistical Pattern Recognition, 2008, pp. 287–297.
  • [43] A. Solé-Ribalta, X. Cortś, and F. Serratosa, “A comparison between structural and embedding methods for graph classification,” in Structural, Syntactic, and Statistical Pattern Recognition, ser. Lecture Notes in Computer Science, 2012, vol. 7626, pp. 234–242.
  • [44] S. A. Nene, S. K. Nayar, and H. Murase, “Columbia object image library: Coil-100,” Department of Computer Science, Columbia University, Tech. Rep. CUCS-006-96, 1996.
  • [45] D. Comaniciu and P. Meer, “Robust analysis of feature spaces: color image segmentation,” in IEEE Conference on Computer Vision and Pattern Recognition, 1997, pp. 750–755.

Tomo Miyazaki received a B.E. degree from the Department of Informatics, Faculty of Engineering, Yamagata University in 2006. He received M.E. and Ph.D. degrees from the Graduate School of Engineering, Tohoku University in 2008 and 2011, respectively. He joined Hitachi, Ltd. in 2011 and has worked at the Graduate School of Engineering, Tohoku University from 2013 to 2014 as a researcher. Since 2015, he has been an Assistant Professor. His research interests include pattern recognition and image processing. Dr. Miyazaki is a member of the Institute of Electronics, Information and Communication Engineers.


Shinichiro Omachi received B.E., M.E., and Doctor of Engineering degrees in Information Engineering from Tohoku University, Japan, in 1988, 1990, and 1993, respectively. He worked as a research associate at the Education Center for Information Processing at Tohoku University from 1993 to 1996. Since 1996, he has been with the Graduate School of Engineering at Tohoku University, where he is currently a professor. From 2000 to 2001, he was a visiting associate professor at Brown University. His research interests include pattern recognition, computer vision, image processing, image coding, and parallel processing. He received the IAPR/ICDAR Best Paper Award in 2007, the Best Paper Method Award of the 33rd Annual Conference of the GfKl in 2010, the ICFHR Best Paper Award in 2010, and the IEICE Best Paper Award in 2012. Dr. Omachi is a member of the Institute of Electronics, Information and Communication Engineers, among others.

Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
Add comment
Loading ...
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test description