A Tutorial on Network Embeddings

A Tutorial on Network Embeddings

Haochen Chen Stony Brook University Bryan Perozzi Google Research Rami Al-Rfou Google Research Steven Skiena Stony Brook University

Network embedding methods aim at learning low-dimensional latent representation of nodes in a network. These representations can be used as features for a wide range of tasks on graphs such as classification, clustering, link prediction, and visualization. In this survey, we give an overview of network embeddings by summarizing and categorizing recent advancements in this research field. We first discuss the desirable properties of network embeddings and briefly introduce the history of network embedding algorithms. Then, we discuss network embedding methods under different scenarios, such as supervised versus unsupervised learning, learning embeddings for homogeneous networks versus for heterogeneous networks, etc. We further demonstrate the applications of network embeddings, and conclude the survey with future work in this area.

1 Introduction

From social networks to the World Wide Web, networks provide a ubiquitous way to organize a diverse set of real-world information. Given a network’s structure, it is often desirable to predict missing information (frequently called attributes or labels) associated with each node in the graph. This missing information can represent a variety of aspects of the data – for example, on a social network they might represent the communities a person belongs to, or the categories of a document’s content on the web.

Because information networks can contain billions of nodes and edges, it can be intractable to perform complex inference procedures on the entire network. One technique which has been proposed to address this problem is network embedding. The central idea is to find a mapping function which converts each node in the network to a low-dimensional latent representation. These representations can then be used as features for common tasks on graphs such as classification, clustering, link prediction, and visualization.

To sum up, we seek to learn network embeddings with the following characteristics:

  • Adaptability - Real networks are constantly evolving; new applications should not require repeating the learning process all over again.

  • Scalability - Real networks are often large in nature, thus network embedding algorithms should be able to process large-scale networks in a short time period.

  • Community aware - The distance between latent representations should represent a metric for evaluating similarity between the corresponding members of the network. This allows generalization in networks with homophily.

  • Low dimensional - When labeled data is scarce, low-dimensional models generalize better, and speed up convergence and inference.

  • Continuous - We require latent representations to model partial community membership in continuous space. In addition to providing a nuanced view of community membership, a continuous representation has smooth decision boundaries between communities which allows more robust classification.

(a) Input: Karate Graph
(b) Output: Network Embedding
Figure 1: Network embedding methods learn latent representation of nodes in a network in . This learned representation encodes community structure so it can be easily exploited by standard classification methods. Here, DeepWalk [33] is used on Zachary’s Karate network [60] to generate a latent representation in . Note the correspondence between community structure in the input graph and the embedding. Vertex colors represent a modularity-based clustering of the input graph.

As a motivating example we show the result of applying DeepWalk [33], which is a widely used network embedding method, to the well-studied Karate network in Figure 1. This network, as typically presented by force-directed layouts, is shown in Figure 1. Figure 0(b) shows the output of DeepWalk with two latent dimensions. Beyond the striking similarity, note that linearly separable portions of (0(b)) correspond to clusters found through modularity maximization in the input graph (0(a)) (shown as vertex colors).

The rest of the survey is organized as follows.111The area of network embedding is rapidly growing, and while we have made an effort to include all relevant work in this survey, there have doubtlessly been accidental omissions. If you are aware of work that would improve the completeness of this survey, please let the authors know. We first provide a general overview of network embedding and give some definitions and notations which will be used later. In Section 2, we introduce unsupervised network embedding methods on homogeneous networks without attributes. Section 3 reviews embedding methods on attributed networks and partially labeled networks. Then, Section 4 discusses heterogeneous network embedding algorithms. We further demonstrate the applications of network embeddings, and conclude the survey with future work in this area.

1.1 A Brief History of Network Embedding

Traditionally, graph embeddings have been described in the context of dimensionality reduction. Classical techniques for dimensionality reduction include principal component analysis (PCA) [51] and multidimensional scaling (MDS) [25]. Both methods seek to represent an matrix as a matrix where . For graphs, is typically an matrix, where could be the adjacency matrix, normalized Laplacian matrix or all-pairs shortest path matrix, to name a few. Both methods are capable of capturing linear structural information, but fails to discover the non-linearity within the input data.

PCA - PCA computes a set of orthogonal principal components, where each principal components is a linear combinations of the original variables. The number of these components could be equal or less than , which is the reason that PCA can serve as a dimensionality reduction technique. Once the principal components are computed, each original data point could be projected to the lower-dimensional space determined by them. For a square matrix, the time complexity of PCA is .

MDS - Multidimensional scaling (MDS) projects each row of to a -dimensional vector, such that the distance between different objects in the original feature matrix is best preserved in the -dimensional space. Specifically, let to be the coordinate of the -th object in the embedding space, metric MDS minimizes the following stress function:


Exact MDS computation requires eigendecomposition of a transformation of , which takes time.

In early 2000s, other methods such as IsoMap [44] and locally linear embeddings (LLE) [39] were proposed to preserve the global structure of non-linear manifolds. We note that both methods are defined abstractly for any type of dataset, and first preprocesses the data points into graphs which capture local neighborhood performance.

Isomap - Isomap [44] is an extension to MDS with the goal of preserving geodesic distances in the neighborhood graph of input data. The neighborhood graph is constructed by connecting each node with either nodes closer than a certain distance or nodes which are -nearest neighbors of . Then, classical MDS is applied to to map data points to a low-dimensional manifold which preserves geodesic distances in .

Local Linear Embeddings (LLE) - Unlike MDS, which preserves pairwise distances between feature vectors, LLE [39] only exploits the local neighborhood of data points and does not attempt to estimate distance between distant data points. LLE assumes that the input data is intrinsically sampled from a latent manifold, and that a data point can be reconstructed from a linear combination of its neighbors. The reconstruction error can be defined as


where is a weight matrix denoting data point ’s contribution to ’s reconstruction, which is computed by minimizing the loss function above. Since reflects the invariant geometric properties of the input data, it can used to find the mapping from a data point to its low-dimensional representation . To compute , LLE minimizes the following embedding cost function:


Since is already fixed, it can be proved that this cost function can be minimized by finding the eigenvectors of an auxiliary matrix.

In general, these methods all offer good performance on small networks. However, the time complexity of these methods are at least quadratic, which makes them impossible to run on large-scale networks.

Another popular class of dimensionality reduction techniques uses the spectral properties (e.g. eigenvectors) of matrices derivable from the graph (e.g. Graph Laplacian) to embed the nodes of the graph. Laplacian eigenmaps (LE) [3], represent each node in the graph by the eigenvectors associated with its -smallest nontrivial eigenvalues. The spectral properties of the Graph Laplacian encode cut information about the graph, and have a rich history of use in graph analysis [15]. Let be the weight of the connection between node and , the diagonal weight matrix can be constructed:


The Laplacian matrix of is then


The solutions to the eigenvector problem:


can be used as the low-dimension embeddings of the input graph.

Tang and Liu [43] examined using eigenvectors of the Graph Laplacian for classification in social networks. They argue that nodes (actors) in a network are associated with different latent affiliations. On the other hand, these social dimensions should also be continuous since the actors might have different magnitude of associations to one affiliation. Another similar method, SocDim [42] proposed using the spectral properties of the modularity matrix as latent social dimensions in networks. However, the performance of these methods has been shown to be lower than neural network-based approaches [33], which we will shortly discuss.

(a) Output: DeepWalk
(b) Output: PCA
(c) Output: MDS
(d) Output: LLE
(e) Output: LE
(f) Output: SVD
Figure 2: Two-dimensional embeddings of the Karate graph using DeepWalk and several early dimension reduction techniques. The input is the adjacency matrix for DeepWalk and SVD, and the geodesic matrix for the other four methods.

1.2 The Age of Deep Learning

DeepWalk [33] was proposed as the first network embedding method using techniques from the representation learning (or deep learning) community. DeepWalk bridges the gap between network embeddings and word embeddings by treating nodes as words and generating short random walks as sentences. Then, neural language models such as Skip-gram [29] can be applied on these random walks to obtain network embedding. DeepWalk has become arguably the most popular network embedding method since then, for several reasons.

First of all, random walks can be generated on demand. Since the Skip-gram model is also optimized per sample, the combination of random walk and Skip-gram makes DeepWalk an online algorithm. Secondly, DeepWalk is scalable. Both the process of generating random walks and optimizing the Skip-gram model are efficient and trivially parallelizable. Most importantly, DeepWalk introduces a paradigm for deep learning on graphs, as shown in Figure 3.

The first part of the DeepWalk paradigm is choosing a matrix associated with the input graph, for which DeepWalk chooses the random walk transition matrix. Indeed, a variety of other choices are also proved to be feasible, such as the normalized Laplacian matrix and the powers of the adjacency matrix.

The second step is graph sampling, where sequences of nodes are implicitly sampled from the chosen matrix. Note that this step is optional; some network embedding algorithms directly compute the exact matrix elements and build embedding models on it. However, in a lot of cases graph sampling is a favorable intermediate step for the following two reasons. First, depends on the matrix of choice, it could take up to quadratic time to compute its exact elements; one example is computing the power series of the adjacency matrix. In this scenario, graph sampling serves as an scalable approach for approximating the matrix. Second, compared to a large-scale and sparse graph which is difficult to model, sequences of symbols are much more easier for deep learning models to deal with. There are a lot of readily available deep learning methods for sequence modeling, such as RNNs and CNNs. DeepWalk generates sequence samples via truncated random walk, which effectively extends the neighborhood of graph nodes.

The third step is learning node embeddings from the generated sequences (or the matrix in the first step). Here, DeepWalk adopts Skip-gram as the model for learning node embeddings, which is one of the most performant and efficient algorithms for learning word embeddings.

The DeepWalk paradigm is highly flexible, which can be expanded in two possible ways:

  1. The complexity of the graphs being modeled can be expanded. For example, HOPE [31] aims at embedding directed graphs, SiNE [49] and SNE [59] are methods for embedding signed networks. In comparison, the methods of [55, 40, 61, 12, 50] are designed for attributed network embedding. Much recent work [7, 64, 27, 22, 53] also attempts to embed heterogeneous networks. Besides these unsupervised methods, network embedding algorithms [46, 57, 24] have been proposed for semi-supervised learning on graphs. We will have detailed discussion of these methods in Section 3 and Section 4.

  2. The complexity of the methods used for the two critical components of DeepWalk, namely sampling sequences from a latent matrix and learning node embeddings from the sampled sequences, can be expanded. Much work on network embedding are extensions to DeepWalk’s basic framework. For instance, [41, 21, 34, 5] propose new strategies for sequence sampling, while [48, 6, 5] present new stategies for modeling sampled sequences. These approachs will be further analyzed in Section 2.

Input: Choose a matrix associated with the input graph

Graph Sampling: Sample sequences from the chosen matrix

Modeling: Learn embeddings from the sequences or the matrix itself


DeepWalk: Random walk transition matrix

DeepWalk: Truncated random walks

DeepWalk: Skip-gram

DeepWalk: Node Embeddings

Figure 3: A paradigm for Deep Learning On Graphs with DeepWalk’s design choice for each building block.

1.3 Notations and Definitions

Here we introduce the definitions of certain concepts we will used throughout this survey:

  • Definition 1

    (graph) A simple undirected graph is a collection of vertices together with a set of edges, which are unordered pairs of the vertices. In other words, the edges in an undirected graph have no orientation.

    The adjacency matrix of is an matrix where if there is an edge between and , and otherwise. Unless otherwise stated, we use both graph and network to refer to a simple undirected graph.

  • Definition 2

    (network embedding) For a given a network , a network embedding is a mapping function , where . This mapping defines the latent representation (or embedding) of each node . Also, we use to denote the embedding vector for node .

  • Definition 3

    (directed graph) A directed graph is a collection of vertices together with a set of edges, which are ordered pairs of the vertices. The only difference between a directed graph and an undirected graph is that the edges in a directed graph have orientation.

  • Definition 4

    (heterogeneous network) A heterogeneous network is a network with multiple types of nodes or multiple types of edges. Formally, is associated with a node type mapping and an edge type mapping , where is the set of all node types and is the set of all edge types.

  • Definition 5

    (signed graph) A signed graph is a graph where each edge is associated with a weight . An edge with weight of 1 denotes a positive link between nodes, whereas an edge with weight of -1 denotes a negative link. Signed graphs can be used to reflect agreement or trust.

2 Unsupervised Network Embeddings

In this section, we introduce network embedding methods on simple undirected networks. We first propose a categorization of the existing methods, and then introduce several representative methods within each category.

Recent scalable network embedding algorithms are inspired by the emergence of neural language models [4] and word embeddings, in particular [29, 30, 32]. Skip-gram [29] is a highly efficient method for learning word embeddings. Its key idea is to learn embeddings which are good at predicting nearby words in sentences. The nearby words (or context words) for a certain word in a sentence are usually defined as the set of words within a pre-defined window size , namely . Specifically, the Skip-gram model minimizes the following objective:


where is calculated using a hierarchical or sampled softmax function:


Here is the distributed representation of when it serves as a context word, and is the vocabulary size.

To summarize, the Skip-gram model consists of two phases. The first phase identifies the context words for each word in each sentence, while the second phase maximizes the conditional probability of observing the context words given a center word.

By capturing the intrinsic similarity between language modeling and network modeling, DeepWalk [33] proposed a two-phase algorithm for learning network embedding. The analogy made by DeepWalk is that nodes in a network can be thought of as words in an artificial language. Similar to the Skip-gram model for learning word embeddings, the first step of DeepWalk is to identify the context nodes for each node. By generating truncated random walks in the network (which are analogous to sentences), the context nodes of can be defined as the set of nodes within a window size in each random walk sequence, which can be seen as a combination of nodes from ’s -hop, -hop, and up to -hop neighbors. In other words, DeepWalk learns the network embedding from the combination of where is the -th power of the adjacency matrix. Once the context nodes have been determined, the second step is same as that of the original Skip-gram model: learn embeddings which maximizes the likelihood of predicting context nodes. DeepWalk uses the same optimization goal and optimization method as Skip-gram, but any other language model could also be used in principle.

0:  network window size embedding size walks per vertex walk length
0:  matrix of vertex representations
1:  Initialization: Sample from
2:  Build a binary Tree from
3:  for  to  do
5:     for each  do
6:        t
7:        SkipGram(, , )
8:     end for
9:  end for
Algorithm 1 DeepWalk(, , , , )

Lines 3-9 in Algorithm 1 shows the core of DeepWalk. The outer loop specifies the number of times, of starting random walks at each node. We can think of each iteration as making a ‘pass’ over the data, sampling one walk per node during this pass. At the start of each pass, DeepWalk generates a random ordering to traverse the vertices.

In the inner loop, DeepWalk iterates over all the vertices of the graph. For each node a random walk is generated, and then used to update network embeddings (Line 7). The Skip-gram algorithm is chosen as the method for updating node representations.

(a) Walklets()
(b) Walklets()
(c) Walklets()
(d) Walklets()
(e) Walklets()
(f) Walklets()
Figure 4: Vastly different information can be encoded, depending on the scale of representation chosen. Shown in Figures 3(a), 3(b), and 3(c) are the distribution of distances to other vertices from in the Cora network, at different scales of network representation. Coarser representations (such as ) ‘flatten’ the distribution, making larger communities close to the source vertex. Graph heatmap of corresponding distances Figures 3(d), 3(e), and 3(f) show the corresponding heatmap of cosine distance from vertex (shown by arrow) in the Cora network through a series of successively coarser representations. Nearby vertices are colored red and distant vertices are colored blue.
Method Source of Context Nodes Embedding Learning Method
DeepWalk [33] Truncated Random Walks Skip-gram with Hierarchical Softmax
LINE [41] 1-hop and 2-hop Neighbors Skip-gram with Negative Sampling
Node2vec [21] Biased Truncated Random Walks Skip-gram with Negative Sampling
Walklets [34] where Skip-gram with Hierarchical Softmax
GraRep [5] where Matrix Factorization
GraphAttention [2] where Graph Likelihood
SDNE [48] 1-hop and 2-hop Neighbors Deep Autoencoder
DNGR [6] Random surfing Stacked Denoising Autoencoder
Table 1: Unsupervised network embedding methods categorized by source of context nodes and method for representation learning.

Most subsequent work on graph embeddings has followed this two-phase framework proposed in DeepWalk, with variations in both phases. Table 1 summarizes several network embedding methods categorized by different definitions of context nodes and different methods for learning embeddings:

  • LINE [41] adopts a breadth-first search strategy for generating context nodes: only nodes which are at most two hops away from a given node are considered as its neighboring nodes. Besides, it uses negative sampling [30] to optimize the Skip-gram model, in contrast to the hierarchical softmax [29] used in DeepWalk.

  • Node2vec [21] is an extension of DeepWalk which introduces a biased random walking procedure which combines BFS style and DFS style neighborhood exploration.

  • Walklets [34] shows that DeepWalk learns network embeddings from a weighted combination of . In particular, DeepWalk is always more biased toward than if . To avoid the above shortcomings, Walklets proposes to learn multiscale network embeddings from each of . Since the time complexity of computing is at least quadratic in the number of nodes in the network, Walklets approximates by skipping over nodes in short random walks. It further learns network embeddings from different powers of to capture the network’s structural information at different granularities.

  • GraRep [5] similarly exploits node co-occurrence information at different scales by raising the graph adjacency matrix to different powers. Singular value decomposition (SVD) [20] is applied to the powers of the adjacency matrix to obtain low-dimensional representation of nodes. There are two major differences between Walklets and GraRep. First, GraRep computes the exact content of , while Walklets approximates it. Second, GraRep adopts SVD to obtain node embeddings with exact factorization, while Walklets uses the Skip-gram model. Interestingly, Levy and Goldberg [26] proves that skip-gram with negative sampling (SGNS) is implicitly factorizing the PMI matrix between nodes and respective context nodes. To sum up, GraRep generates network embedding using a process with less noise, but Walklets proves much more scalable.

The models discussed so far rely on some manually chosen parameters to control the distribution of context nodes of each node in the graph. For DeepWalk, the window size determines the context node. Furthermore, the Skip-gram model used has hidden hyper-parameters that determine the importance of an example, based on how far in the context it is. For Walklets and GraRep, the power to which the graph adjacency matrix is raised to should be decided beforehand. Selecting these hyperparameters is non-trivial, since they will significantly affect the performance of the network embedding algorithms.

GraphAttention [2] proposes an attention model that learns a multi-scale representation which best predicts links in the original graph. Instead of pre-determining hyperparameters to control the context nodes distribution, GraphAttention automatically learns the attention over the power-series of the graph transition matrix. Formally, let be the co-occurence matrix derived from random walks and be the initial random walk starting positions matrix. GraphAttention parameterizes the expectation of with a probability distribution :


This probability distribution can then be learned by backpropagation from the data itself, e.g. by modeling it as the output a softmax layer with parameters (, …, ),


This allows every graph to learn its own distribution Q with a bespoke sparsity and decay form.

The expressiveness of deep learning methods makes them suitable for embedding networks. SDNE [48] learns node representations that preserve the proximity between 2-hop neighbors with a deep autoencoder. It further preserves the proximity between adjacent nodes by minimizing the Euclidean distance between their representations. DNGR [6] is another deep neural network-based method for learning network embeddings. They adopt a random surfing strategy for capturing graph structural information. They further transform these structural information into a PPMI matrix, and train a stacked denoising autoencoder (SDAE) to embed nodes.

All of these papers focus on embedding simple undirected graphs. In the next section, we will introduce methods on embedding graphs with different properties, such as directed graphs and signed graphs.

2.1 Directed Graph Embeddings

The graph embeddings discussed in the previous section were designed to operate on undirected networks. However, as shown in [66], they can be naturally generalized to directed graphs by employing directed random walks as the training data for the network. Several other recent methods have also been proposed for modeling directed graphs.

HOPE [31] is a graph embedding method specifically designed for directed graphs. HOPE is a general framework for asymmetric transitivity preserving graph embedding, which incorporates several popular proximity measurements such as Katz index, rooted PageRank and common neighbors as special cases. The optimization goal of HOPE is efficiently solved using generalized SVD.

Abu-El-Haija et al. [1] propose two separate representations for each node, one where it is a source, and the other where it is a destination. In this sense, edge embeddings could be thought of as simply a concatenation of the source embedding of the source and the destination of the destination. These ‘edge representations’ (discussed further in Section 2.2), implicitly preserve the directed nature of the graph.

2.2 Edge Embeddings

Tasks like link prediction require accurate modeling of graph edges. An unsupervised way of constructing a representation for edge is to apply a binary operator over and :


In node2vec [21], several binary operators are considered, such as average, Hardmard product, L1 distance and L2 distance. However, these symmetric binary operators always assign same representations to edges and , ignoring the direction of edges.

To alleviate this problem, Abu-El-Haija et al. [1] propose to learn edge representations via low-rank asymmetric projections. Their method consists of three steps. In the first step, embedding vectors are learned for every with node2vec. Then, a DNN is learned to reduce the dimensionality of embedding vectors. Finally, for each node pair , a low-rank asymmetric projection transforms and into their corresponding representations as source and destination nodes, and is represented as:


where is the low-rank projection matrix. The model’s architecture is further illustrated in Figure 5.

Figure 5: Depiction of the edge representation method in [1]. On the left: a graph, showing a random walk in dotted-red, where nodes are “close” in the walk (i.e. within a configurable context window parameter). Their method accesses the trainable embeddings and for the nodes and feed them as input to Deep Neural Network (DNN) . The DNN outputs manifold coordinates and for nodes and , respectively. A low-rank asymmetric projection transforms and to their source and destination representations, which are used by to represent an edge.

2.3 Signed Graph Embeddings

Recall that in a signed graph, an edge with weight of 1 denotes a positive link between nodes, whereas an edge with weight of -1 denotes a negative link.

SiNE [49] is a deep neural network-based model for learning signed network embeddings. Based on the structural balance theory, nodes should be closer to their friends (linked with positive edges) than their foes (linked with negative edges). SiNE preserves this property by maximizing the margin between the embedding similarity of friends and the embedding similarity of foes. Formally, given a triplet where and have a positive link while and have a negative link, the following property holds:


where is a similarity metric between node embeddings and is a tunable margin. However, negative links are much rarer than positive links in real social networks. Thus, such a triplet may not exist for many nodes in the network, since there are only positive links in their 2-hop networks. To solve this problem, an additional virtual node is connected to such nodes with a negative link. Similarly, given a triplet where a positive link connects and while a negative link connects and , we have another objective function:


The node embeddings are learned by jointly minimizing Eq. 13 and Eq. 14.

SNE [59] is a log-bilinear model for signed network embedding. SNE predicts the representation of a target node by linearly combines the representation of its context nodes. To capture the signed relationships between nodes, two signed-type vectors are incorporated into the log-bilinear model.

2.4 Subgraph Embeddings

Another branch of research concerns embedding larger-scale components of graphs, such as graph sub-structures or whole graphs. Yanardag and Vishwanathan [54] present the deep graph kernel, which is a general framework for modeling sub-structure similarity in graphs. Traditionally, the kernel between two graphs and is given by


where represents dot product in a RKHS .

Many sub-structures have been developed to compute this kernel, such as graphlets, subtrees and shortest paths. However, these representations fail to uncover the similarity between different but similar sub-structures. That is, even if two graphlets only differ by one edge or one node, they are still considered to be totally different. This kernel definition causes the diagonal dominance problem: a graph is only similar to itself, but not to any other graph. To overcome this problem, Yanardag and Vishwanathan [54] present an alternative kernel definition as follows:


where is the similarity matrix between all pairs of sub-structures in the input graph.

To build , their algorithm first generates the co-occurrence matrix of graph sub-structures. Then, the Skip-gram model is trained on the co-occurrence matrix to obtain the latent representation of sub-structures, which is subsequently used to compute .

2.5 Meta-strategies for Improving Network Embeddings

Despite the success of neural methods for network embedding, all methods to date have several shared weaknesses. Firstly, they are all local approaches – limited to the structure immediately around a node. DeepWalk and node2vec adopt short random walks to explore the local neighborhoods of nodes, while LINE is concerned with even closer relationships (nodes at most two hops away). This focus on local structure implicitly ignores long-distance global relationships, and the learned representations can fail to uncover important global structural patterns. Secondly, they all rely on a non-convex optimization goal solved using stochastic gradient descent [29] which can become stuck in a local minima (e.g. perhaps as a result of a poor initialization). In other words, these techniques for learning network embedding can accidentally learn embedding configurations which disregard important structural features of their input graph.

To solve these problems, HARP[11] proposes a meta strategy for embedding graph datasets which preserves higher-order structural features. HARP recursively coalesces the nodes and edges in the original graph to get a series of successively smaller graphs with similar structure. These coalesced graphs, each with a different granularity, provide us a view of the original graph’s global structure. Starting from the most simplified form, each graph is used to learn a set of initial representations which serve as good initializations for embedding the next, more detailed graph. This process is repeated until we get an embedding for each node in the original graph.

(a) Can_187
(b) LINE
(c) HARP
(d) Poisson 2D
(e) LINE
(f) HARP
Figure 6: Comparison of two-dimensional embeddings from LINE and HARP, for two distinct graphs. Observe how HARP’s embedding better preserves the higher order structure of a ring and a plane.

HARP is a general meta-strategy to improve all of the state-of-the-art neural algorithms for embedding graphs, including DeepWalk, LINE, and Node2vec. The effectiveness of the HARP paradigm is illustrated in Figure 6, by visualizing the two-dimension embeddings from LINE and the improvement to it, HARP(LINE). Each of the small graphs we consider have an obvious global structure (that of a ring (5(a)) and a grid (5(d))) which is easily exposed by a force directed layout [23]. The center figures represent the two-dimensional embedding obtained by LINE for the ring (5(b)) and grid (5(e)). In these embeddings, the global structure is lost (i.e. that is, the ring and plane are unidentifiable). However, the embeddings produced by using HARP to improve LINE (right) capture both the local and global structure of the given graphs (5(c), 5(f)).

3 Attributed Network Embeddings

The methods we have discussed above leverage only network structural information to obtain network embedding. However, nodes and edges in real-world networks are often associated with additional features, which are called attributes. For example, in a social network site such as Twitter, the textual contents posted by users (nodes) are available. Therefore, it is desirable that network embedding methods also learn from the rich content in node attributes and edge attributes. In the discussion below, we assume that attributes are only associated with nodes, since most existing work focus on exploiting node attributes. Different strategies have been proposed for different types of attributes. In particular, researchers are interested in two categories of attributes: high-level features such as text or images, and node labels.

These high-level features are usually high-dimensional sparse features of the nodes, so it is a common practice to use unsupervised text embedding or image embedding models to convert these sparse features into dense embedding features. Once the embedding features are learned, the major challenge is how to incorporate them into an existing network embedding framework.

TADW [55] studies the case when nodes are associated with text features. The authors of TADW first prove that DeepWalk is essentially factorizing a transition probability matrix into two low-dimensional matrices and where . Inspired by this result, TADW incorporates the text feature matrix into the matrix factorization process, by factorizing into the product of , and . Finally, and are concatenated as the latent representations of nodes.

Another idea is to jointly model network stucture and node features. Intuitively, in addition to enforcing the embedding similarity between nodes in the same neighborhood, we should also enforce the embedding similarity between nodes with similar feature vectors. CENE [40] is a network embedding method which jointly models network structure and textual content in nodes. CENE treats text content as a special type of node, and leverages both node-node links and node-content links for node embedding. The optimization goal is to jointly minimize the loss on both types of links. HSCA [61] is a network embedding method for attributed graphs which models homophily, network topological structure and node features simultaneously.

Besides textual attributes, node labels are another important type of attribute. In a citation network, the labels associated with papers might be their venue or year of publication. In a social network, the labels of people cmight be the groups they belong to. A typical approach to incorporating label information is jointly optimizing the loss for generating node embeddings and for predicting node labels. GENE [12] considers the situation when group information is associated with nodes. GENE follows the idea of DeepWalk, but instead of only predicting context nodes in random walk sequences, it also predicts the group information of context nodes as a part of the optimization goal. Wang et al. [50] present a modularized nonnegative matrix factorization-based method for network embedding which preserves the community structures within network. On the level of nodes, their model preserves first-order and second-order proximities between nodes with matrix factorization; on the level of communities, a modularity constraint term is applied during the matrix factorization process for community detection.

It is also common in real-world networks that node labels are only available for a portion of nodes. Semi-supervised network embedding methods have been developed for joint learning on both node labels and network structure in such case. Planetoid [57] is a semi-supervised network embedding method which learns node representations by jointly predicting the label and the context nodes for each node in the graph. It works under both inductive and transductive scenarios. Max-margin DeepWalk (MMDW) [46] is a semi-supervised approach which learns node representations in a partially labeled network. MMDW consists of two parts: the first part is a node embedding model based on matrix factorization, while the second part takes in the learned representations as features to train a max-margin SVM classifier on the labeled nodes. By introducing biased gradients, the parameters in both parts can be updated jointly.

4 Heterogeneous Network Embeddings

Recall that heterogeneous networks have multiple classes of nodes or edges. To model the nodes and edges of different types, most network embedding methods we introduce below learn node embeddings via jointly minimizing the loss over each modality. These methods either directly learn all node embeddings in the same latent space, or construct the embeddings for each modality beforehand and then map them to the same latent space.

Chang et al. [7] present a deep embedding framework for heterogeneous networks. Their model first constructs a feature representation for each modality (such as image, text), then maps the embeddings of different modalities into the same embedding space. The optimization goal is to maximize the similarity between the embeddings of linked nodes, while minimizing that of the unlinked nodes. Note that edges can be between both nodes within the same modality as well as nodes from different modalities.

Zhao et al. [64] is another such framework for constructing node representations in a heterogeneous network. Specifically, they consider the Wikipedia network with three types of nodes: entities, words and categories. The co-occurrence matrices between same and different types of nodes are built, and the representations for entities, words and categories are jointly learned from all matrices using coordinate matrix factorization.

Li et al. [27] propose a neural network model for learning user representations in a heterogeneous social network. Their method jointly models user-generated texts, user networks and multifaceted relationships between users and user attributes.

HEBE [22] is an algorithm for embedding large-scale heterogeneous event networks, where an event is defined as the interaction between a set of nodes (possibly of different types) in the network. While previous work decomposes an event into the pairwise interaction between each pair of nodes involved in the event, HEBE treat the whole event as a hyperedge and preserves the proximity between all participating nodes simultaneously. Specifically, for each node in a hyperedge, HEBE considers it as the target node and the remaining nodes in the hyperedge as context nodes. Thus, the underlying optimization goal is to predict the target node given all context nodes.

EOE [53] is a network embedding method for coupled heterogeneous networks, where two homogeneous networks are connected with inter-network edges. EOE learns latent node representations for both networks, and utilize a harmonious embedding matrix to transform the representations of different networks into the same space.

Besides modeling heterogeneous nodes and edges jointly, another promising direction of work is on extending random walks and embedding learning methods to a heterogeneous scenario. Metapath2vec [17] is an extension to DeepWalk which works for heterogeneous networks. For constructing random walks, metapath2vec uses meta-path-based walks which capture the relationship between different types of nodes. For learning representation from random walk sequences, they propose heterogeneous Skip-gram which considers node type information during model optimization.

5 Applications of Network Embeddings

Network embeddings have been widely employed in practice, due to their ease of use in turning adjacency data into actionable features. Here we review several representative applications of network embeddings to demonstrate how they can be used:

5.1 Knowledge Representation

The problem of knowledge representation is concerned with encoding facts about the world using short sentences (or tuples) composed of subjects, predicates, and objects. While it can be viewed as strictly as a heterogeneous network, it is an important enough application area to mention here in its own right:

  • GenVector [58] studies the problem of learning social knowledge graphs, where the goal is to connect online social networks to knowledge bases. Their multi-modal Bayesian embedding model utilizes DeepWalk for generating user representations in social networks.

  • RDF2Vec [38] is an approach for learning latent entity representations in Resource Description Framework (RDF) graphs. RDF2Vec first converts RDF graphs into sequences of graph random walks and Weisfeiler-Lehman graph kernels, and then adopt CBOW and Skip-gram models on the sequences to build entity representations.

5.2 Recommender Systems

Another branch of work attempts to incorporate network embeddings into recommender systems. Naturally, the interactions between users, users’ queries and items altogether form a heterogeneous network which encodes the latent preferences of users over items. Network embedding on such interaction graphs could serve as an enhancement to recommender systems.

  • Chen et al. [8] exploit the usage of social listening graph to enhance music recommendation models. They utilize DeepWalk to learn latent node representations in the social listening graph, and incorporate these latent representations into factorization machines.

  • Chen et al. [9] propose Heterogeneous Preference Embedding to embed user preference and query intention into low-dimensional vector space. With both user preference embedding and query embedding available, recommendations can be made based on the similarity between items and queries.

5.3 Natural Language Processing

State-of-the-art network embedding methods are mostly inspired by advances in the field of natural language processing, especially neural language models. At the same time, network embedding methods also lead to better modeling of human language.

  • PLE [37] studies the problem of label noise reduction in entity typing. Their model jointly learns the representations of entity mentions, text features and entity types in the same feature space. These representations are further used to estimate the type-path for each training example.

  • CANE [45] is a context-aware network embedding framework. They argue that one node may exhibit different properties when interacting with different neighbors, thus its embedding with respect to these neighbors should be different. CANE achieves this goal by employing mutual attention mechanism.

  • Fang et al. [18] propose a community-based question answering (cQA) framework which leverages the social interactions in the community for better question-answering matching. Their framework treats users, questions and answers and the interactions between them as a heterogeneous network and trains a deep neural network on random walks in the network.

  • Zhao et al. [65] study the problem of expert finding in community-based question answering (cQA) site. Their method adopts the random-walk method in DeepWalk for embedding social relations between users and RNNs for modeling users’ relative quality rank to questions.

5.4 Social Network Analysis

Social networks are prevailing in the real world, and it is not suprising that network embedding methods have become popular in social network analysis. Network embeddings on social network have prove to be powerful features for a wide spectrum of applications, leading to improved performance on a lot of downstream tasks.

  • Perozzi et al. [35] study the problem of predicting the exact age of users in social networks. They learn the user representations in social networks with DeepWalk, and adopts linear regression on these user representations for age prediction.

  • Yang et al. [56] propose a neural network model for modeling social networks and mobile trajectories simultaneously. They adopt DeepWalk to generate node embeddings in social networks and the RNN and GRU models for generating mobile trajectories.

  • Dallmann et al. [16] show that by learning Wikipedia page representations from both the Wikipedia link network and Wikipedia click stream network with DeepWalk, they can obtain concept embeddings of higher quality compared to counting-based methods on the Wikipedia networks.

  • Liu et al. [28] propose Input-output Network Embedding (IONE), which use network embeddings to align users across different social networks. IONE achieves this by preserving the proximity of users with similar followers and followees in a common embedding space.

  • Chen and Skiena [14] demonstrate the efficacy of network embedding methods in measuring similarity between historical figures. They construct a network between historical figures from the interlinks between their Wikipedia pages, and use DeepWalk to obtain vector representations of historical figures. It is shown that the similarity between the DeepWalk representations of historical figures can be used as an effctive decent similarity measurement.

  • DeepBrowse [10] is an approach for browsing through large lists in the absence of a predefined hierarchy. DeepBrowse is defined by the interaction of two fixed, globally-defined permutations on the space of objects: one ordering the items by similarity, the second based on magnitude or importance. The similarity between items is computed by using DeepWalk embeddings generated over the interaction graph of objects.

  • TransNet [47] is a translation-based network embedding model which exploits the rich semantic information in graph edges for relation prediction on edges. TransNet treats the interactions between nodes as a translation operation and further employ a deep autoencoder to construct edge representations.

5.5 Other Applications

  • Geng et al. [19] and Zhang et al. [62] develop deep neural network models which learns distributed representations of both users and images from an user-image co-occurrence network. The representation learning process in the network is analogous to that of DeepWalk [33], except that they also incorporate image features extracted with a DCNN into the optimization process.

  • Wu et al. [52] treat the click data collected from users’ searching behavior in image search engines as a heterogeneous graph. The nodes in the click graph are text queries and images returned as search results, while the edges indicates the click count of an image given a search query. By proposing a neural network model based on truncated random walks, their method learns multimodal representations of text and images, which are shown to boost cross-modal retrieval performance on unseen queries or images.

  • Zhang et al. [63] apply DeepWalk to large-scale social image-tag collections to learn both image features and word features in a unified embedding space.

These applications only represent the tip of the iceberg. The future of network embeddings seems bright, with new algorithmic approaches producing better embeddings to feed increasingly sophisticated neural networks as Deep Learning continues to grow in popularity and importance.

6 Conclusions and Future Directions

Network embedding is an exciting and rapidly growing research area which attracts researchers from various communities, especially data mining, machine learning and natural language processing. While most work concerned about general methods for network embedding, we argue that the applications of network embedding is even more underresearched. We anticipate a large body of work on additional applications of network embeddings, such as improving the performance of natural language processing and information retrieval models, mining biology network and social networks, to name a few.

Also, much work has been done for graphs which possess different properties and from different domains. In terms of graph properties, various methods are proposed for directed graphs, signed graphs, heterogeneous graphs and attributed graphs. In terms of application domains, network embedding methods are applied to a wide spectrum of graphs including knowledge graphs, biology graphs and social networks. However, doubtlessly much more work can be done on this front by exploiting the unique characteristics of these graphs.

6.1 The search for the right context

Inspired by the two-phase network embedding learning framework presented in DeepWalk, various strategies have been proposed for searching for the right context, as discussed in Table 1. However, most of these strategies relies on a rigid definition of context nodes identical for all networks, which is not desirable.

Under this background, there is much effort recently on unifying different network embedding under a general framework [13, 36]. GEM-D [13] decomposes graph embedding algorithms into three building blocks: node proximity function, warping function and loss function. They show that algorithms such as Laplacian Eigenvectors, DeepWalk, LINE, and node2vec can all be unified under this framework. By testing different design choices for each building block on real-world graphs, they pick the triple which works the best empirically: the combination of the finite-step transition matrix, exponential warping function and warped Frobenius norm loss. However, such design decisions are purely made based on models’ empirical performance on a limited number of networks, which may not work well for all networks.

A promising approach is the attention model recently proposed in GraphAttention[2]. By parameterizing the attention over the power series of the transition matrix, GraphAttention automatically learns different attention parameters for different networks.

6.2 Improved Losses / Optimization Models

Another issue with the neural embedding methods is their dependence upon general loss functions and optimization models, such as Skip-gram. These optimization goals and models are not tuned for any particular task. As a result, though the learned network embeddings have been proven to achieve competitive performance on a variety of tasks such as node classification and link prediction, they are suboptimal when compared with end-to-end embeddings methods designed specifically for a task.

Thus, another future direction for network embedding algorithms is to design loss functions and optimization models for a specific task. From this perspective, the semi-supervised network embedding methods can been seen as specifically designed for the node classification task. Another attempt is made by Abu-El-Haija et al. [1], where the graph likelihood is proposed as a novel objective tuned for link prediction. Given a training graph , its graph likelihood is defined as a product of an edge estimate over all node pairs:


where is a trainable edge estimator.


  • [1] Sami Abu-El-Haija, Bryan Perozzi, and Rami Al-Rfou. Learning edge representations via low-rank asymmetric projections. CIKM ’17, 2017.
  • [2] Sami Abu-El-Haija, Bryan Perozzi, Rami Al-Rfou, and Alex Alemi. Watch your step: Learning graph embeddings through attention. arXiv preprint arXiv:1710.09599, 2017.
  • [3] Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps and spectral techniques for embedding and clustering. In Advances in neural information processing systems, pages 585–591, 2002.
  • [4] Yoshua Bengio, Réjean Ducharme, Pascal Vincent, and Christian Jauvin. A neural probabilistic language model. Journal of machine learning research, 3(Feb):1137–1155, 2003.
  • [5] Shaosheng Cao, Wei Lu, and Qiongkai Xu. Grarep: Learning graph representations with global structural information. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, pages 891–900. ACM, 2015.
  • [6] Shaosheng Cao, Wei Lu, and Qiongkai Xu. Deep neural networks for learning graph representations. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, pages 1145–1152. AAAI Press, 2016.
  • [7] Shiyu Chang, Wei Han, Jiliang Tang, Guo-Jun Qi, Charu C Aggarwal, and Thomas S Huang. Heterogeneous network embedding via deep architectures. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 119–128. ACM, 2015.
  • [8] Chih-Ming Chen, Po-Chuan Chien, Yu-Ching Lin, Ming-Feng Tsai, and Yi-Hsuan Yang. Exploiting latent social listening representations for music recommendations. In Proc Ninth ACM Int. Conf. Recommender Syst. Poster, 2015.
  • [9] Chih-Ming Chen, Ming-Feng Tsai, Yu-Ching Lin, and Yi-Hsuan Yang. Query-based music recommendations via preference embedding. In Proceedings of the 10th ACM Conference on Recommender Systems, pages 79–82. ACM, 2016.
  • [10] Haochen Chen, Arvind Ram Anantharam, and Steven Skiena. Deepbrowse: Similarity-based browsing through large lists. In International Conference on Similarity Search and Applications, pages 300–314. Springer, 2017.
  • [11] Haochen Chen, Bryan Perozzi, Yifan Hu, and Steven Skiena. Harp: Hierarchical representation learning for networks. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence. AAAI Press, 2018.
  • [12] Jifan Chen, Qi Zhang, and Xuanjing Huang. Incorporate group information to enhance network embedding. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, pages 1901–1904. ACM, 2016.
  • [13] Siheng Chen, Sufeng Niu, Leman Akoglu, Jelena Kovačević, and Christos Faloutsos. Fast, warped graph embedding: Unifying framework and one-click algorithm. arXiv preprint arXiv:1702.05764, 2017.
  • [14] Yanqing Chen, Bryan Perozzi, and Steven Skiena. Vector-based similarity measurements for historical figures. Information Systems, 64:163–174, 2017.
  • [15] Fan RK Chung. Spectral graph theory. Number 92. American Mathematical Soc., 1997.
  • [16] Alexander Dallmann, Thomas Niebler, Florian Lemmerich, and Andreas Hotho. Extracting semantics from random walks on wikipedia: Comparing learning and counting methods. In Tenth International AAAI Conference on Web and Social Media, 2016.
  • [17] Yuxiao Dong, Nitesh V Chawla, and Ananthram Swami. metapath2vec: Scalable representation learning for heterogeneous networks. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 135–144. ACM, 2017.
  • [18] Hanyin Fang, Fei Wu, Zhou Zhao, Xinyu Duan, Yueting Zhuang, and Martin Ester. Community-based question answering via heterogeneous social network learning. In Thirtieth AAAI Conference on Artificial Intelligence, 2016.
  • [19] Xue Geng, Hanwang Zhang, Jingwen Bian, and Tat-Seng Chua. Learning image and user features for recommendation in social networks. In Proceedings of the IEEE International Conference on Computer Vision, pages 4274–4282, 2015.
  • [20] Gene H Golub and Christian Reinsch. Singular value decomposition and least squares solutions. Numerische mathematik, 14(5):403–420, 1970.
  • [21] Aditya Grover and Jure Leskovec. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 855–864. ACM, 2016.
  • [22] Huan Gui, Jialu Liu, Fangbo Tao, Meng Jiang, Brandon Norick, and Jiawei Han. Large-scale embedding learning in heterogeneous event data. 2016.
  • [23] Yifan Hu. Efficient, high-quality force-directed graph drawing. Mathematica Journal, 10(1):37–71, 2005.
  • [24] Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.
  • [25] Joseph B Kruskal and Myron Wish. Multidimensional scaling, volume 11. Sage, 1978.
  • [26] Omer Levy and Yoav Goldberg. Neural word embedding as implicit matrix factorization. In Advances in neural information processing systems, pages 2177–2185, 2014.
  • [27] Jiwei Li, Alan Ritter, and Dan Jurafsky. Learning multi-faceted representations of individuals from heterogeneous evidence using neural networks. arXiv preprint arXiv:1510.05198, 2015.
  • [28] Li Liu, William K Cheung, Xin Li, and Lejian Liao. Aligning users across social networks using network embedding. In IJCAI, pages 1774–1780, 2016.
  • [29] Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781, 2013.
  • [30] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. Distributed representations of words and phrases and their compositionality. In Advances in neural information processing systems, pages 3111–3119, 2013.
  • [31] Mingdong Ou, Peng Cui, Jian Pei, Ziwei Zhang, and Wenwu Zhu. Asymmetric transitivity preserving graph embedding. In Proc. of ACM SIGKDD, pages 1105–1114, 2016.
  • [32] Jeffrey Pennington, Richard Socher, and Christopher D Manning. Glove: Global vectors for word representation. In EMNLP, volume 14, pages 1532–1543, 2014.
  • [33] Bryan Perozzi, Rami Al-Rfou, and Steven Skiena. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 701–710. ACM, 2014.
  • [34] Bryan Perozzi, Vivek Kulkarni, Haochen Chen, and Steven Skiena. Don’t walk, skip! online learning of multi-scale network embeddings. In 2017 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM). IEEE/ACM, 2017.
  • [35] Bryan Perozzi and Steven Skiena. Exact age prediction in social networks. In Proceedings of the 24th International Conference on World Wide Web, pages 91–92. ACM, 2015.
  • [36] Jiezhong Qiu, Yuxiao Dong, Hao Ma, Jian Li, Kuansan Wang, and Jie Tang. Network embedding as matrix factorization: Unifying deepwalk, line, pte, and node2vec. arXiv preprint arXiv:1710.02971, 2017.
  • [37] Xiang Ren, Wenqi He, Meng Qu, Clare R Voss, Heng Ji, and Jiawei Han. Label noise reduction in entity typing by heterogeneous partial-label embedding. arXiv preprint arXiv:1602.05307, 2016.
  • [38] Petar Ristoski and Heiko Paulheim. Rdf2vec: Rdf graph embeddings for data mining. In International Semantic Web Conference, pages 498–514. Springer, 2016.
  • [39] Sam T Roweis and Lawrence K Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290(5500):2323–2326, 2000.
  • [40] Xiaofei Sun, Jiang Guo, Xiao Ding, and Ting Liu. A general framework for content-enhanced network representation learning. arXiv preprint arXiv:1610.02906, 2016.
  • [41] Jian Tang, Meng Qu, Mingzhe Wang, Ming Zhang, Jun Yan, and Qiaozhu Mei. Line: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, pages 1067–1077. ACM, 2015.
  • [42] Lei Tang and Huan Liu. Relational learning via latent social dimensions. In Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 817–826. ACM, 2009.
  • [43] Lei Tang and Huan Liu. Leveraging social media networks for classification. Data Mining and Knowledge Discovery, 23(3):447–478, 2011.
  • [44] Joshua B Tenenbaum, Vin De Silva, and John C Langford. A global geometric framework for nonlinear dimensionality reduction. Science, 290(5500):2319–2323, 2000.
  • [45] Cunchao Tu, Han Liu, Zhiyuan Liu, and Maosong Sun. Cane: Context-aware network embedding for relation modeling. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), volume 1, pages 1722–1731, 2017.
  • [46] Cunchao Tu, Weicheng Zhang, Zhiyuan Liu, and Maosong Sun. Max-margin deepwalk: discriminative learning of network representation. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI 2016), pages 3889–3895, 2016.
  • [47] Cunchao Tu, Zhengyan Zhang, Zhiyuan Liu, and Maosong Sun. Transnet: Translation-based network representation learning for social relation extraction. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, IJCAI, pages 19–25, 2017.
  • [48] Daixin Wang, Peng Cui, and Wenwu Zhu. Structural deep network embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 1225–1234. ACM, 2016.
  • [49] Suhang Wang, Jiliang Tang, Charu Aggarwal, Yi Chang, and Huan Liu. Signed network embedding in social media. SDM, 2017.
  • [50] Xiao Wang, Peng Cui, Jing Wang, Jian Pei, Wenwu Zhu, and Shiqiang Yang. Community preserving network embedding. 2017.
  • [51] Svante Wold, Kim Esbensen, and Paul Geladi. Principal component analysis. Chemometrics and intelligent laboratory systems, 2(1-3):37–52, 1987.
  • [52] Fei Wu, Xinyan Lu, Jun Song, Shuicheng Yan, Zhongfei Mark Zhang, Yong Rui, and Yueting Zhuang. Learning of multimodal representations with random walks on the click graph. IEEE Transactions on Image Processing, 25(2):630–642, 2016.
  • [53] Linchuan Xu, Xiaokai Wei, Jiannong Cao, and Philip S Yu. Embedding of embedding (eoe): Joint embedding for coupled heterogeneous networks. In Proceedings of the Tenth ACM International Conference on Web Search and Data Mining, pages 741–749. ACM, 2017.
  • [54] Pinar Yanardag and SVN Vishwanathan. Deep graph kernels. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 1365–1374. ACM, 2015.
  • [55] Cheng Yang, Zhiyuan Liu, Deli Zhao, Maosong Sun, and Edward Y Chang. Network representation learning with rich text information. In IJCAI, pages 2111–2117, 2015.
  • [56] Cheng Yang, Maosong Sun, Wayne Xin Zhao, Zhiyuan Liu, and Edward Y Chang. A neural network approach to joint modeling social networks and mobile trajectories. arXiv preprint arXiv:1606.08154, 2016.
  • [57] Zhilin Yang, William Cohen, and Ruslan Salakhutdinov. Revisiting semi-supervised learning with graph embeddings. arXiv preprint arXiv:1603.08861, 2016.
  • [58] Zhilin Yang, Jie Tang, and William Cohen. Multi-modal bayesian embeddings for learning social knowledge graphs. arXiv preprint arXiv:1508.00715, 2015.
  • [59] Shuhan Yuan, Xintao Wu, and Yang Xiang. Sne: Signed network embedding. arXiv preprint arXiv:1703.04837, 2017.
  • [60] Wayne W Zachary. An information flow model for conflict and fission in small groups. Journal of anthropological research, 33(4):452–473, 1977.
  • [61] Daokun Zhang, Jie Yin, Xingquan Zhu, and Chengqi Zhang. Homophily, structure, and content augmented network representation learning. In Data Mining (ICDM), 2016 IEEE 16th International Conference on, pages 609–618. IEEE, 2016.
  • [62] Hanwang Zhang, Xindi Shang, Huanbo Luan, Meng Wang, and Tat-Seng Chua. Learning from collective intelligence: Feature learning using social images and tags. ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM), 13(1):1, 2016.
  • [63] Hanwang Zhang, Xindi Shang, Huanbo Luan, Yang Yang, and Tat-Seng Chua. Learning features from large-scale, noisy and social image-tag collection. In Proceedings of the 23rd ACM international conference on Multimedia, pages 1079–1082. ACM, 2015.
  • [64] Yu Zhao, Zhiyuan Liu, and Maosong Sun. Representation learning for measuring entity relatedness with rich information. In IJCAI, pages 1412–1418, 2015.
  • [65] Zhou Zhao, Qifan Yang, Deng Cai, Xiaofei He, and Yueting Zhuang. Expert finding for community-based question answering via ranking metric network learning. In IJCAI, pages 3000–3006, 2016.
  • [66] Chang Zhou, Yuqiong Liu, Xiaofei Liu, Zhongyi Liu, and Jun Gao. Scalable graph embedding for asymmetric proximity. In AAAI, pages 2942–2948, 2017.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
Add comment
Loading ...
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test description