# Identifying Transition States of Chemical Kinetic Systems using Network Embedding Techniques

## Abstract

Using random walk sampling methods for feature learning on networks, we develop a method for generating low-dimensional node embeddings for directed graphs and identifying transition states of stochastic chemical reacting systems. We modified objective functions adopted in existing random walk based network embedding methods to handle directed graphs and neighbors of different degrees. Through optimization via gradient ascent, we embed the weighted graph vertices into a low-dimensional vector space while preserving the neighborhood of each node. We then demonstrate the effectiveness of the method on dimension reduction through several examples regarding identification of transition states of chemical reactions, especially for entropic systems.

Email: mercuri6@msu.edu

## 1 Introduction

Networks provide a natural framework for many chemical and biochemical systems, including chemical kinetics and dynamical interactions between biomolecules. The possible stages of a reaction, for example, can be viewed as vertices in a directed graph. Frequently, real-world networks are high-dimensional and complex in structure, which leads to difficulties in interpretation and analysis. In order to take full advantage of the structural information contained in such networks, efficient and robust methods of effective network analysis are essential to produce low dimensional representations while retaining principal correlations of the graph.

Networks analysis is especially useful for understanding behaviors of nonlinear chemical processes such as biochemical switches. Biochemical switches are systems of chemical reactions with multiple steady states. Without random perturbations, the system will spend all its time at the metastable states, where energy is minimized, while the presence of stochastic noise in the dynamics can lead to transitions between metastable states. Difficulties arise in the simulation of such systems as a result of the multiple time scales at work; transitions between steady states are typically rare, but occur rapidly.

The different possible states of the reaction, given by the varying concentrations of reactant and product species, can be viewed as the vertices of a weighted directed graph. The progression of the reaction over time can then be viewed as a Markov jump process, where the jump probabilities are represented by the weights on the edges between each pair of nodes. Then, methods for network analysis can be used to identify interactions and pathways between the components of the biochemical switch. One objective of this paper is to suggest a method to analyze this Markov process using node embedding techniques. This paper will focus on the problem: and are taken as the reactant and product states, respectively. The aim of this paper is to investigate the mechanism by which this transition occurs.

Recently, methods for analyzing networks through feature representation and graph embedding have received increasing attention. For an overview on this subject, a number of recent review papers are available [14, 7]. While much of this research has been focused on undirected graphs, the directed case has also been investigated, for example [10, 9, 24]. In particular, several methods have been proposed that utilize random walks for node embedding.

Node embedding (or feature learning) methods aim to represent each node of a given network as a vector in a low-dimensional continuous vector space, in such a way that information about the nodes and relationships between them are retained. Specifically, nodes that are similar to each other according to some measure in the original graph will also have similar representations in the embedding space. Usually, this means that the embedding will be chosen in a way that maximizes the inner products of embedding vectors for embedded nodes that are similar to each other. Compared with the original complex high-dimensional networks, these low dimensional continuous node representations have the benefit of being convenient to work with for downstream machine learning techniques.

Compared with alternative methods that embed edges or entire graphs instead of nodes, node embeddings are more adaptive for numerous tasks, including node classification, clustering, and link prediction [7]. Node classification is a process by which class labels are assigned to nodes based on a small sample of labelled nodes, where similar nodes have the same labels. Clustering algorithms group the representations of similar nodes together in the target vector space. Link prediction seeks to predict missing edges in an incomplete graph.

The recent interest in feature learning has led to a collection of different node embedding methods, which can be broadly classified [14] by the analytical and computational techniques employed such as matrix factorization, random walk sampling, and deep learning network. The matrix factorization based methods generate embeddings by first using a matrix to represent the relationships between nodes, which include the Laplacian matrix as in the Laplacian eigenmaps [3], the adjacency matrix as in locally linear embedding (LLE) [20] or graph factorization (GF) [1], etc. Though the above methods focus on the case of undirected methods, similar techniques for the directed case have also been suggested [10], where the development of the Laplacian matrix, a combinatorial Laplacian matrix, and Cheeger inequality allow the above approaches to be extended for directed graphs. Factorization based methods such as eigenvalue decomposition can apply for certain matrices (e.g. positive semidefinite), and additional difficulties may arise with scaling for large data sets.

Random walk methods are often used for large graphs, or graphs that cannot be observed in their entirety. The general idea involves first simulating random walks on a network, then using the output to infer relationships between nodes. Random walk methods can be used to study either undirected or directed graphs, although much of the previous work has focused on the undirected case. For example, DeepWalk [19] uses short unbiased random walks to find similarities between nodes, with node pairs that tend to co-occur in the same random walks having higher similarities. This method performs well for detecting homophily – similarity based on node adjacency and shared neighbors. Similar methods, such as node2vec [15], use biased random walks to capture similarity based on both homophily and structural equivalence – similarities in the nodes’ structural roles in the graph. Both of these methods specifically address undirected graphs. A similar approach designed for the directed case is proposed in [9], which uses Markov random walks with transition probabilities given by ratios of edge weights to nodes’ out degrees, together with the stationary distribution of the walks, to measure local relationships between nodes.

Theoretical frameworks for the study of transition events of biochemical processes include Transition State Theory (TST), Transition Path Sampling (TPS), Transition Path Theory (TPT) and Forward Flux Sampling (FFS). The main idea of TST is that in order for the system to move from the reactant state to the product state, the system must pass through a saddle point on the potential energy surface, known as a Transition State [4]. The TPS technique uses Monte Carlo sampling of transition paths to study the full collection of transition paths of a given Markov process [5]. Transition Path Theory (TPT) studies reactive trajectories of a system by considering statistical properties such as rate functions and probability currents between states [6]. Forward flux sampling [2], designed specifically to capture rare switching events in biochemical networks, uses a series of interfaces between the initial and final states to calculate rate constants and generate transition paths. For the purpose of this paper, we will be using TPT, but other such techniques can fit into the method presented here as well.

This paper will suggest a method of using a random walk network embedding approach for node classification and clustering on directed graphs, as well as identification of transition states for the specific case of entropy effects. The method presented reduces the dimension of networks, thereby allowing analysis and interpretation of the system. First, in Section 2 we will introduce some background regarding feature learning on networks, and provide a brief overview of some basic principles from TPT. Then Section 3 will introduce a method for node classification for directed graphs, based on random walk node embedding. Finally, in Section 4, we will study several examples, showing the effectiveness of the method on identifying transition states of entropic systems.

## 2 Background

In this paper, we focus on the mechanism by which a system such as a biochemical switch passes through its energy landscape following a reactive trajectory. A trajectory is said to be reactive if it leaves the reactant state and later arrives at the product state without first returning to state . Such a trajectory can be viewed as a sequence of transitions between metastable states where the energy is locally minimized. In particular, we treat these sequences as Markov jump processes, and study the probability space of these reactive trajectories in order to identify and understand the transition events of the system.

Though this paper will be utilizing the techniques of Transition Path Theory for this purpose, it is also possible to adopt other frameworks, such as Transition Path sampling, Forward Flux sampling, etc., to investigate the transition events of a system. As a Monte Carlo technique, Transition path sampling [5] is essentially a generalization of importance sampling in trajectory space. The basic idea is to perform a random walk in trajectory space, biased so that important regions are visited more often. Forward flux sampling (FFS) [2], which generates trajectories from to using a sequence of nonintersecting parametrizable surfaces, has the advantage that knowledge of the phase space density is not required.

Consider a Markov jump process on a countable state space . Let represent the infinitesimal generator of the process. In other words, for , represents the probability that a process that is in state at time will jump to state during the time interval . Then the entries of are transition rates, satisfying

(2.1) |

Let represent an equilibrium sample trajectory of the Markov process, such that is right-continuous with left limits. At time , let the probability distribution of this process be denoted by . Then the time evolution of follows the forward kolmogorov equation (or master equation)

Denote the time-reversed process by , and the infinitesimal generator of this process by . The stationary probability distribution of both processes, and , is given by the solution to

### 2.1 Network Embedding for Directed Graphs

Networks can have large numbers of vertices and complex structures, which can lead to challenges in network analysis. To overcome these difficulties and improve understanding of these networks, the nodes of a given network can be embedded into a low-dimensional vector space, a process called feature learning. The techniques used in this paper to investigate transition states and pathways is a combination of neural network-based and random walk-based methods for learning latent representations of a network’s nodes, as discussed in [15, 19, 9]. The idea behind these methods is that the nodes in a given network can be embedded into a low-dimensional vector space in such a way that nodes similar to each other according to some measure in the original graph will also have embeddings similar to each other in the embedding space.

Feature learning, or node embedding, requires three main components: an encoder function, a definition of node similarity and a loss function. The encoder function is the function that maps each node to a vector in the embedding space. To find the optimal embeddings, the encoder function can be chosen by minimizing the loss function so that the similarity between nodes in the original network corresponds as close as possible to the similarity between their vector representations in the continuous space. Node similarity can have different interpretations, depending on the network and the task to be accomplished. For example, nodes that are connected, share neighbors, or share a common structural role in the original graph can be claimed to be similar, and to be embedded in the vector space by representations of close similarity. For the embedded vectors, a common approach is to use inner products as the similarity measure in the feature space–that is, if and are two nodes in a graph with representations and in a low-dimensional vector space, then should approximate the similarity between representations and .

#### Factorization Approaches

Many feature learning methods obtain node embeddings via factorization of a matrix containning information regarding relationships between nodes. Locally Linear Embedding (LLE) [20] and Graph Factorization (GF) [1], utilize the node adjacency matrix entry being the weight of edge in the graph. Laplacian Eigenmaps and related methods instead factorize the Laplacian matrix of the whole graph. Optimization of the loss function is typically solve as an eigenvalue problem.

While the methods mentioned here mainly preserve first-order node relationships, i.e., two nodes that are directly connected by an edge are similar, other factorization methods such as GraRep, Cauchy graph embedding and structure preserving embedding can detect higher-order relationships between nodes, or alternative notions of similarity such as structural equivalence [8, 14, 7].

#### Random Walk Approaches

One technique for finding node embeddings is through the use of random walks. This technique has been applied to undirected graphs in methods such as DeepWalk and node2vec [19, 15]. In these methods, the general approach is to use short random walks to determine similarities for each pair of nodes. The notion of node similarity is co-occurence within a random walk; two nodes are similar if there is a higher probability that a walk containing one node will also contain the other. The various methods using this approach differ in the implementation of this general idea: DeepWalk, for example, uses straightforward unbiased walks, while node2vec uses adjustable parameters to bias random walks toward breadth-first or depth-first sampling for a more flexible notion of similarity.

The first stage in the feature learning process is to simulate a certain number of random walks starting from each vertex. Using the results of these random walks, the neighbors of each node can be determined. The objective is to maximize the probability of observing the neighborhood of each node, conditioned on the vector representation of the node. In several of the random walk-based feature representation methods for undirected graphs, this is achieved through the Loss function of the form:

(2.2) |

where is the set of all vertices in the graph, is the neighborhood of the node , and is the encoder function representating the embedded vector. If we assume that the probabilities of observing a particular neighbor are independent, we can have

(2.3) |

Since these methods typically aim to maximize the inner products for neighboring nodes, they often utilize a softmax function to model each conditional probability:

For networks where is large, techniques such as hierarchical softmax [19] or negative sampling [18] can be used to increase efficiency. To determine the optimal embedding function , the Loss function can be optimized using gradient ascent respect to parameters of .

In [9], a random walk-based method specifically intended for directed graphs is proposed. The function to be optimized in this method, when embedding into , is

where is the one-dimensional embedding of , is the stationary probability for the node and represents the random walk’s transition probability from to . The transition probabilities are calculated according to

where is the weight on the edge . Note that this method, though it uses transition probabilities of a random walk, does not actually require simulation of random walks, since all the summations can be done deterministically.

According to the spectral graph theory [10], a circulation on a directed graph is a function such that, for each , . The Laplacian of a directed graph is therefore defined as

where is a diagonal matrix with diagonal entries given by for some circulation function . The following definition of the combinatorial Laplacian is also due to Chung[10]:

where is the transition matrix, i.e. , is the diagonal matrix of the stationary distribution, i.e. . Note that the combinatorial Laplacian is symmetric and semi-positive definite. It is shown in [9] that

where . This result demonstrates that this method is analogous to the Laplacian-based methods for undirected graphs, e.g. Laplacian eigenmaps [14].

Random walk methods have been shown to perform well compared to other methods, and can be useful for a variety of different notions of node similarity. Previous work has shown them to be robust, efficient, and capable of completing a diverse range of tasks, including node classification, link prediction, clustering, and more [14, 7, 19, 15]. Because they do not examine the entire graph at once, these methods are also useful for very large networks and networks that cannot be observed in their entirety.

#### Neural Network Approaches

A third class of graph embedding methods involves the use of deep learning techniques, including neural networks [7, 14], e.g. Graph Neural Network (GNN) [7, 21]. Such methods have been shown to perform well, particularly for dimension reduction and identifying nonlinear relationships among data. Random walk based methods, such as DeepWalk, can incorporate deep learning algorithms like SkipGram for graph embedding, where random walks serve to provide neighborhood information as input.

In general, neural network methods typically work by assigning a weight to each node representation, and then combining those terms via a transfer function. This result is then used as the input for an activation function, such as a sigmoid function, i.e. . This process may then repeat for multiple layers. In a neural network method, the parameters to be optimized are the weights, which are updated after each layer, typically by gradient descent. In the GNN [21], the objective function to be minimized is of the form

where is the learning set, the and are the nodes and target outputs, and the goal is to choose the parameter in such a way that closely approximates the target outputs. This optimization occurs through gradient descent, where the gradient is computed with respect to the weights .

### 2.2 Transition Path Theory

In order to identify the transition paths of a given Markov process, we will adopt the framework of Transition Path Theory (TPT). The following notations are mostly from [17]. Consider a state space , for an initial state and final state , a trajectory is said to be reactive if it begins from state and reaches at state before returning to state .

To determine whether a given reaction path is reactive, we will need the following forward and backward committor functions. For each , the forward committor is the probability that a process initially at state will reach state before it reaches state . Similarly, the discrete backward committor denotes the probability that a process arriving at state was more recently in state rather than in state . The forward committors satisfy the following Dirichlet equation

and similar equations are satisfied by backward committors.

The probability current, or flux, of reactive trajectories gives the average rate at which reactive paths transition from state to state. For trajectories from to , the probability current is defined for all such that

Also, for all , . It can be shown that

Since we are primarily interested in the flow from to , and the process can move in either direction between two adjacent nodes on the path, the following effective current accounts for the net average number of reactive paths that jump from to per time unit:

We can also define the total effective current for a node :

In [11], the effective current is generalized to a connected subnetwork of nodes such that

which leads to the following definition of transition states of a time reversible process as subsets of the state space:

For the non-time reversible case, transition states can be identified similarly, replacing the exponential argument with . The above definition gives the flexibility of identifying transition states as subsets of the states space, which will be very useful when dealing with complex dynamics such as entropy effects.

## 3 Identifying Transition States using Network Embedding

We now introduce a method to identify transition states and paths of Markov processes, by random walk based Network Embedding techniques for directed networks. Consider a Markov jump process in a countable state space . Using the infinitesimal generator for the process, we can calculate the stationary probability distribution and the forward and backward committors, based on which probability current and effective currents and can then be computed. Once these quantities have been obtained, the discretized state space can be viewed as a weighted, directed graph . The nodes are the grid points representing states of the system, and each pair of adjacent nodes are connected by an edge with weights given by the effective probability current, such that the direction of the edge will be determined by the sign of the effective current.

Once is constructed, we can apply feature learning techniques for node classification to identify transition states. Random walk trials for similarities between nodes will start from each node in the space, using the edge weights as transition probabilities. The outputs of the random walks can then be used to compute experimental “neighbor probabilities”, i.e., the probability that a random walk of length starting at node will contain node . If the neighbor probability for a node pair is sufficiently high, then the node is claimed as a neighbor of . Note that this is a directed process; can be a neighbor of even if is not a neighbor of .

After finding these conditional probabilities and identifying the neighborhoods of each node, the next step is to maximize an objective function in order to solve for the node embeddings. As in the node2vec and DeepWalk methods for the undirected case, the conditional probabilities are modeled using softmax functions. In order to adapt this technique for the case of directed graphs, we will use a modified version of the objective function:

(3.1) |

Optimizing this function will give the encoder function that maximizes the probability of the neighborhood of each node, with the most relevant nodes having the most influence on the feature representations.

In the original form of the objective function (3.1) for the undirected graph methods described above, the softmax functions depend only on whether nodes are neighbors, and is independent of the probability (frequency) that a particular neighbor will show up in a random walk trial. In other words, all neighbors are treated equally, while in reality some node pairs are more likely to co-occur than others. To address this, the objective function used here incorporates the neighboring probabilities, denoted such that

Assuming , for being a matrix, the objective defined above can then be maximized using gradient ascent to obtain the matrix and therefore the feature embeddings. In practice, the linear assumption on the map from nodes to embeddings appears to be insufficiently flexible, particularly for higher-dimensional problems. To improve upon this, a neural network can be incorporated into the feature learning process, with the sigmoid function where is the representation in the embedding space of the node . The current algorithm uses a relatively small number of layers; raising this number results in node representations that are clustered more closely to their neighbors.

To obtain transition states for a particular process, we will examine the similarities between the node representations of each state and representation of metastable state . Here, by “similarities” we mean the conditional probabilities given by the softmax units discussed previously. By construction, these values will naturally be small for nodes that are further from node , due to hopping distances that are longer than the length of the random walks. So, to find the transition states, we combine similarities of node pairs with shorter hopping distances via matrix propagation, e.g.

(3.2) |

Note that a node for which this value is high is a higher-order neighbor of the reactant node ; therefore, such a node will have a higher probability of lying along a transition path. We then define transition states as nodes with high probability of lying on a transition path, which are not direct neighbors of the reactant or product states.

## 4 Examples and Results

Next, we will illustrate the suggested method on several simple examples. In the examples presented here, we have used random walk trials, each of length steps. Smaller numbers of trials increase the impact of noise in the results. Due to the matrix propagation technique described above, changes to the walk length parameter do not have a noticeable affect.

### 4.1 Diffusion Process with Energy Barrier

First we will consider a two-dimensional diffusion process representing the position of a particle which follows the SDE

where is the potential

(4.1) |

and and are independent Brownian motion processes.

This potential has two metastable states, or local minima, at (-1,0) and (1,0), as well as a saddle point at the origin. Here I will take (-1,0) to be the reactant state , and (1,0) to be the product state . The diffusion process can be approximated using a Markov (birth-death) jump process on a discrete state space. For this example, we will study the diffusion process on the domain by examining the Markov jump process on a grid , with .

In order to examine the behavior of this Markov process, we should first compute the probability currents for the process. The infinitesimal transition matrix can be found by using the jump rates for the birth-death process for each pair of adjacent nodes [12]. We define the following constants for a state :

Also let if , if , and if or respectively, and similarly for and . Then the infinitesimal generator can be expressed in terms of its action on a test function such that

Figure 1 assigns each point in the grid a color based on its similarity to the node of , for epsilon values and in equation (4.1), respectively. To account for the fact that nodes with greater hopping distances from will have lower similarities to this node, we propagate the similarity matrix according to (3.2) in order to assign similarities to more distant nodes. The red area passing from to represents the region that reactive trajectories will pass through with the highest probability for each value of epsilon.

In Figure 2, the node embeddings corresponding to a subset of nodes with higher similarity values to are shown, for the case where , where is the optimal linear encoding. Notice that in this figure, the embeddings for nodes with the highest probability of occurring in a reactive trajectory are clustered together, while nodes with lower probabilities tend to be grouped with other nodes with similar probabilities; more specifically, nodes with similarity values in the range have embeddings clustered near the upper left or lower right of the figure, while nodes with values from will have embeddings which appear in one of the two orange clusters.

In the context of the diffusion process with the potential , the entropy effect refers to changes in the system’s observed dynamics in response to the change of the parameter . That is, as is decreased, shrinks and therefore the negative term of becomes small. As a result, the shape and size of the saddle point located at the origin is altered, and we can expect the y-coordinate to take on a greater variety of values with higher probabilities in this case. In particular, around the origin, smaller values of should result in higher similarities between the nodes above and below the origin, since in this case the process is more likely to move up and down. For larger values of , the process is expected to move straight forward from to , with a smaller probability of moving up or down.

### 4.2 Diffusion Process with Barriers

Next we consider a pure diffusion process () on the domain with two barriers as depicted in Figure 3. This example was also discussed in [16]. Here we take the node at to be the initial state , and as the final state . The particle spends most of its time between the reactant and product states, in the region between the barriers (). The only factor affecting the probability that the particle will reach before returning to is its current distance to , since there are no energetic obstacles to overcome. In order to travel from to , the process must find its way past the obstacles by chance, thereby overcoming an entropic barrier.

Figure 4 shows the node embeddings for a subset of nodes with high similarity values to reactant . Note that the color scheme used in this figure corresponds to the color scheme in Figure 3; node embeddings that are colored red in the scatter plot correspond to red areas in the surface plot. In Figure 4, node embeddings again tend to be grouped by the corresponding values of the similarity function: nodes that are very similar to the initial state (similarity function value ) appear in the red cluster toward the lower right. Other clusters of different colors correspond to groups of nodes with lower probabilities of appearing in a reactive trajectory. The yellow cluster contains embeddings of nodes with similarity function values approximately , while the blue cluster corresponds to nodes with similarity values near .

### 4.3 Three-Dimensional Toggle Switch

Now we will consider a higher-dimension example, a stochastic model of a 3D genetic toggle switch consisting of three genes that each inhibit the others’ expression [11]. We consider the production and degradation of the three gene products, and :

where the parameters are defined:

with , , , , , and . From the stationary probability distribution, it can be seen that this model has three metastable states:

Here we will choose to be the reactant state and to be the product state.

Figure 5 assigns colors to each node in based on their similarity to node sin , propagating the similarity matrix for nodes with larger hopping distance from as described previously. This figure shows two potential transition states for this system. A reactive trajectory from to will be most likely to pass directly from to along the higher-probability trajectory, represented in red. However, such a trajectory may instead reach after travelling through state , as indicated by the fainter region connecting and via .

Below, Figure 6 shows the two-dimensional node embeddings in , with the embeddings of nodes with zero similarity to omitted for clarity. There are several distinct clusters present in Figure 6, each representing a cluster of similar nodes. The nodes with the highest similarities to after propagation of the similarity matrix (represented in red) are shown in the cluster to the lower left of the figure. Hence, the nodes in this cluster represent the nodes expected to appear in the transition path.

### 4.4 Stochastic Virus Model

Now we will consider a higher-dimension example, a 3D stochastic model for virus propagation consisting of three species which are necessary for virus production [23]. The three species involved in this system are the template (tem), viral genome (gen), and structural (struct) proteins of the virus.

The cellular concentrations of these proteins are controlled by six reactions: producing tem from gen, using tem as a catalyst to produce struct and gen, degradation of tem and struct, and propagation of the virus using struct and gen such that

where we adopt . From [23] we know there are two steady states for this system: an unstable trivial solution at , and a stable steady state, which with the given parameter values is . Ignoring the stochastic effects, as a macro scale limit, this system can be modeled by the ODEs:

For a stochastic perspective, we model the system by a Markov process, where we follow the First Reactions method of the Gillespie algorithm[13], which at each time step assumes that the reaction that could take place in the shortest amount of time will occur next.

From the upper left to the lower right, these clusters contain the representations for nodes near approximately , , and . Direct simulations confirm that these are transition states: reactive trajectories will tend to pass through nodes in each of these clusters. Applying the feature learning method to this system produces a two-dimensional feature representation for each three-dimensional node . Here, we take state to be the origin and . Figure 7 plots the two-dimensional feature representations for this system. In this figure, the nodes with higher similarities to the initial state (represented in red) and those with lower such similarities (represented in blue) appear in separate clusters. Figure 8 plots these feature representations with a logarithmic scale on the -axis to better display the clusters.

Five distinct clusters are discernible in Figure8: the red node in the top left is the feature representation for the reactant state, the red node at the bottom right is the representation of the product state, and the three clusters between each represent a transition state that reactive trajectories pass through while traveling from to . From the upper left to the lower right, these three clusters contain the representations for nodes near approximately , , and . Direct simulations confirm that these are transition states: reactive trajectories will tend to pass through nodes in each of these clusters.

### 4.5 E. Coli Sigma-32 Heat Response Circuit

Finally we will address a higher-dimension example, to illustrate the effectiveness of the method on reducing the dimension of a graph. Consider the E. coli -32 stress circuit, a network of regulatory pathways controlling the -32 protein, which is essential in the E. coli response to heat shock [22]. The systems consists of 10 processes:

In this context, FtsH and GroEL are stress response proteins, E is a holoenzyme, and (or J-complex) represents several chaperone proteins that are lumped as a simplification. Then denotes the protein complex formed when E binds to , which catalyzes downstream synthesis reactions. is a product of translation, which we assume to occur at a rate corresponding to . It can also associate and dissociate of the J-complex. In accordance with [22] the other rate constants are taken to be: and .

For the above reaction rates, this system has a metastable states where the concentrations of FtsH, GroEL, , and are approximately , and the concentrations of the other three species are approximately . Another metastable state occurs when the concentrations of FtsH, GroEL,, and are approximately . Here we take the former to be initial state and the latter to be the product state . Following a procedure similar to that in the above virus propagation example, three-dimensional representations are generated for each node. The plot of these node representations is given in Figure 9. The color scale in this figure is determined by the value of the similarity function between each node and the node , with the most similar nodes shown in red. Nodes with lower similarities have been omitted for clarity. Four distinct clusters can be seen in Figure 9.

## 5 Conclusions

In this paper, we presented a method for analyzing metastable chemical reaction systems via feature learning on directed networks using random walk sampling. We have shown how this method may be used to identify transition states of Markov jump processes by interpreting such processes in terms of directed networks and taking advantage of Transition Path Theory. We have illustrated the efficacy of this method through several low-dimensional examples involving various energetic and entropic barriers. As noted above, for more complex, realistic examples, the method can be used for dimensional reduction, enabling the extraction and analysis of high-dimensional graph information. Further work will be necessary to fully understand this approach from a probabilistic standpoint. Another direction for future work is improving efficiency and development of faster algorithms. However, this method still forms the groundwork for a potential new way of analyzing directed networks and jump processes, particularly in the context of chemical kinetics.

### References

- A. Ahmed, N. Shervashidze, S. Narayanamurthy, V. Josifovski and A. Smola, Distributed large-scale natural graph factorization, in WWW 2013 - Proceedings of the 22nd International Conference on World Wide Web, 2013, 37–48.
- R. Allen, P. Warren and P. Wolde, Sampling rare switching events in biochemical networks, Physical review letters, 94 (2005), 018104.
- M. Belkin and P. Niyogi, Laplacian eigenmaps and spectral techniques for embedding and clustering, Advances in Neural Information Processing System, 14.
- E. Wigner, The transition state method, Trans. of the Faraday Society, 34, 29-41, 1938.
- P. Bolhuis, D. Chandler, C. Dellago and P. Geissler, Transition path sampling: Throwing ropes over rough mountain passes in the dark, Ann. Rev. Phys. Chem., 53, 291-318, 2002.
- W. E and E. Vanden-Eijnden, Towards a theory of transition paths, J. Stat. Phys., 123, 503-523, 2006.
- H. Cai, V. Zheng and K. Chang, A comprehensive survey of graph embedding: Problems, techniques and applications, IEEE Transactions on Knowledge and Data Engineering.
- S. Cao, W. Lu and Q. Xu, Grarep, in CIKM ’15: Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, 2015, 891–900.
- M. Chen, Q. Yang and X. Tang, Directed graph embedding, in IJCAI International Joint Conference on Artificial Intelligence, 2007, 2707–2712.
- F. Chung, Laplacians and the cheeger inequality for directed graphs, Annals of Combinatorics, 9 (2005), 1–19.
- J. Du and D. Liu, Transition states of stochastic chemical reaction networks, Comm. Comp. Phys., to appear.
- C. Gardiner, U. Bhat, D. Stoyan, D. Daley, Y. Kutoyants and B. Rao, Handbook of stochastic methods for physics, chemistry and the natural sciences., Biometrics, 42 (1986), 226.
- D. Gillespie, Approximate accelerated stochastic simulation of chemically reacting systems, Journal of Chemical Physics, 115 (2001), 1716–1733.
- P. Goyal and E. Ferrara, Graph embedding techniques, applications, and performance: A survey, Knowledge-Based Systems.
- A. Grover and J. Leskovec, node2vec: Scalable feature learning for networks, in KDD : proceedings. International Conference on Knowledge Discovery & Data Mining, vol. 2016, 2016, 855–864.
- P. Metzner, C. Schutte and E. Vanden-Eijnden, Illustration of transition path theory on a collection of simple examples, The Journal of chemical physics, 125 (2006), 084110.
- P. Metzner, C. Schutte and E. Vanden-Eijnden, Transition path theory for markov jump processes, Mult. Mod. Sim., 7 (2008), 1192–1219.
- T. Mikolov, I. Sutskever, K. Chen, G. Corrado and J. Dean, Distributed representations of words and phrases and their compositionality, Advances in Neural Information Processing Systems, 26.
- B. Perozzi, R. Al-Rfou and S. Skiena, Deepwalk: Online learning of social representations, Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.
- S. Roweis and L. Saul, Nonlinear dimensionality reduction by locally linear embedding, Science (New York, N.Y.), 290 (2001), 2323–6.
- F. Scarselli, M. Gori, A. C. Tsoi, M. Hagenbuchner and G. Monfardini, The graph neural network model, IEEE Transactions on Neural Networks, 20 (2009), 61–80.
- R. Srivastava, M. S. Peterson and W. E. Bentley, Stochastic kinetic analysis of the escherichia coli stress circuit using 32-targeted antisense, Biotechnology and Bioengineering, 75 (2001), 120–129, URL https://onlinelibrary.wiley.com/doi/abs/10.1002/bit.1171.
- R. Srivastava, L. You, J. Summers and J. Yin, Stochastic vs. deterministic modeling of intracellular viral kinetics, Journal of theoretical biology, 218 (2002), 309–21.
- D. Zhou, J. Huang and B. Schoelkopf, Learning from labeled and unlabeled data on a directed graph, Framework.