Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 28 tok/s
Gemini 2.5 Pro 40 tok/s Pro
GPT-5 Medium 16 tok/s Pro
GPT-5 High 13 tok/s Pro
GPT-4o 103 tok/s Pro
Kimi K2 197 tok/s Pro
GPT OSS 120B 471 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

DLGNet: Hyperedge Classification through Directed Line Graphs for Chemical Reactions (2410.06969v1)

Published 9 Oct 2024 in cs.LG and cs.AI

Abstract: Graphs and hypergraphs provide powerful abstractions for modeling interactions among a set of entities of interest and have been attracting a growing interest in the literature thanks to many successful applications in several fields. In particular, they are rapidly expanding in domains such as chemistry and biology, especially in the areas of drug discovery and molecule generation. One of the areas witnessing the fasted growth is the chemical reactions field, where chemical reactions can be naturally encoded as directed hyperedges of a hypergraph. In this paper, we address the chemical reaction classification problem by introducing the notation of a Directed Line Graph (DGL) associated with a given directed hypergraph. On top of it, we build the Directed Line Graph Network (DLGNet), the first spectral-based Graph Neural Network (GNN) expressly designed to operate on a hypergraph via its DLG transformation. The foundation of DLGNet is a novel Hermitian matrix, the Directed Line Graph Laplacian, which compactly encodes the directionality of the interactions taking place within the directed hyperedges of the hypergraph thanks to the DLG representation. The Directed Line Graph Laplacian enjoys many desirable properties, including admitting an eigenvalue decomposition and being positive semidefinite, which make it well-suited for its adoption within a spectral-based GNN. Through extensive experiments on chemical reaction datasets, we show that DGLNet significantly outperforms the existing approaches, achieving on a collection of real-world datasets an average relative-percentage-difference improvement of 33.01%, with a maximum improvement of 37.71%.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper introduces DLGNet, a spectral Graph Neural Network designed for hyperedge classification using a novel Directed Line Graph representation.
  • DLGNet utilizes complex-valued edge weights within its Directed Line Graph to capture and leverage the directional information present in directed hypergraphs.
  • Experimental results demonstrate that DLGNet significantly outperforms existing methods across chemical reaction datasets, highlighting the value of modeling directionality.

The paper introduces Directed Line Graph Network (DLGNet), a spectral-based Graph Neural Network (GNN) designed for hyperedge classification in directed hypergraphs, with a specific application to chemical reaction classification.

The authors define the concept of a Directed Line Graph (DLG) associated with a directed hypergraph H\vec H. In this DLG(%%%%1%%%%), vertices represent the hyperedges of H\vec H, and edges connect vertices if their corresponding hyperedges in H\vec H share at least one vertex. Complex-valued edge weights in DLG(H\vec H) encode the directionality of interactions within H\vec H.

Key contributions include:

  • A formal definition of a directed line graph associated with a directed hypergraph H\vec H, denoted as DLG(H\vec H).
  • The Directed Line Graph Laplacian LN\mathbb{\vec L}_N, a Hermitian matrix capturing both directed and undirected relationships between hyperedges in a directed hypergraph via its DLG. The paper proves that LN\mathbb{\vec L}_N possesses spectral properties such as being positive semidefinite.
  • DLGNet, a spectral-based GNN designed to operate on directed line graphs, convolving hyperedge features.

The paper defines an undirected hypergraph as an ordered pair H=(V,E)H = (V, E), with n:=Vn := |V| and m:=Em := |E|, where VV is the set of vertices and E2V{}E \subseteq 2^{V} \setminus \{\} is the set of hyperedges. The hyperedges' weights are stored in the diagonal matrix WRm×mW \in \mathbb{R}^{m \times m}, where wew_e is the weight of hyperedge eEe \in E. The vertex degree dud_u and hyperedge degree δe\delta_e are defined as du:=eE:uewed_u := \sum_{e \in E: u \in e} |w_e| for uVu \in V, and δe:=e\delta_e := |e| for eEe \in E, stored in diagonal matrices DvRn×nD_v \in \mathbb{R}^{n \times n} and DeRm×mD_e \in \mathbb{R}^{m \times m}. For 2-uniform hypergraphs, the adjacency matrix ARn×nA \in \mathbb{R}^{n \times n} is defined such that Auv=weA_{uv} = w_e for each e={u,v}Ee=\{u,v\} \in E and Auv=0A_{uv} = 0 otherwise. Directed hypergraph H\vec H is defined as a hypergraph where each hyperedge eEe \in E is partitioned in a head set H(e)H(e) and a tail set T(e)T(e).

The relationship between vertices and hyperedges in a undirected hypergraph H{H} is classically represented via an incidence matrix BB of size V×E|V| \times |E|, where

Bve={1amp;if ve 0amp;otherwisevV,eE.{B_{ve} = \begin{cases} 1 &amp; \text{if } v \in e \ 0 &amp; \text{otherwise} \end{cases} \qquad v \in V, e \in E.}From the incidence matrixBB, one can derive the Signless Laplacian MatrixQQas well as its normalized versionQNQ_N:Q:=BWB<sup>Q := B W B<sup>\topandQN</sup>:=DvBWDe<sup>1</sup>B<sup></sup>DvQ_N</sup> := {D_v} B W D_e<sup>{-1}</sup> B<sup>\top</sup> {D_v},

where W,De,DvW, D_e, D_v are the diagonal matrices defined above.

The Laplacian for a general undirected hypergraph is defined as:

Δ:=IQN\Delta := I-Q_N.

Given a Laplacian matrix L\mathcal{L} of a hypergraph HH that admits an eigenvalue decomposition L=UΛU\mathcal{L} = U \Lambda U^*, where UCn×nU \in \mathbb{C}^{n \times n} represents the eigenvectors, UU^* is its conjugate transpose, and ΛRn×n\Lambda \in \mathbb{R}^{n \times n} is the diagonal matrix containing the eigenvalues, the convolution yxy \circledast x between xx and another graph signal yCny \in \mathbb{C}^n is defined in the frequency space as yx=Udiag(Uy)Uxy \circledast x = U \text{diag}(U^* y) U^* x.

The adjacency matrix of L(H)L(H) is defined as:

A(L(H)):=QWDeA(L(H)) := \mathbb{Q} - W D_e,

where Q:=BB\mathbb{Q} := B^\top B is the Signless Laplacian of L(H)L(H). The normalized Signless Laplacian QN\mathbb{Q}_{N} and the normalized Laplacian LN\mathbb{L}_N are defined as:

Q:=WBBW\mathbb{Q} := {W}B^\top B {W}, ${\mathbb{Q}_{N} := {D}_e} {W} {B}^\top {D}_v^{-1}{B} {W} {D}_e}$, and LN:=IQN\mathbb{L}_N := I - \mathbb{Q}_{N}.

The complex-valued incidence matrix B\vec{B} preserves the directionality of H\vec{H}:

Bve:={1amp;if vH(e), iamp;if vT(e), 0amp;otherwise.vV,eE.\vec{B}_{ve} := \begin{cases} 1 &amp; \text{if } v \in H(e), \ -i &amp; \text{if } v \in T(e), \ 0 &amp; \text{otherwise}. \end{cases} \qquad v \in V, e \in E.The adjacency matrix is computed as:A(DLG(H))=WB<sup></sup>BWWDeA(DLG(\vec{H})) = {W}\vec{B}<sup>*</sup> \vec{B}{W} - W D_e.

The normalized Signless Laplacian $\mathbb{\vec{Q}_{N}$ and the normalized Laplacian $\mathbb{\vec{L}_N$ of DLGDLG are:

$\mathbb{\vec{Q}_{N} := {\vec{D}_e} {W} \vec{B}^* \vec{D}_v^{-1} \vec{B} {W} {\vec{D}_e}$ and $\mathbb{\vec{L}_N := I - \mathbb{\vec{Q}_{N}}$.

The scalar form of LN(ij)\mathbb{\vec L}_N(ij) for a pair of hyperedges i,jEi, j \in E is:

$\mathbb{\vec L}<em>N(ij)= \left{ \begin{array}{lr} \displaystyle 1 - \sum</em>{ u \in i}\frac{w_i}{d_u\delta_i}&amp; i = j\ \displaystyle \left(-\hspace{-.4cm}\sum_{\substack{%i,j \in E: \ u \in H(i) \cap H(j) \ \vee u \in T(i) \cap T(j)} \hspace{-.5cm}\frac{w_i} {w_j}{d_u} - i \left(\sum_{\substack{%i, j \in E:\ u \in H(i) \cap T(j)}%\ \wedge u \in T(j)} \hspace{-.3cm}\frac{w_i} {w_j}{d_u} - \sum_{\substack{%i,j \in E:\ u \in T(i) \cap H(j)} %\ \wedge u \in H(j)} \hspace{-.3cm}\frac{w_i} {w_j}{d_u}\right)\right) \frac{1}{\delta_i} \frac{1}{\delta_j} &amp; i \neq j \end{array} \right.$The Euclidean norm induced byL<em>N\mathbb{\vec L}<em>Nof a complex-valued signalx=a+ibC<sup>nx = a + i b \in \mathbb{C}<sup>{n}reads:12</sup></em>uV1d(u)i,jEw(i)(((aiδ(i)ajδ(j))<sup>2</sup>+(biδ(i)bjδ(j))<sup>2)</sup>1<em>uH(i)H(j)uT(i)T(j) +((aiδ(i)bjδ(j))<sup>2</sup>+(ajδ(j)+biδ(i))<sup>2)</sup>1</em>uH(i)T(j) +((aiδ(i)+bjδ(j))<sup>2</sup>+(ajδ(j)biδ(i))<sup>2)</sup>1uT(i)H(j))w(j)\frac{1}{2}</sup> \sum</em>{u \in V} \frac{1}{d(u)} \sum_{i, j \in E} {w(i)} \Bigg( \left(\left(\frac{a_i}{\delta(i)} - \frac{a_j}{\delta(j)} \right)<sup>2</sup> + \left(\frac{b_i}{\delta(i)} - \frac{b_j}{\delta(j)}\right)<sup>2\right)</sup> \mathbf{1}<em>{ u \in H(i) \cap H(j) \vee u \in T(i) \cap T(j)}\ + \left(\left(\frac{a_i}{\delta(i)} - \frac{b_j}{\delta(j)} \right)<sup>2</sup> + \left(\frac{a_j}{\delta(j)} + \frac{b_i}{\delta(i)}\right)<sup>2\right)</sup> \mathbf{1}</em>{u\in H(i) \cap T(j)} \ + \left( \left(\frac{a_i}{\delta(i)} + \frac{b_j}{\delta(j)} \right)<sup>2</sup> + \left(\frac{a_j}{\delta(j)} - \frac{b_i}{\delta(i)}\right)<sup>2\right)</sup> \mathbf{1}_{u\in T(i) \cap H(j)} \Bigg) {w(j)}.

The convolution operator is defined as Y^x=θ0I+θ1LN\hat{Y} x = \theta_0 I + \theta_1 \mathbb{\vec L}_N.

Given XCm×c0X \in \mathbb{C}^{m \times c_0} as a c0c_0-dimensional graph signal, the feature matrix for the vertices of DGL(H)DGL(\vec H) is defined as X=BXX = \vec{B}^* X', where XCn×c0X' \in \mathbb{C}^{n \times c_0} is the feature matrix of the nodes of H\vec H.

The convolution is computed as:

Z(X)=ϕ(IXΘ0+LNXΘ1)Z (X) = \phi\left(IX\Theta_0 + \mathbb{\vec L}_N X\Theta_1\right),

where ϕ\phi is a complex ReLU activation function, and Θ0,Θ1Cc0×c\Theta_0, \Theta_1 \in \mathbb{C}^{c_0 \times c} are learnable parameters.

The paper presents experiments conducted on three real-world chemical reaction datasets: {\tt Dataset-1} (50K reactions from USPTO granted patents), {\tt Dataset-2} (5300 reactions from five different sources), and {\tt Dataset-3} (649 competitive reactions extracted from \cite{von2020thousands}). Node features are based on Morgan Fingerprints (MFs).

The results demonstrate that DLGNet outperforms existing methods, achieving an average relative percentage difference improvement of 33.01\% over the second-best method across three real-world datasets. Specifically, DLGNet achieves the best improvement on {\tt Dataset-3}, with an average RPD improvement of approximately 37.71\% and an average additive improvement of 31.65 percentage points.

An ablation paper demonstrates the importance of directionality, showing that DLGNet consistently outperforms its undirected counterpart.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-Up Questions

We haven't generated follow-up questions for this paper yet.