Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Embedding Compression with Hashing for Efficient Representation Learning in Large-Scale Graph (2208.05648v1)

Published 11 Aug 2022 in cs.LG, cs.AI, and cs.DB

Abstract: Graph neural networks (GNNs) are deep learning models designed specifically for graph data, and they typically rely on node features as the input to the first layer. When applying such a type of network on the graph without node features, one can extract simple graph-based node features (e.g., number of degrees) or learn the input node representations (i.e., embeddings) when training the network. While the latter approach, which trains node embeddings, more likely leads to better performance, the number of parameters associated with the embeddings grows linearly with the number of nodes. It is therefore impractical to train the input node embeddings together with GNNs within graphics processing unit (GPU) memory in an end-to-end fashion when dealing with industrial-scale graph data. Inspired by the embedding compression methods developed for NLP tasks, we develop a node embedding compression method where each node is compactly represented with a bit vector instead of a floating-point vector. The parameters utilized in the compression method can be trained together with GNNs. We show that the proposed node embedding compression method achieves superior performance compared to the alternatives.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Chin-Chia Michael Yeh (43 papers)
  2. Mengting Gu (4 papers)
  3. Yan Zheng (102 papers)
  4. Huiyuan Chen (43 papers)
  5. Javid Ebrahimi (7 papers)
  6. Zhongfang Zhuang (32 papers)
  7. Junpeng Wang (53 papers)
  8. Liang Wang (512 papers)
  9. Wei Zhang (1489 papers)
Citations (16)