Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Effective and Efficient Embedding via an Adaptively-Masked Twins-based Layer (2108.11513v1)

Published 24 Aug 2021 in cs.LG and cs.AI

Abstract: Embedding learning for categorical features is crucial for the deep learning-based recommendation models (DLRMs). Each feature value is mapped to an embedding vector via an embedding learning process. Conventional methods configure a fixed and uniform embedding size to all feature values from the same feature field. However, such a configuration is not only sub-optimal for embedding learning but also memory costly. Existing methods that attempt to resolve these problems, either rule-based or neural architecture search (NAS)-based, need extensive efforts on the human design or network training. They are also not flexible in embedding size selection or in warm-start-based applications. In this paper, we propose a novel and effective embedding size selection scheme. Specifically, we design an Adaptively-Masked Twins-based Layer (AMTL) behind the standard embedding layer. AMTL generates a mask vector to mask the undesired dimensions for each embedding vector. The mask vector brings flexibility in selecting the dimensions and the proposed layer can be easily added to either untrained or trained DLRMs. Extensive experimental evaluations show that the proposed scheme outperforms competitive baselines on all the benchmark tasks, and is also memory-efficient, saving 60\% memory usage without compromising any performance metrics.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Bencheng Yan (7 papers)
  2. Pengjie Wang (51 papers)
  3. Kai Zhang (542 papers)
  4. Wei Lin (207 papers)
  5. Jian Xu (209 papers)
  6. Bo Zheng (205 papers)
  7. Kuang-chih Lee (23 papers)
Citations (24)

Summary

We haven't generated a summary for this paper yet.