Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Simple Hash-Based Early Exiting Approach For Language Understanding and Generation (2203.01670v1)

Published 3 Mar 2022 in cs.CL

Abstract: Early exiting allows instances to exit at different layers according to the estimation of difficulty. Previous works usually adopt heuristic metrics such as the entropy of internal outputs to measure instance difficulty, which suffers from generalization and threshold-tuning. In contrast, learning to exit, or learning to predict instance difficulty is a more appealing way. Though some effort has been devoted to employing such "learn-to-exit" modules, it is still unknown whether and how well the instance difficulty can be learned. As a response, we first conduct experiments on the learnability of instance difficulty, which demonstrates that modern neural models perform poorly on predicting instance difficulty. Based on this observation, we propose a simple-yet-effective Hash-based Early Exiting approach (HashEE) that replaces the learn-to-exit modules with hash functions to assign each token to a fixed exiting layer. Different from previous methods, HashEE requires no internal classifiers nor extra parameters, and therefore is more efficient. Experimental results on classification, regression, and generation tasks demonstrate that HashEE can achieve higher performance with fewer FLOPs and inference time compared with previous state-of-the-art early exiting methods.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Tianxiang Sun (35 papers)
  2. Xiangyang Liu (23 papers)
  3. Wei Zhu (290 papers)
  4. Zhichao Geng (7 papers)
  5. Lingling Wu (5 papers)
  6. Yilong He (1 paper)
  7. Yuan Ni (11 papers)
  8. Guotong Xie (31 papers)
  9. Xuanjing Huang (287 papers)
  10. Xipeng Qiu (257 papers)
Citations (39)

Summary

We haven't generated a summary for this paper yet.