Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Full Stack Optimization of Transformer Inference: a Survey (2302.14017v1)

Published 27 Feb 2023 in cs.CL and cs.LG

Abstract: Recent advances in state-of-the-art DNN architecture design have been moving toward Transformer models. These models achieve superior accuracy across a wide range of applications. This trend has been consistent over the past several years since Transformer models were originally introduced. However, the amount of compute and bandwidth required for inference of recent Transformer models is growing at a significant rate, and this has made their deployment in latency-sensitive applications challenging. As such, there has been an increased focus on making Transformer models more efficient, with methods that range from changing the architecture design, all the way to developing dedicated domain-specific accelerators. In this work, we survey different approaches for efficient Transformer inference, including: (i) analysis and profiling of the bottlenecks in existing Transformer architectures and their similarities and differences with previous convolutional models; (ii) implications of Transformer architecture on hardware, including the impact of non-linear operations such as Layer Normalization, Softmax, and GELU, as well as linear operations, on hardware design; (iii) approaches for optimizing a fixed Transformer architecture; (iv) challenges in finding the right mapping and scheduling of operations for Transformer models; and (v) approaches for optimizing Transformer models by adapting the architecture using neural architecture search. Finally, we perform a case study by applying the surveyed optimizations on Gemmini, the open-source, full-stack DNN accelerator generator, and we show how each of these approaches can yield improvements, compared to previous benchmark results on Gemmini. Among other things, we find that a full-stack co-design approach with the aforementioned methods can result in up to 88.7x speedup with a minimal performance degradation for Transformer inference.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Sehoon Kim (30 papers)
  2. Coleman Hooper (16 papers)
  3. Thanakul Wattanawong (2 papers)
  4. Minwoo Kang (11 papers)
  5. Ruohan Yan (2 papers)
  6. Hasan Genc (9 papers)
  7. Grace Dinh (7 papers)
  8. Qijing Huang (14 papers)
  9. Kurt Keutzer (199 papers)
  10. Michael W. Mahoney (233 papers)
  11. Yakun Sophia Shao (13 papers)
  12. Amir Gholami (60 papers)
Citations (86)
X Twitter Logo Streamline Icon: https://streamlinehq.com