Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Lightening-Transformer: A Dynamically-operated Optically-interconnected Photonic Transformer Accelerator (2305.19533v3)

Published 31 May 2023 in cs.ET, cs.AR, and physics.optics

Abstract: The wide adoption and significant computing resource of attention-based transformers, e.g., Vision Transformers and LLMs (LLM), have driven the demand for efficient hardware accelerators. There is a growing interest in exploring photonics as an alternative technology to digital electronics due to its high energy efficiency and ultra-fast processing speed. Photonic accelerators have shown promising results for CNNs, which mainly rely on weight-static linear operations. However, they encounter issues when efficiently supporting Transformer architectures, questioning the applicability of photonics to advanced ML tasks. The primary hurdle lies in their inefficiency in handling unique workloads in Transformers, i.e., dynamic and full-range tensor multiplication. In this work, we propose Lightening-Transformer, the first light-empowered, high-performance, and energy-efficient photonic Transformer accelerator. To overcome prior designs' fundamental limitations, we introduce a novel dynamically-operated photonic tensor core, DPTC, a crossbar array of interference-based optical vector dot-product engines supporting highly parallel, dynamic, and full-range matrix multiplication. Furthermore, we design a dedicated accelerator that integrates our novel photonic computing cores with photonic interconnects for inter-core data broadcast, fully unleashing the power of optics. Comprehensive evaluations show that ours achieves >2.6x energy and >12x latency reductions compared to prior photonic accelerators and delivers the lowest energy cost and 2 to 3 orders of magnitude lower energy-delay product compared to electronic Transformer accelerators, all while maintaining digital-comparable accuracy. Our work highlights the immense potential of photonics for advanced ML workloads, such as Transformer-backboned LLM. Our work is available at https://github.com/zhuhanqing/Lightening-Transformer.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Hanqing Zhu (22 papers)
  2. Jiaqi Gu (70 papers)
  3. Hanrui Wang (49 papers)
  4. Zixuan Jiang (16 papers)
  5. Zhekai Zhang (11 papers)
  6. Rongxing Tang (2 papers)
  7. Chenghao Feng (20 papers)
  8. Song Han (155 papers)
  9. Ray T. Chen (43 papers)
  10. David Z. Pan (70 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.