Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Pruning random resistive memory for optimizing analogue AI (2311.07164v1)

Published 13 Nov 2023 in cs.ET, cs.AI, and cs.AR

Abstract: The rapid advancement of AI has been marked by the LLMs exhibiting human-like intelligence. However, these models also present unprecedented challenges to energy consumption and environmental sustainability. One promising solution is to revisit analogue computing, a technique that predates digital computing and exploits emerging analogue electronic devices, such as resistive memory, which features in-memory computing, high scalability, and nonvolatility. However, analogue computing still faces the same challenges as before: programming nonidealities and expensive programming due to the underlying devices physics. Here, we report a universal solution, software-hardware co-design using structural plasticity-inspired edge pruning to optimize the topology of a randomly weighted analogue resistive memory neural network. Software-wise, the topology of a randomly weighted neural network is optimized by pruning connections rather than precisely tuning resistive memory weights. Hardware-wise, we reveal the physical origin of the programming stochasticity using transmission electron microscopy, which is leveraged for large-scale and low-cost implementation of an overparameterized random neural network containing high-performance sub-networks. We implemented the co-design on a 40nm 256K resistive memory macro, observing 17.3% and 19.9% accuracy improvements in image and audio classification on FashionMNIST and Spoken digits datasets, as well as 9.8% (2%) improvement in PR (ROC) in image segmentation on DRIVE datasets, respectively. This is accompanied by 82.1%, 51.2%, and 99.8% improvement in energy efficiency thanks to analogue in-memory computing. By embracing the intrinsic stochasticity and in-memory computing, this work may solve the biggest obstacle of analogue computing systems and thus unleash their immense potential for next-generation AI hardware.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (20)
  1. Yi Li (482 papers)
  2. Songqi Wang (8 papers)
  3. Yaping Zhao (17 papers)
  4. Shaocong Wang (10 papers)
  5. Woyu Zhang (6 papers)
  6. Yangu He (9 papers)
  7. Ning Lin (25 papers)
  8. Binbin Cui (2 papers)
  9. Xi Chen (1036 papers)
  10. Shiming Zhang (11 papers)
  11. Hao Jiang (230 papers)
  12. Peng Lin (33 papers)
  13. Xumeng Zhang (10 papers)
  14. Xiaojuan Qi (133 papers)
  15. Zhongrui Wang (32 papers)
  16. Xiaoxin Xu (9 papers)
  17. Dashan Shang (16 papers)
  18. Qi Liu (485 papers)
  19. Kwang-Ting Cheng (96 papers)
  20. Ming Liu (421 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.