Pruning random resistive memory for optimizing analogue AI (2311.07164v1)
Abstract: The rapid advancement of AI has been marked by the LLMs exhibiting human-like intelligence. However, these models also present unprecedented challenges to energy consumption and environmental sustainability. One promising solution is to revisit analogue computing, a technique that predates digital computing and exploits emerging analogue electronic devices, such as resistive memory, which features in-memory computing, high scalability, and nonvolatility. However, analogue computing still faces the same challenges as before: programming nonidealities and expensive programming due to the underlying devices physics. Here, we report a universal solution, software-hardware co-design using structural plasticity-inspired edge pruning to optimize the topology of a randomly weighted analogue resistive memory neural network. Software-wise, the topology of a randomly weighted neural network is optimized by pruning connections rather than precisely tuning resistive memory weights. Hardware-wise, we reveal the physical origin of the programming stochasticity using transmission electron microscopy, which is leveraged for large-scale and low-cost implementation of an overparameterized random neural network containing high-performance sub-networks. We implemented the co-design on a 40nm 256K resistive memory macro, observing 17.3% and 19.9% accuracy improvements in image and audio classification on FashionMNIST and Spoken digits datasets, as well as 9.8% (2%) improvement in PR (ROC) in image segmentation on DRIVE datasets, respectively. This is accompanied by 82.1%, 51.2%, and 99.8% improvement in energy efficiency thanks to analogue in-memory computing. By embracing the intrinsic stochasticity and in-memory computing, this work may solve the biggest obstacle of analogue computing systems and thus unleash their immense potential for next-generation AI hardware.
- Yi Li (482 papers)
- Songqi Wang (8 papers)
- Yaping Zhao (17 papers)
- Shaocong Wang (10 papers)
- Woyu Zhang (6 papers)
- Yangu He (9 papers)
- Ning Lin (25 papers)
- Binbin Cui (2 papers)
- Xi Chen (1036 papers)
- Shiming Zhang (11 papers)
- Hao Jiang (230 papers)
- Peng Lin (33 papers)
- Xumeng Zhang (10 papers)
- Xiaojuan Qi (133 papers)
- Zhongrui Wang (32 papers)
- Xiaoxin Xu (9 papers)
- Dashan Shang (16 papers)
- Qi Liu (485 papers)
- Kwang-Ting Cheng (96 papers)
- Ming Liu (421 papers)