Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Few-Shot Testing: Estimating Uncertainty of Memristive Deep Neural Networks Using One Bayesian Test Vector (2405.18894v1)

Published 29 May 2024 in cs.LG, cs.AI, and cs.ET

Abstract: The performance of deep learning algorithms such as neural networks (NNs) has increased tremendously recently, and they can achieve state-of-the-art performance in many domains. However, due to memory and computation resource constraints, implementing NNs on edge devices is a challenging task. Therefore, hardware accelerators such as computation-in-memory (CIM) with memristive devices have been developed to accelerate the most common operations, i.e., matrix-vector multiplication. However, due to inherent device properties, external environmental factors such as temperature, and an immature fabrication process, memristors suffer from various non-idealities, including defects and variations occurring during manufacturing and runtime. Consequently, there is a lack of complete confidence in the predictions made by the model. To improve confidence in NN predictions made by hardware accelerators in the presence of device non-idealities, in this paper, we propose a Bayesian test vector generation framework that can estimate the model uncertainty of NNs implemented on memristor-based CIM hardware. Compared to the conventional point estimate test vector generation method, our method is more generalizable across different model dimensions and requires storing only one test Bayesian vector in the hardware. Our method is evaluated on different model dimensions, tasks, fault rates, and variation noise to show that it can consistently achieve $100\%$ coverage with only $0.024$ MB of memory overhead.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (45)
  1. S. W. Keckler et al., “Gpus and the future of parallel computing,” IEEE micro, vol. 31, no. 5, pp. 7–17, 2011.
  2. Y. Chen et al., “A survey of accelerator architectures for deep neural networks,” Engineering, vol. 6, no. 3, pp. 264–274, 2020.
  3. S. Yu, “Neuro-inspired computing with emerging nonvolatile memorys,” Proceedings of the IEEE, vol. 106, no. 2, 2018.
  4. H.-S. P. Wong et al., “Metal–oxide rram,” Proceedings of the IEEE, vol. 100, no. 6, pp. 1951–1970, 2012.
  5. G. W. Burr et al., “Experimental demonstration and tolerancing of a large-scale neural network (165 000 synapses) using phase-change memory as the synaptic weight element,” IEEE Transactions on Electron Devices, vol. 62, no. 11, pp. 3498–3507, 2015.
  6. A. F. Vincent et al., “Spin-transfer torque magnetic memory as a stochastic memristive synapse for neuromorphic systems,” IEEE transactions on biomedical circuits and systems, vol. 9, no. 2, pp. 166–174, 2015.
  7. K. Hofmann et al., “Comprehensive statistical investigation of stt-mram thermal stability,” in Symposium on VLSI Technology: Digest of Technical Papers, June 2014, pp. 1–2.
  8. C.-Y. Chen et al., “Rram defect modeling and failure analysis based on march test and a novel squeeze-search scheme,” IEEE Transactions on Computers, vol. 64, no. 1, pp. 180–190, 2014.
  9. J. Park et al., “Investigation of state stability of low-resistance state in resistive memory,” IEEE Electron Device Letters, vol. 31, no. 5, pp. 485–487, 2010.
  10. Y.-X. Chen and J.-F. Li, “Fault modeling and testing of 1t1r memristor memories,” in 2015 IEEE 33rd VLSI Test Symposium (VTS), 2015, pp. 1–6.
  11. W. Sun et al., “Understanding memristive switching via in situ characterization and device modeling,” Nature communications, vol. 10, no. 1, p. 3453, 2019.
  12. S. M. Nair et al., “Defect characterization and test generation for spintronic-based compute-in-memory,” in IEEE ETS.   IEEE, 2020.
  13. D. Niu et al., “Impact of process variations on emerging memristor,” in Proceedings of the 47th Design Automation Conference, 2010, pp. 877–882.
  14. C. Zhou et al., “Noisy machines: Understanding noisy neural networks and enhancing robustness to analog hardware errors using distillation,” 2020.
  15. H. Kim et al., “Efficient precise weight tuning protocol considering variation of the synaptic devices and target accuracy,” Neurocomputing, vol. 378, pp. 189–196, 2020.
  16. A. Chaudhuri et al., “Functional test generation for ai accelerators using bayesian optimization,” in 2023 IEEE 41th VLSI Test Symposium (VTS).   IEEE, 2023, pp. 1–7.
  17. C.-Y. Chen and K. Chakrabarty, “On-line functional testing of memristor-mapped deep neural networks using backdoored checksums,” in 2021 IEEE ITC, 2021.
  18. W. Li et al., “Rramedy: Protecting reram-based neural network from permanent and soft faults during its lifetime,” in IEEE ICCD), 2019.
  19. B. Luo et al., “On functional test generation for deep neural network ips,” in 2019 Design, Automation & Test in Europe Conference & Exhibition (DATE).   IEEE, 2019, pp. 1010–1015.
  20. S. T. Ahmed and M. B. Tahoori, “One-shot online testing of deep neural networks based on distribution shift detection,” 2023.
  21. S. T. Ahmed et al., “Binary bayesian neural networks for efficient uncertainty estimation leveraging inherent stochasticity of spintronic devices,” in IEEE/ACM NANOARCH, 2022.
  22. ——, “SpinDrop: Dropout-Based Bayesian Binary Neural Networks with spintronic Implementation,” To appear at IEEE Journal on Emerging and Selected Topics in Circuits and Systems, 2023.
  23. A. Malhotra et al., “Exploiting Oxide Based Resistive RAM Variability for Bayesian Neural Network Hardware Design,” IEEE TNANO, 2020.
  24. S. T. Ahmed et al., “Scalable spintronics-based bayesian neural network for uncertainty estimation,” in 2023 Design, Automation and Test in Europe (DATE).   IEEE, 2023.
  25. K. Yang et al., “All-spin bayesian neural networks,” IEEE Transactions on Electron Devices, vol. 67, no. 3, pp. 1340–1347, 2020.
  26. M. Liu et al., “Online fault detection in reram-based computing systems for inferencing,” IEEE Trans. on VLSI Systems, vol. 30, no. 4, pp. 392–405, 2022.
  27. X. Ding et al., “Repvgg: Making vgg-style convnets great again,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 13 733–13 742.
  28. J. Deng et al., “Imagenet: A large-scale hierarchical image database,” in 2009 IEEE conference on computer vision and pattern recognition.   Ieee, 2009, pp. 248–255.
  29. S. Yoshikiyo et al., “Nn algorithm aware alternate layer retraining on computation-in-memory for write variation compensation of non-volatile memories at edge ai,” in 2023 7th IEEE Electron Devices Technology & Manufacturing Conference (EDTM).   IEEE, 2023, pp. 1–3.
  30. S. T. Ahmed and M. B. Tahoori, “Fault-tolerant neuromorphic computing with memristors using functional atpg for efficient re-calibration,” IEEE Design & Test, 2023.
  31. K. He et al., “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
  32. A. Krizhevsky et al., “Learning multiple layers of features from tiny images,” 2009.
  33. M. Sandler et al., “Mobilenetv2: Inverted residuals and linear bottlenecks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 4510–4520.
  34. C. Szegedy et al., “Rethinking the inception architecture for computer vision,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 2818–2826.
  35. G. Huang et al., “Densely connected convolutional networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 4700–4708.
  36. O. Ronneberger et al., “U-net: Convolutional networks for biomedical image segmentation,” in Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18.   Springer, 2015, pp. 234–241.
  37. M. Buda et al., “Association of genomic subtypes of lower-grade gliomas with shape features automatically extracted by a deep learning algorithm,” Computers in biology and medicine, vol. 109, pp. 218–225, 2019.
  38. J. Long et al., “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3431–3440.
  39. T.-Y. Lin et al., “Microsoft coco: Common objects in context,” in Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13.   Springer, 2014, pp. 740–755.
  40. A. Radford et al., “Unsupervised representation learning with deep convolutional generative adversarial networks,” 2016.
  41. N. Rostamzadeh et al., “Fashion-gen: The generative fashion dataset and challenge,” 2018.
  42. D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” arXiv preprint arXiv:1312.6114, 2013.
  43. Y. Chen, “Pytorch cifar models,” https://github.com/chenyaofo/pytorch-cifar-models, 2013.
  44. S. T. Ahmed and M. B. Tahoori, “Compact functional test generation for memristive deep learning implementations using approximate gradient ranking,” in 2022 IEEE International Test Conference (ITC).   IEEE, 2022, pp. 239–248.
  45. L.-H. Tsai et al., “Robust processing-in-memory neural networks via noise-aware normalization,” 2020.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Soyed Tuhin Ahmed (10 papers)
  2. Mehdi Tahoori (5 papers)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com