Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Bridging Classical and Quantum Machine Learning: Knowledge Transfer From Classical to Quantum Neural Networks Using Knowledge Distillation (2311.13810v1)

Published 23 Nov 2023 in quant-ph, cs.AI, and cs.LG

Abstract: Very recently, studies have shown that quantum neural networks surpass classical neural networks in tasks like image classification when a similar number of learnable parameters are used. However, the development and optimization of quantum models are currently hindered by issues such as qubit instability and limited qubit availability, leading to error-prone systems with weak performance. In contrast, classical models can exhibit high-performance owing to substantial resource availability. As a result, more studies have been focusing on hybrid classical-quantum integration. A line of research particularly focuses on transfer learning through classical-quantum integration or quantum-quantum approaches. Unlike previous studies, this paper introduces a new method to transfer knowledge from classical to quantum neural networks using knowledge distillation, effectively bridging the gap between classical machine learning and emergent quantum computing techniques. We adapt classical convolutional neural network (CNN) architectures like LeNet and AlexNet to serve as teacher networks, facilitating the training of student quantum models by sending supervisory signals during backpropagation through KL-divergence. The approach yields significant performance improvements for the quantum models by solely depending on classical CNNs, with quantum models achieving an average accuracy improvement of 0.80% on the MNIST dataset and 5.40% on the more complex Fashion MNIST dataset. Applying this technique eliminates the cumbersome training of huge quantum models for transfer learning in resource-constrained settings and enables re-using existing pre-trained classical models to improve performance.Thus, this study paves the way for future research in quantum machine learning (QML) by positioning knowledge distillation as a core technique for advancing QML applications.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (67)
  1. Comparing concepts of quantum and classical neural network models for image classification task. In Progress in Image Processing, Pattern Recognition and Communication Systems: Proceedings of the Conference (CORES, IP&C, ACS)-June 28-30 2021 12, pages 61–71. Springer, 2022.
  2. Qnet: A scalable and noise-resilient quantum neural network architecture for noisy intermediate-scale quantum computers. Frontiers in Physics, 9:702, 2022.
  3. Patrick J Coles. Seeking quantum advantage for neural networks. Nature Computational Science, 1(6):389–390, 2021.
  4. Training deep quantum neural networks. Nature communications, 11(1):808, 2020.
  5. Learnability of quantum neural networks. PRX Quantum, 2(4):040337, 2021.
  6. Challenges and opportunities in quantum machine learning. Nature Computational Science, 2(9):567–576, 2022.
  7. Quantum neural networks: Concepts, applications, and challenges. In 2021 Twelfth International Conference on Ubiquitous and Future Networks (ICUFN), pages 413–416. IEEE, 2021.
  8. A review of quantum neural networks: methods, models, dilemma. arXiv preprint arXiv:2109.01840, 2021.
  9. Knowledge distillation in quantum neural network using approximate synthesis. In Proceedings of the 28th Asia and South Pacific Design Automation Conference, pages 639–644, 2023.
  10. Knowledge distillation for optimization of quantized deep neural networks. In 2020 IEEE Workshop on Signal Processing Systems (SiPS), pages 1–6. IEEE, 2020.
  11. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2015.
  12. Knowledge distillation: A survey. International Journal of Computer Vision, 129:1789–1819, 2021.
  13. Li Deng. The mnist database of handwritten digit images for machine learning research [best of the web]. IEEE signal processing magazine, 29(6):141–142, 2012.
  14. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747, 2017.
  15. Quantum supremacy using a programmable superconducting processor. Nature, 574(7779):505–510, 2019.
  16. Quantum computational advantage with a programmable photonic processor. Nature, 606(7912):75–81, 2022.
  17. Peter W Shor. Scheme for reducing decoherence in quantum computer memory. Physical review A, 52(4):R2493, 1995.
  18. Noisy intermediate-scale quantum algorithms. Reviews of Modern Physics, 94(1):015004, 2022.
  19. Qubits made by advanced semiconductor manufacturing. Nature Electronics, 5(3):184–190, 2022.
  20. Coherent phonon manipulation in coupled mechanical resonators. Nature Physics, 9(8):480–484, 2013.
  21. Quantum boltzmann machine. Physical Review X, 8(2):021050, 2018.
  22. Quantum machine learning. Nature, 549(7671):195–202, 2017.
  23. The power of quantum neural networks. Nature Computational Science, 1(6):403–409, 2021.
  24. Supervised learning with quantum-enhanced feature spaces. Nature, 567(7747):209–212, 2019.
  25. Experimental quantum speed-up in reinforcement learning agents. Nature, 591(7849):229–233, 2021.
  26. Variational quantum algorithms. Nature Reviews Physics, 3(9):625–644, 2021.
  27. Prospects and challenges of quantum finance. arXiv preprint arXiv:2011.06492, 2020.
  28. Fundamental limits of quantum error mitigation. npj Quantum Information, 8(1):114, 2022.
  29. Error statistics and scalability of quantum error mitigation formulas. npj Quantum Information, 9(1):35, 2023.
  30. Reference-state error mitigation: A strategy for high accuracy quantum computation of chemistry. Journal of Chemical Theory and Computation, 19(3):783–789, 2023.
  31. Error mitigation extends the computational reach of a noisy quantum processor. Nature, 567(7749):491–495, 2019.
  32. Learning-based quantum error mitigation. PRX Quantum, 2(4):040330, 2021.
  33. Unifying and benchmarking state-of-the-art quantum error mitigation techniques. Quantum, 7:1034, 2023.
  34. Unified approach to data-driven quantum error mitigation. Physical Review Research, 3(3):033098, 2021.
  35. Error mitigation for quantum approximate optimization. arXiv preprint arXiv:2301.05042, 2023.
  36. Optimizing resource efficiencies for scalable full-stack quantum computers. PRX Quantum, 4(4):040319, 2023.
  37. Next steps in quantum computing: Computer science’s role. arXiv preprint arXiv:1903.10541, 2019.
  38. Artificial intelligence computing at the quantum level. Data, 7(3):28, 2022.
  39. Quantum computing opportunities in renewable energy. SN Computer Science, 2(5):393, 2021.
  40. Transfer learning for scalability of neural-network quantum states. Physical Review E, 101(5):053301, 2020.
  41. Outsmarting quantum chemistry through transfer learning. 2018.
  42. Approaching coupled cluster accuracy with a general-purpose neural network potential through transfer learning. Nature communications, 10(1):2903, 2019.
  43. Transfer learning in hybrid classical-quantum neural networks. Quantum, 4:340, 2020.
  44. An integrated framework for covid-19 classification based on classical and quantum transfer learning from a chest radiograph. Concurrency and Computation: Practice and Experience, 34(20):e6434, 2022.
  45. Using a novel transfer learning method for designing thin film solar cells with enhanced quantum efficiencies. Scientific reports, 9(1):5034, 2019.
  46. A new model for brain tumor detection using ensemble transfer learning and quantum variational classifier. Computational intelligence and neuroscience, 2022, 2022.
  47. Quantum behaved particle swarm optimization-based deep transfer learning model for sugarcane leaf disease detection and classification. Mathematical Problems in Engineering, 2022, 2022.
  48. Enhancing materials property prediction by leveraging computational and experimental data using deep transfer learning. Nature communications, 10(1):5316, 2019.
  49. Classical-to-quantum transfer learning for spoken command recognition based on quantum neural networks. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 8627–8631. IEEE, 2022.
  50. Classical-to-quantum convolutional neural network transfer learning. Neurocomputing, 555:126643, 2023.
  51. Generalized cross entropy loss for training deep neural networks with noisy labels. Advances in neural information processing systems, 31, 2018.
  52. On information and sufficiency. The annals of mathematical statistics, 22(1):79–86, 1951.
  53. John Von Neumann. Mathematical foundations of quantum mechanics: New edition, volume 53. Princeton university press, 2018.
  54. Differentiable learning of quantum circuit born machines. Physical Review A, 98(6):062324, 2018.
  55. Adagrad stepsizes: Sharp convergence over nonconvex landscapes. The Journal of Machine Learning Research, 21(1):9047–9076, 2020.
  56. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  57. Variants of rmsprop and adagrad with logarithmic regret bounds. In International conference on machine learning, pages 2545–2553. PMLR, 2017.
  58. Adaptive subgradient methods for online learning and stochastic optimization. Journal of machine learning research, 12(7), 2011.
  59. Quantum machine learning in feature hilbert spaces. Physical review letters, 122(4):040504, 2019.
  60. Classification with quantum neural networks on near term processors. arXiv preprint arXiv:1802.06002, 2018.
  61. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.
  62. Alex Krizhevsky. One weird trick for parallelizing convolutional neural networks. arXiv preprint arXiv:1404.5997, 2014.
  63. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
  64. Qiskit: An open-source framework for quantum computing. Accessed on: Mar, 16, 2019.
  65. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  66. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
  67. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2818–2826, 2016.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Mohammad Junayed Hasan (4 papers)
  2. M. R. C. Mahdy (77 papers)

Summary

We haven't generated a summary for this paper yet.