Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
173 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Reinforcement learning with learned gadgets to tackle hard quantum problems on real hardware (2411.00230v2)

Published 31 Oct 2024 in quant-ph, cs.AI, and cs.LG

Abstract: Designing quantum circuits for specific tasks is challenging due to the exponential growth of the state space. We introduce gadget reinforcement learning (GRL), which integrates reinforcement learning with program synthesis to automatically generate and incorporate composite gates (gadgets) into the action space. This enhances the exploration of parameterized quantum circuits (PQCs) for complex tasks like approximating ground states of quantum Hamiltonians, an NP-hard problem. We evaluate GRL using the transverse field Ising model under typical computational budgets (e.g., 2- 3 days of GPU runtime). Our results show improved accuracy, hardware compatibility and scalability. GRL exhibits robust performance as the size and complexity of the problem increases, even with constrained computational resources. By integrating gadget extraction, GRL facilitates the discovery of reusable circuit components tailored for specific hardware, bridging the gap between algorithmic design and practical implementation. This makes GRL a versatile framework for optimizing quantum circuits with applications in hardware-specific optimizations and variational quantum algorithms. The code is available at: https://github.com/Aqasch/Gadget_RL

Definition Search Book Streamline Icon: https://streamlinehq.com
References (54)
  1. Tetris-adapt-vqe: An adaptive algorithm that yields shallower, denser circuit ansätze. Physical Review Research, 6(1):013254, 2024.
  2. The option-critic architecture. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, AAAI’17, pp.  1726–1734. AAAI Press, 2017.
  3. A strategy for quantum algorithm design assisted by machine learning. New Journal of Physics, 16(7):073017, 2014.
  4. Quantum approximate optimization via learning-based adaptive optimization. Communications Physics, 7(1):83, 2024.
  5. Bootstrap learning via modular concept discovery. In Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence, IJCAI ’13, pp.  1302–1309. AAAI Press, 2013. ISBN 9781577356332.
  6. Evolutionary quantum architecture search for parametrized quantum circuits. In Proceedings of the Genetic and Evolutionary Computation Conference Companion, pp.  2190–2195, 2022.
  7. Dreamcoder: growing generalizable, interpretable knowledge with wake–sleep bayesian program learning. Philosophical Transactions of the Royal Society A, 381, 2020. URL https://api.semanticscholar.org/CorpusID:219687434.
  8. Diversity is all you need: Learning skills without a reward function. In International Conference on Learning Representations, 2019. URL https://openreview.net/forum?id=SJx63jRqFm.
  9. Overlap-adapt-vqe: practical quantum chemistry on quantum computers via overlap-guided compact ansätze. Communications Physics, 6(1):192, 2023.
  10. Stochastic neural networks for hierarchical reinforcement learning. In International Conference on Learning Representations, 2017. URL https://openreview.net/forum?id=B1oK8aoxe.
  11. Reinforcement learning for variational quantum circuits design. preprint arXiv:2409.05475, 2024.
  12. Quantum circuit optimization with deep reinforcement learning. arXiv preprint arXiv:2103.07585, 2021.
  13. META LEARNING SHARED HIERARCHIES. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=SyX0IeWAW.
  14. Variational intrinsic control, 2016. URL https://arxiv.org/abs/1611.07507.
  15. An adaptive variational algorithm for exact molecular simulations on a quantum computer. Nature communications, 10(1):3007, 2019.
  16. Grover, L. K. A fast quantum mechanical algorithm for database search. In Proceedings of the twenty-eighth annual ACM symposium on Theory of computing, pp.  212–219, 1996.
  17. Gsqas: graph self-supervised quantum architecture search. Physica A: Statistical Mechanics and its Applications, 630:129286, 2023a.
  18. A gnn-based predictor for quantum architecture search. Quantum Information Processing, 22(2):128, 2023b.
  19. A meta-trained generator for quantum architecture search. EPJ Quantum Technology, 11(1):44, 2024.
  20. IBM Quantum Documentation. transpiler — docs.quantum.ibm.com. https://docs.quantum.ibm.com/api/qiskit/transpiler, 2024. [Accessed 30-10-2024].
  21. Kingma, D. P. Adam: A method for stochastic optimization. preprint arXiv:1412.6980, 2014.
  22. Artificial intelligence and machine learning for quantum technologies. Phys. Rev. A, 107:010101, Jan 2023. doi: 10.1103/PhysRevA.107.010101. URL https://link.aps.org/doi/10.1103/PhysRevA.107.010101.
  23. Ddco: Discovery of deep continuous options for robot learning from demonstrations. preprint arXiv:1710.05421, 2017. URL https://api.semanticscholar.org/CorpusID:11787854.
  24. Kundu, A. Reinforcement learning-assisted quantum architecture search for variational quantum algorithms. preprint arXiv:2402.13754, 2024.
  25. Hamiltonian-oriented homotopy quantum approximate optimization algorithm. Physical Review A, 109(2):022611, 2024a.
  26. Kanqas: Kolmogorov arnold network for quantum architecture search. preprint arXiv:2406.17630, 2024b.
  27. Quantum architecture search via deep reinforcement learning. preprint arXiv:2104.07715, 2021.
  28. Continuous evolution for efficient quantum architecture search. EPJ Quantum Technology, 11(1):54, 2024.
  29. Temporal abstraction in reinforcement learning with the successor representation. J. Mach. Learn. Res., 24(1), March 2024. ISSN 1532-4435.
  30. Implementing grover’s algorithm on the ibm quantum computers. In 2018 IEEE international conference on big data (big data), pp.  2531–2537. IEEE, 2018.
  31. Melo, F. S. Convergence of q-learning: A simple proof. Institute Of Systems and Robotics, Tech. Rep, pp.  1–4, 2001.
  32. Cost explosion for efficient reinforcement learning optimisation of quantum circuits. In 2023 IEEE International Conference on Rebooting Computing (ICRC), pp.  1–5. IEEE, 2023.
  33. Realization of a scalable shor algorithm. Science, 351(6277):1068–1070, 2016.
  34. Data-efficient hierarchical reinforcement learning. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS’18, pp.  3307–3317, Red Hook, NY, USA, 2018. Curran Associates Inc.
  35. Perturbed ferromagnetic chain: Tunable test of hardness in the transverse-field ising model. Phys. Rev. A, 105:022410, Feb 2022. doi: 10.1103/PhysRevA.105.022410. URL https://link.aps.org/doi/10.1103/PhysRevA.105.022410.
  36. Reinforcement learning for optimization of variational quantum circuit architectures. Advances in Neural Information Processing Systems, 34:18182–18194, 2021.
  37. Reinforcement learning assisted recursive qaoa. EPJ Quantum Technology, 11(1):6, 2024a.
  38. Curriculum reinforcement learning for quantum architecture search under hardware errors. preprint arXiv:2402.03500, 2024b.
  39. Pierce, B. C. Types and programming languages. MIT Press, Cambridge, Mass, 2002. ISBN 978-0-262-16209-8. URL 10.5555/509043.
  40. Powell, M. J. A direct search optimization method that models the objective and constraint functions by linear interpolation. Springer, 1994.
  41. Quantum Circuit Optimization with AlphaTensor. preprint arXiv:2402.14396, 2 2024.
  42. A quantum information theoretic analysis of reinforcement learning-assisted quantum architecture search. preprint arXiv:2404.06174, 2024.
  43. Discovering quantum circuit components with program synthesis. Machine Learning: Science and Technology, 5(2):025029, may 2024. doi: 10.1088/2632-2153/ad4252. URL https://dx.doi.org/10.1088/2632-2153/ad4252.
  44. Shor, P. W. Polynomial-time algorithms for prime factorization and discrete logarithms on a quantum computer. SIAM review, 41(2):303–332, 1999.
  45. Hybrid quantum-classical algorithm for the transverse-field Ising model in the thermodynamic limit. preprint arXiv:2310.07600, 10 2023.
  46. Quantum architecture search with unsupervised representation learning. preprint arXiv:2401.11576, 2024.
  47. qubit-adapt-vqe: An adaptive algorithm for constructing hardware-efficient ansätze on a quantum processor. PRX Quantum, 2(2):020310, 2021.
  48. Alpharouter: Quantum circuit routing with reinforcement learning and tree search. preprint arXiv:2410.05115, 2024.
  49. Automated gadget discovery in the quantum domain. Machine Learning: Science and Technology, 4(3):035043, sep 2023. doi: 10.1088/2632-2153/acf098. URL https://dx.doi.org/10.1088/2632-2153/acf098.
  50. Scipy 1.0: fundamental algorithms for scientific computing in python. Nature methods, 17(3):261–272, 2020.
  51. Quantum architecture search via continual reinforcement learning. preprint arXiv:2112.05779, 2021.
  52. Neural predictor based quantum architecture search. Machine Learning: Science and Technology, 2(4):045027, 2021.
  53. Quantum approximate optimization algorithm: Performance, mechanism, and implementation on near-term devices. Physical Review X, 10(2):021067, 2020.
  54. Adaptive quantum approximate optimization algorithm for solving combinatorial problems on a quantum computer. Physical Review Research, 4(3):033029, 2022.

Summary

  • The paper introduces Gadget Reinforcement Learning (GRL), a novel approach leveraging learned composite gates (gadgets) to build efficient parameterized quantum circuits (PQCs).
  • GRL iteratively learns gadgets from solving easy quantum problems and integrates them into the RL agent's action space to effectively tackle more complex tasks like finding the TFIM ground state.
  • Experiments show GRL achieves significantly higher accuracy and more compact circuits compared to pure RL, demonstrating viability for real-world quantum hardware like IBM Heron by reducing transpilation and increasing fidelity.

Tackling Quantum Problems with Gadget Reinforcement Learning

The paper “From Easy to Hard: Tackling Quantum Problems with Learned Gadgets For Real Hardware” by Akash Kundu and Leopoldo Sarra presents an innovative approach to building quantum circuits using reinforcement learning (RL) to enhance the efficiency and practicality of quantum algorithms on current hardware. The focus is on a novel algorithm, Gadget Reinforcement Learning (GRL), which addresses the limitations of finding parameterized quantum circuits (PQCs) for simulating quantum systems, specifically targeting the NP-hard problem of obtaining the ground state of the transverse field Ising model (TFIM).

Methodological Advances

The GRL algorithm improves upon traditional RL algorithms by introducing learned composite gates, called “gadgets,” as additional actions within the RL framework. This extension to the action space is designed to overcome the exponential scaling issue typical in quantum state preparation and optimization tasks. The method begins with solving simple quantum problems and extracting frequently occurring sequences of gates or “gadgets” from these solutions. These gadgets are then iteratively integrated into the action space of the RL agent to tackle more complex instances, effectively creating a dynamically evolving decision-making space for the agent.

The architecture harnesses a double deep Q-network with an epsilon-greedy policy to sequentially build a quantum circuit by adding gates from a defined action space. The reward scheme is tailored to focus the agent’s search by providing feedback through a stringent cost function tied to the estimation of the TFIM ground state’s energy.

Experimental Results and Impact

The paper's results highlight a significant improvement in achieving compact PQCs that approximate the TFIM ground state with increased accuracy and reduced error margins, up to 10710^7 times better than pure RL approaches. The experiments demonstrate that GRL maintains scalability to larger systems and more complex quantum problems, establishing its viability for real-world quantum hardware applications.

Further, the work underscores the advantage of crafting circuits with a native gateset of the IBM Heron processor, enabling more straightforward implementation on quantum hardware. The GRL algorithm produces circuits that require less transpilation, enhancing fidelity and execution efficiency by minimizing unnecessary quantum gate operations. This practical optimization has tangible implications for the deployment of quantum algorithms under constrained, noise-prone quantum environments.

Implications and Future Directions

This research introduces a promising strategy for the automated design of VQAs by leveraging learned knowledge through gadgets. Its applications extend beyond the immediate problem space, potentially influencing methods in adaptive algorithm design where similar computational challenges exist. Moreover, refining GRL can lead to enhanced integration with noise-aware quantum computing models, thus contributing to more resilient quantum applications across various platforms.

The paper opens avenues for further exploration, such as integrating noise models directly into the learning protocol to assess the robustness of derived algorithms. Another promising direction is the exploration of hierarchical RL frameworks where gadgets form the basis for modular, scalable long-term strategies.

In summary, the paper presents a significant methodological contribution to machine learning for quantum circuits, displaying a thorough evaluation of GRL in optimizing the PQC search process. Its strategic innovation in utilizing learned gadgets illustrates the nuanced way to better align the potential of RL with the requirements and limitations of current quantum hardware, marking a step toward pragmatic quantum computing solutions.