Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Heuristics for Template-based CEGIS of Loop Invariants with Reinforcement Learning (2107.09766v3)

Published 16 Jul 2021 in cs.AI, cs.LG, and cs.PL

Abstract: Loop-invariant synthesis is the basis of program verification. Due to the undecidability of the problem in general, a tool for invariant synthesis necessarily uses heuristics. Despite the common belief that the design of heuristics is vital for the performance of a synthesizer, heuristics are often engineered by their developers based on experience and intuition, sometimes in an \emph{ad-hoc} manner. In this work, we propose an approach to systematically learning heuristics for template-based CounterExample-Guided Inductive Synthesis (CEGIS) with reinforcement learning. As a concrete example, we implement the approach on top of PCSat, which is an invariant synthesizer based on template-based CEGIS. Experiments show that PCSat guided by the heuristics learned by our framework not only outperforms existing state-of-the-art CEGIS-based solvers such as HoICE and the neural solver Code2Inv, but also has slight advantages over non-CEGIS-based solvers such as Eldarica and Spacer in linear Constrained Horn Clause (CHC) solving.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Minchao Wu (3 papers)
  2. Takeshi Tsukada (17 papers)
  3. Hiroshi Unno (13 papers)
  4. Taro Sekiyama (17 papers)
  5. Kohei Suenaga (28 papers)

Summary

We haven't generated a summary for this paper yet.