Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Compilable Neural Code Generation with Compiler Feedback (2203.05132v1)

Published 10 Mar 2022 in cs.CL, cs.AI, and cs.PL

Abstract: Automatically generating compilable programs with (or without) natural language descriptions has always been a touchstone problem for computational linguistics and automated software engineering. Existing deep-learning approaches model code generation as text generation, either constrained by grammar structures in decoder, or driven by pre-trained LLMs on large-scale code corpus (e.g., CodeGPT, PLBART, and CodeT5). However, few of them account for compilability of the generated programs. To improve compilability of the generated programs, this paper proposes COMPCODER, a three-stage pipeline utilizing compiler feedback for compilable code generation, including LLM fine-tuning, compilability reinforcement, and compilability discrimination. Comprehensive experiments on two code generation tasks demonstrate the effectiveness of our proposed approach, improving the success rate of compilation from 44.18 to 89.18 in code completion on average and from 70.3 to 96.2 in text-to-code generation, respectively, when comparing with the state-of-the-art CodeGPT.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Xin Wang (1306 papers)
  2. Yasheng Wang (91 papers)
  3. Yao Wan (70 papers)
  4. Fei Mi (56 papers)
  5. Yitong Li (95 papers)
  6. Pingyi Zhou (9 papers)
  7. Jin Liu (151 papers)
  8. Hao Wu (623 papers)
  9. Xin Jiang (242 papers)
  10. Qun Liu (230 papers)
Citations (56)