Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Grammar-Based Structural CNN Decoder for Code Generation (1811.06837v1)

Published 14 Nov 2018 in cs.LG, cs.SE, and stat.ML

Abstract: Code generation maps a program description to executable source code in a programming language. Existing approaches mainly rely on a recurrent neural network (RNN) as the decoder. However, we find that a program contains significantly more tokens than a natural language sentence, and thus it may be inappropriate for RNN to capture such a long sequence. In this paper, we propose a grammar-based structural convolutional neural network (CNN) for code generation. Our model generates a program by predicting the grammar rules of the programming language; we design several CNN modules, including the tree-based convolution and pre-order convolution, whose information is further aggregated by dedicated attentive pooling layers. Experimental results on the HearthStone benchmark dataset show that our CNN code generator significantly outperforms the previous state-of-the-art method by 5 percentage points; additional experiments on several semantic parsing tasks demonstrate the robustness of our model. We also conduct in-depth ablation test to better understand each component of our model.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zeyu Sun (33 papers)
  2. Qihao Zhu (27 papers)
  3. Lili Mou (79 papers)
  4. Yingfei Xiong (33 papers)
  5. Ge Li (213 papers)
  6. Lu Zhang (373 papers)
Citations (117)

Summary

We haven't generated a summary for this paper yet.