Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Natural Language to Code Generation in Interactive Data Science Notebooks (2212.09248v1)

Published 19 Dec 2022 in cs.CL and cs.SE

Abstract: Computational notebooks, such as Jupyter notebooks, are interactive computing environments that are ubiquitous among data scientists to perform data wrangling and analytic tasks. To measure the performance of AI pair programmers that automatically synthesize programs for those tasks given natural language (NL) intents from users, we build ARCADE, a benchmark of 1082 code generation problems using the pandas data analysis framework in data science notebooks. ARCADE features multiple rounds of NL-to-code problems from the same notebook. It requires a model to understand rich multi-modal contexts, such as existing notebook cells and their execution states as well as previous turns of interaction. To establish a strong baseline on this challenging task, we develop PaChiNCo, a 62B code LLM (LM) for Python computational notebooks, which significantly outperforms public code LMs. Finally, we explore few-shot prompting strategies to elicit better code with step-by-step decomposition and NL explanation, showing the potential to improve the diversity and explainability of model predictions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Pengcheng Yin (42 papers)
  2. Wen-Ding Li (19 papers)
  3. Kefan Xiao (7 papers)
  4. Abhishek Rao (4 papers)
  5. Yeming Wen (14 papers)
  6. Kensen Shi (15 papers)
  7. Joshua Howland (5 papers)
  8. Paige Bailey (5 papers)
  9. Michele Catasta (9 papers)
  10. Henryk Michalewski (42 papers)
  11. Alex Polozov (5 papers)
  12. Charles Sutton (74 papers)
Citations (46)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com