Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

CoditT5: Pretraining for Source Code and Natural Language Editing (2208.05446v2)

Published 10 Aug 2022 in cs.SE and cs.LG

Abstract: Pretrained LLMs have been shown to be effective in many software-related generation tasks; however, they are not well-suited for editing tasks as they are not designed to reason about edits. To address this, we propose a novel pretraining objective which explicitly models edits and use it to build CoditT5, a LLM for software-related editing tasks that is pretrained on large amounts of source code and natural language comments. We fine-tune it on various downstream editing tasks, including comment updating, bug fixing, and automated code review. By outperforming standard generation-based models, we demonstrate the generalizability of our approach and its suitability for editing tasks. We also show how a standard generation model and our edit-based model can complement one another through simple reranking strategies, with which we achieve state-of-the-art performance for the three downstream editing tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jiyang Zhang (11 papers)
  2. Sheena Panthaplackel (9 papers)
  3. Pengyu Nie (19 papers)
  4. Junyi Jessy Li (79 papers)
  5. Milos Gligoric (23 papers)
Citations (80)