Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Grammatical Error Generation Based on Translated Fragments (2104.09933v1)

Published 20 Apr 2021 in cs.CL

Abstract: We perform neural machine translation of sentence fragments in order to create large amounts of training data for English grammatical error correction. Our method aims at simulating mistakes made by second language learners, and produces a wider range of non-native style language in comparison to state-of-the-art synthetic data creation methods. In addition to purely grammatical errors, our approach generates other types of errors, such as lexical errors. We perform grammatical error correction experiments using neural sequence-to-sequence models, and carry out quantitative and qualitative evaluation. A model trained on data created using our proposed method is shown to outperform a baseline model on test data with a high proportion of errors.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Eetu Sjöblom (2 papers)
  2. Mathias Creutz (8 papers)
  3. Teemu Vahtola (5 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.