Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
149 tokens/sec
GPT-4o
9 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Evaluating Prompting Strategies for Grammatical Error Correction Based on Language Proficiency (2402.15930v1)

Published 24 Feb 2024 in cs.CL

Abstract: The writing examples of English language learners may be different from those of native speakers. Given that there is a significant differences in second language (L2) learners' error types by their proficiency levels, this paper attempts to reduce overcorrection by examining the interaction between LLM's performance and L2 language proficiency. Our method focuses on zero-shot and few-shot prompting and fine-tuning models for GEC for learners of English as a foreign language based on the different proficiency. We investigate GEC results and find that overcorrection happens primarily in advanced language learners' writing (proficiency C) rather than proficiency A (a beginner level) and proficiency B (an intermediate level). Fine-tuned LLMs, and even few-shot prompting with writing examples of English learners, actually tend to exhibit decreased recall measures. To make our claim concrete, we conduct a comprehensive examination of GEC outcomes and their evaluation results based on language proficiency.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (12)
  1. Language Models are Few-Shot Learners. In H Larochelle, M Ranzato, R Hadsell, M F Balcan, and H Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 1877–1901. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf.
  2. Automatic Annotation and Evaluation of Error Types for Grammatical Error Correction. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 793–805, Vancouver, Canada, 2017. Association for Computational Linguistics. URL http://aclweb.org/anthology/P17-1074.
  3. The BEA-2019 Shared Task on Grammatical Error Correction. In Proceedings of the Fourteenth Workshop on Innovative Use of NLP for Building Educational Applications, pages 52–75, Florence, Italy, 8 2019. Association for Computational Linguistics. doi: 10.18653/v1/W19-4406. URL https://aclanthology.org/W19-4406.
  4. Analyzing the Performance of GPT-3.5 and GPT-4 in Grammatical Error Correction, 2023. URL https://arxiv.org/abs/2303.14342.
  5. Is ChatGPT a Highly Fluent Grammatical Error Correction System? A Comprehensive Evaluation, 2023. URL https://arxiv.org/abs/2304.01746.
  6. Criterial Features in Learner Corpora: Theory and Illustrations. English Profile Journal, 1(1), 2010. URL http://journals.cambridge.org/action/displayAbstract?fromPage=online&aid=7908278&fileId=S2041536210000103.
  7. Exploring Effectiveness of GPT-3 in Grammatical Error Correction: A Study on Performance and Controllability in Prompt-Based Methods. In Proceedings of the 18th Workshop on Innovative Use of NLP for Building Educational Applications (BEA 2023), pages 205–219, Toronto, Canada, 7 2023. Association for Computational Linguistics. URL https://aclanthology.org/2023.bea-1.18.
  8. GECToR – Grammatical Error Correction: Tag, Not Rewrite. In Proceedings of the Fifteenth Workshop on Innovative Use of NLP for Building Educational Applications, pages 163–170, Seattle, WA, USA → Online, 7 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.bea-1.16. URL https://aclanthology.org/2020.bea-1.16.
  9. A Simple Recipe for Multilingual Grammatical Error Correction. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 702–707, Online, 8 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.acl-short.89. URL https://aclanthology.org/2021.acl-short.89.
  10. Prompt-and-Rerank: A Method for Zero-Shot and Few-Shot Arbitrary Textual Style Transfer with Small Language Models. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 2195–2222, Abu Dhabi, United Arab Emirates, 12 2022. Association for Computational Linguistics. URL https://aclanthology.org/2022.emnlp-main.141.
  11. Developing an automated writing placement system for ESL learners. Applied Measurement in Education, 31(3):251–267, 7 2018. ISSN 0895-7347. doi: 10.1080/08957347.2018.1464447. URL https://doi.org/10.1080/08957347.2018.1464447.
  12. Effects of the types of error, proficiency level of the learners and nature of the recasts on the uptake of learners. Educational Research and Reviews, 12(20):1008–1014, 2017. doi: https://doi.org/10.5897/ERR2017.3367.
Citations (1)

Summary

We haven't generated a summary for this paper yet.