Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Paraphrase Types Elicit Prompt Engineering Capabilities (2406.19898v3)

Published 28 Jun 2024 in cs.CL

Abstract: Much of the success of modern LLMs depends on finding a suitable prompt to instruct the model. Until now, it has been largely unknown how variations in the linguistic expression of prompts affect these models. This study systematically and empirically evaluates which linguistic features influence models through paraphrase types, i.e., different linguistic changes at particular positions. We measure behavioral changes for five models across 120 tasks and six families of paraphrases (i.e., morphology, syntax, lexicon, lexico-syntax, discourse, and others). We also control for other prompt engineering factors (e.g., prompt length, lexical diversity, and proximity to training data). Our results show a potential for LLMs to improve tasks when their prompts are adapted in specific paraphrase types (e.g., 6.7% median gain in Mixtral 8x7B; 5.5% in LLaMA 3 8B). In particular, changes in morphology and lexicon, i.e., the vocabulary used, showed promise in improving prompts. These findings contribute to developing more robust LLMs capable of handling variability in linguistic expression.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Jan Philip Wahle (31 papers)
  2. Terry Ruas (46 papers)
  3. Yang Xu (277 papers)
  4. Bela Gipp (98 papers)
Citations (2)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets