2000 character limit reached
AdaptEval: Evaluating Large Language Models on Domain Adaptation for Text Summarization (2407.11591v3)
Published 16 Jul 2024 in cs.CL
Abstract: Despite the advances in the abstractive summarization task using LLMs (LLM), there is a lack of research that asses their abilities to easily adapt to different domains. We evaluate the domain adaptation abilities of a wide range of LLMs on the summarization task across various domains in both fine-tuning and in-context learning settings. We also present AdaptEval, the first domain adaptation evaluation suite. AdaptEval includes a domain benchmark and a set of metrics to facilitate the analysis of domain adaptation. Our results demonstrate that LLMs exhibit comparable performance in the in-context learning setting, regardless of their parameter scale.