Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Is It Worth the (Environmental) Cost? Limited Evidence for Temporal Adaptation via Continuous Training (2210.07365v2)

Published 13 Oct 2022 in cs.CL

Abstract: Language is constantly changing and evolving, leaving LLMs to become quickly outdated. Consequently, we should continuously update our models with new data to expose them to new events and facts. However, that requires additional computing, which means new carbon emissions. Do any measurable benefits justify this cost? This paper looks for empirical evidence to support continuous training. We reproduce existing benchmarks and extend them to include additional time periods, models, and tasks. Our results show that the downstream task performance of temporally adapted English models for social media data do not improve over time. Pretrained models without temporal adaptation are actually significantly more effective and efficient. However, we also note a lack of suitable temporal benchmarks. Our findings invite a critical reflection on when and how to temporally adapt LLMs, accounting for sustainability.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Giuseppe Attanasio (21 papers)
  2. Debora Nozza (17 papers)
  3. Federico Bianchi (47 papers)
  4. Dirk Hovy (57 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com