Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Training language models to summarize narratives improves brain alignment (2212.10898v2)

Published 21 Dec 2022 in cs.CL and q-bio.NC

Abstract: Building systems that achieve a deeper understanding of language is one of the central goals of NLP. Towards this goal, recent works have begun to train LLMs on narrative datasets which require extracting the most critical information by integrating across long contexts. However, it is still an open question whether these models are learning a deeper understanding of the text, or if the models are simply learning a heuristic to complete the task. This work investigates this further by turning to the one language processing system that truly understands complex language: the human brain. We show that training LLMs for deeper narrative understanding results in richer representations that have improved alignment to human brain activity. We further find that the improvements in brain alignment are larger for character names than for other discourse features, which indicates that these models are learning important narrative elements. Taken together, these results suggest that this type of training can indeed lead to deeper language understanding. These findings have consequences both for cognitive neuroscience by revealing some of the significant factors behind brain-NLP alignment, and for NLP by highlighting that understanding of long-range context can be improved beyond LLMing.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Khai Loong Aw (4 papers)
  2. Mariya Toneva (23 papers)
Citations (21)

Summary

We haven't generated a summary for this paper yet.