- The paper finds that human word prediction improves gradually with repeated text while GPT-2 achieves near-perfect accuracy after a single repeat.
- Detailed analysis identifies specific GPT-2 attention heads that drive its rapid pattern recognition, highlighting a contrast with human short-term memory limitations.
- Introducing a human-like recency bias in GPT-2 leads to more aligned predictions with human behavior, though it reduces overall accuracy in non-repeating contexts.
In recently published research, a comparison between human cognitive behavior and the performance of LLMs (LMs), specifically the GPT-2 model, was conducted in the field of word prediction. The paper centered on a compelling question: Do humans and artificial intelligence diverge when predicting repeating text?
The paper first set up an experiment where human participants were asked to predict the next word in sequences that were repeated up to four times. Predictably, human performance on this task improved slightly with each repeat, as familiarity with the text helped refine their predictions.
In stark contrast, the paper revealed that the GPT-2 model excelled after just one repetition, achieving nearly perfect performance from therein. This sharp deviation pointed to a fundamental difference in memory mechanisms—the humans relying on relatively fallible short-term memory, and GPT-2 leveraging its capacity to recognize and recall repeated sequences with almost flawless precision.
Upon further analysis, the researchers identified specific attention heads within GPT-2's neural network architecture that facilitated this pattern recognition. These findings throw a spot of doubt on previously held beliefs that LMs mimic human cognitive functions closely.
Seeking to bridge this gap, the researchers introduced a novel method within the model that skewed the attention heads to favor recent information over older data, simulating a form of recency bias akin to human memory patterns. Surprisingly, with this adjustment, the model demonstrated behavior more closely resembling that of the human participants, suggesting that such modifications could make LMs better proxies for human cognition.
However, this human-like performance came at a cost: the LM's overall word-prediction accuracy decreased when applied to non-repeating text. This trade-off revealed an intriguing insight: the LM's exceptional prediction capabilities might be rooted more in its superior memory recall than in its mimicry of human thought processes.
In conclusion, the paper not only shed light on the distinctive memory operations in humans and LMs but also proposed potential steps forward. The work implies that, by refining LMs to exhibit behavior similar to human memory patterns, we may advance closer to creating AI that genuinely reflects human cognitive processes. The findings also hint at optimization opportunities in LM design, perhaps leading to more efficient and effective artificial intelligence in the future.