Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Scaling laws for language encoding models in fMRI (2305.11863v4)

Published 19 May 2023 in cs.CL and cs.AI

Abstract: Representations from transformer-based unidirectional LLMs are known to be effective at predicting brain responses to natural language. However, most studies comparing LLMs to brains have used GPT-2 or similarly sized LLMs. Here we tested whether larger open-source models such as those from the OPT and LLaMA families are better at predicting brain responses recorded using fMRI. Mirroring scaling results from other contexts, we found that brain prediction performance scales logarithmically with model size from 125M to 30B parameter models, with ~15% increased encoding performance as measured by correlation with a held-out test set across 3 subjects. Similar logarithmic behavior was observed when scaling the size of the fMRI training set. We also characterized scaling for acoustic encoding models that use HuBERT, WavLM, and Whisper, and we found comparable improvements with model size. A noise ceiling analysis of these large, high-performance encoding models showed that performance is nearing the theoretical maximum for brain areas such as the precuneus and higher auditory cortex. These results suggest that increasing scale in both models and data will yield incredibly effective models of language processing in the brain, enabling better scientific understanding as well as applications such as decoding.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Richard Antonello (8 papers)
  2. Aditya Vaidya (1 paper)
  3. Alexander G. Huth (11 papers)
Citations (39)
X Twitter Logo Streamline Icon: https://streamlinehq.com