Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 62 tok/s
Gemini 2.5 Pro 47 tok/s Pro
GPT-5 Medium 12 tok/s Pro
GPT-5 High 10 tok/s Pro
GPT-4o 91 tok/s Pro
Kimi K2 139 tok/s Pro
GPT OSS 120B 433 tok/s Pro
Claude Sonnet 4 31 tok/s Pro
2000 character limit reached

Generative causal testing to bridge data-driven models and scientific theories in language neuroscience (2410.00812v2)

Published 1 Oct 2024 in cs.CL and q-bio.NC

Abstract: Representations from LLMs are highly effective at predicting BOLD fMRI responses to language stimuli. However, these representations are largely opaque: it is unclear what features of the language stimulus drive the response in each brain area. We present generative causal testing (GCT), a framework for generating concise explanations of language selectivity in the brain from predictive models and then testing those explanations in follow-up experiments using LLM-generated stimuli.This approach is successful at explaining selectivity both in individual voxels and cortical regions of interest (ROIs), including newly identified microROIs in prefrontal cortex. We show that explanatory accuracy is closely related to the predictive power and stability of the underlying predictive models. Finally, we show that GCT can dissect fine-grained differences between brain areas with similar functional selectivity. These results demonstrate that LLMs can be used to bridge the widening gap between data-driven models and formal scientific theories.

Citations (1)

Summary

  • The paper introduces GEM-V, a framework that translates deep language model features into human-readable explanations validated through fMRI experiments.
  • It leverages encoding models from large LLMs like LLaMA and OPT to generate n-gram driven explanations mapping to voxel-level brain responses.
  • Validation reveals significant BOLD responses and specific ROI activations, establishing causal links between data-driven models and neuroscience theories.

A Generative Framework to Bridge Data-Driven Models and Scientific Theories in Language Neuroscience

Overview

The paper "A generative framework to bridge data-driven models and scientific theories in language neuroscience" describes the GEM-V (Generative Explanation-Mediated Validation) framework, a novel methodological approach designed to translate deep learning models of language selectivity in the brain into concise, human-readable explanations. This approach subsequently validates these explanations through follow-up neuroimaging experiments using synthetic stimuli.

The GEM-V Framework

GEM-V is an iterative framework that operates in two main phases:

  1. Explanation Generation:
    • The framework starts by using encoding models built from representations extracted from LLMs. Specifically, the paper utilizes the 30-billion parameter LLaMA and OPT models.
    • These models predict the brain's voxelwise responses using features derived from textual stimuli.
    • GEM-V then automates the generation of natural-language explanations that describe the function computed by each voxel's encoding model. This is achieved by summarizing the n-grams that most strongly drive these voxels.
  2. Validation via Synthetic Stimuli:
    • The generated explanations are tested through new neuroimaging experiments.
    • Synthetic stories containing these explanations are created using an instructive-finetuned LLM.
    • These stories are designed to elicit responses corresponding to specific voxels or regions of interest (ROIs) when presented to subjects during a subsequent fMRI scan.
    • Comparisons are then made between the predicted activations and the actual brain responses to validate the accuracy of these explanations.

Results

Single Voxel and ROI Validation

  • The framework successfully generated explanations for individual voxels and several well-known ROIs, revealing both expected and novel patterns of selectivity in the brain.
  • For single voxels, the generated explanations provided a significant increase in BOLD response when subjects were exposed to the corresponding synthetic stimuli, validating the causal relationship of the explanations.
  • In the ROI validation phase, explanations such as "Body parts" for the extrastriate body area (EBA) and "Scenes and settings" for the parahippocampal place area (PPA) were generated and shown to drive responses significantly above baseline levels.

Granularity and Fine-Tuning

  • A thorough analysis revealed that voxels sharing semantically related explanations often exhibited similar activation patterns when driven by synthetic stimuli.
  • For further nuance, GEM-V could differentiate between ROIs with overlapping semantic selectivity by generating more specific explanations, as demonstrated with location-selective ROIs (RSC, PPA, and OPA).

Factors Influencing Effective Explanations

  • Stability: Stability scores, measuring the agreement between predictions of different encoding models, were found to be a strong predictor of explanatory accuracy.
  • Temporal Dynamics: The presence of key driving n-grams in stimuli showed a typical hemodynamic response peaking around six seconds, further validating the temporal specificity of the responses.

Implications

The GEM-V framework bridges the gap between opaque data-driven models and formal scientific theories in neuroscience. By providing a systematic approach to translate and validate model predictions, it paves the way for more interpretable and causally testable models of brain function. This approach can significantly impact how neuroscientific theories are developed, allowing for rapid hypothesis generation and testing using naturalistic data. Furthermore, it can lead to more robust and nuanced theories of brain function, applicable across different cognitive domains.

Future Directions

The paper suggests several future directions:

  • Enhanced Models: Building more intricate and higher-fidelity models could improve the granularity of explanations.
  • Broader Applications: Extending GEM-V to other modalities such as vision, and to finer-grained data collection methods like 7T laminar fMRI or ECoG.
  • Automated Hypothesis Testing: Developing a fully automated system that continuously refines hypotheses based on ongoing experimental results.

Conclusion

Overall, the GEM-V framework exemplifies a significant step in integrating data-driven modeling with scientific theorization in language neuroscience. It allows for robust, interpretable, and causally valid models of brain function, providing a framework that can be extended and adapted to various cognitive neuroscience domains. By leveraging the predictive power of LLMs and validating these predictions through precisely controlled neuroimaging experiments, GEM-V represents a highly impactful methodological advancement in understanding the neural basis of language.

Lightbulb On Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 4 posts and received 163 likes.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube