Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fine-tuned vs. Prompt-tuned Supervised Representations: Which Better Account for Brain Language Representations? (2310.01854v1)

Published 3 Oct 2023 in cs.AI and cs.CL

Abstract: To decipher the algorithm underlying the human brain's language representation, previous work probed brain responses to language input with pre-trained artificial neural network (ANN) models fine-tuned on NLU tasks. However, full fine-tuning generally updates the entire parametric space and distorts pre-trained features, cognitively inconsistent with the brain's robust multi-task learning ability. Prompt-tuning, in contrast, protects pre-trained weights and learns task-specific embeddings to fit a task. Could prompt-tuning generate representations that better account for the brain's language representations than fine-tuning? If so, what kind of NLU task leads a pre-trained model to better decode the information represented in the human brain? We investigate these questions by comparing prompt-tuned and fine-tuned representations in neural decoding, that is predicting the linguistic stimulus from the brain activities evoked by the stimulus. We find that on none of the 10 NLU tasks, full fine-tuning significantly outperforms prompt-tuning in neural decoding, implicating that a more brain-consistent tuning method yields representations that better correlate with brain data. Moreover, we identify that tasks dealing with fine-grained concept meaning yield representations that better decode brain activation patterns than other tasks, especially the syntactic chunking task. This indicates that our brain encodes more fine-grained concept information than shallow syntactic information when representing languages.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (37)
  1. Predicting neural activity patterns associated with sentences using a neurobiologically motivated model of semantic representation. Cerebral Cortex, 27(9):4379–4395, 2016.
  2. Where is the semantic system? a critical review and meta-analysis of 120 functional neuroimaging studies. Cerebral Cortex, 19(12):2767–2796, 2009.
  3. Language models are few-shot learners. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 1877–1901. Curran Associates, Inc., 2020.
  4. The brain’s default network. Annals of the New York Academy of Sciences, 1124(1):1–38, 2008.
  5. Model-based analysis of brain activity reveals the hierarchy of language in 305 subjects. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 3635–3644, Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics.
  6. Knowprompt: Knowledge-aware prompt-tuning with synergistic optimization for relation extraction. In Proceedings of the ACM Web Conference 2022, pages 2778–2788, 2022.
  7. Multi-task connectivity reveals flexible hubs for adaptive task control. Nature neuroscience, 16(9):1348–1355, 2013.
  8. Bert: Pre-training of deep bidirectional transformers for language understanding. Proceedings of the Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), 2018.
  9. Functional specificity for high-level linguistic processing in the human brain. Proceedings of the National Academy of Sciences, 108(39):16428–16433, 2011.
  10. Alona Fyshe. Corpora and Cognition: The Semantic Composition of Adjectives and Nouns in the Human Brain. PhD thesis, Doctoral dissertation, Air Force Research Laboratory, 2015.
  11. Linking artificial and human neural representations of language. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing (EMNLP) and the 9th International Joint Conference on Natural Language Processing, 2019.
  12. How concepts are encoded in the human brain: a modality independent, category-based cortical organization of semantic knowledge. Neuroimage, 135:232–242, 2016.
  13. Natural speech reveals the semantic maps that tile human cerebral cortex. Nature, 532(7600):453, 2016.
  14. Large language models are zero-shot reasoners. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022.
  15. Fine-tuning can distort pretrained features and underperform out-of-distribution. In International Conference on Learning Representations, 2022.
  16. The power of scale for parameter-efficient prompt tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3045–3059, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics.
  17. Prefix-tuning: Optimizing continuous prompts for generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4582–4597, Online, August 2021. Association for Computational Linguistics.
  18. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Comput. Surv., sep 2022. Just Accepted.
  19. P-tuning: Prompt tuning can be comparable to fine-tuning across scales and tasks. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 61–68, Dublin, Ireland, May 2022. Association for Computational Linguistics.
  20. Generating natural language descriptions for semantic representations of human brain activity. Proceedings of the ACL 2016 Student Research Workshop, pages 22–29, 2016.
  21. fmri semantic category decoding using linguistic encoding of word embeddings. In International Conference on Neural Information Processing, pages 3–15. Springer, 2018.
  22. Neural language taskonomy: Which NLP tasks are the most predictive of fMRI brain activity? In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 3220–3237, Seattle, United States, July 2022. Association for Computational Linguistics.
  23. Training language models to follow instructions with human feedback. Proceedings of Advances in neural information processing systems (NeurIPS), 2022.
  24. Continual lifelong learning with neural networks: A review. Neural Networks, 113:54–71, 2019.
  25. Generating text from functional brain images. Frontiers in human neuroscience, 5:72, 2011.
  26. Toward a universal decoder of linguistic meaning from brain activation. Nature communications, 9(1):963, 2018.
  27. Learning how to ask: Querying LMs with mixtures of soft prompts. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5203–5212, Online, June 2021. Association for Computational Linguistics.
  28. Pre-trained models for natural language processing: A survey. Science China Technological Sciences, 63(10):1872–1897, 2020.
  29. Language models are unsupervised multitask learners. OpenAI Blog, 1(8):9, 2019.
  30. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4938–4947, 2020.
  31. The neural architecture of language: Integrative modeling converges on predictive processing. Proceedings of the National Academy of Sciences, 118(45):e2105646118, 2021.
  32. Towards sentence-level brain decoding with distributed representations. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 7047–7054, 2019.
  33. Distill and replay for continual language learning. In Proceedings of the 28th International Conference on Computational Linguistics, pages 3569–3579, Barcelona, Spain (Online), December 2020. International Committee on Computational Linguistics.
  34. Neural encoding and decoding with distributed sentence representations. IEEE Transactions on Neural Networks and Learning Systems, 32(2):589–603, 2020.
  35. Attention is all you need. Proceedings of Advances in neural information processing systems (NIPS), 2017.
  36. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pages 353–355, Brussels, Belgium, November 2018. Association for Computational Linguistics.
  37. An fmri dataset for concept representation with semantic feature annotations. Scientific Data, 9(1):721, 2022.
Citations (7)

Summary

We haven't generated a summary for this paper yet.