Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

From Language to Language-ish: How Brain-Like is an LSTM's Representation of Nonsensical Language Stimuli? (2010.07435v1)

Published 14 Oct 2020 in cs.CL and cs.LG

Abstract: The representations generated by many models of language (word embeddings, recurrent neural networks and transformers) correlate to brain activity recorded while people read. However, these decoding results are usually based on the brain's reaction to syntactically and semantically sound language stimuli. In this study, we asked: how does an LSTM (long short term memory) LLM, trained (by and large) on semantically and syntactically intact language, represent a language sample with degraded semantic or syntactic information? Does the LSTM representation still resemble the brain's reaction? We found that, even for some kinds of nonsensical language, there is a statistically significant relationship between the brain's activity and the representations of an LSTM. This indicates that, at least in some instances, LSTMs and the human brain handle nonsensical data similarly.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Maryam Hashemzadeh (3 papers)
  2. Greta Kaufeld (1 paper)
  3. Martha White (89 papers)
  4. Andrea E. Martin (6 papers)
  5. Alona Fyshe (23 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.