Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Compositional Processing Emerges in Neural Networks Solving Math Problems (2105.08961v1)

Published 19 May 2021 in cs.LG, cs.AI, and cs.CL

Abstract: A longstanding question in cognitive science concerns the learning mechanisms underlying compositionality in human cognition. Humans can infer the structured relationships (e.g., grammatical rules) implicit in their sensory observations (e.g., auditory speech), and use this knowledge to guide the composition of simpler meanings into complex wholes. Recent progress in artificial neural networks has shown that when large models are trained on enough linguistic data, grammatical structure emerges in their representations. We extend this work to the domain of mathematical reasoning, where it is possible to formulate precise hypotheses about how meanings (e.g., the quantities corresponding to numerals) should be composed according to structured rules (e.g., order of operations). Our work shows that neural networks are not only able to infer something about the structured relationships implicit in their training data, but can also deploy this knowledge to guide the composition of individual meanings into composite wholes.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Jacob Russin (7 papers)
  2. Roland Fernandez (14 papers)
  3. Hamid Palangi (52 papers)
  4. Eric Rosen (20 papers)
  5. Nebojsa Jojic (43 papers)
  6. Paul Smolensky (31 papers)
  7. Jianfeng Gao (344 papers)
Citations (13)

Summary

We haven't generated a summary for this paper yet.