Papers
Topics
Authors
Recent
Search
2000 character limit reached

Unravelling the Mechanisms of Manipulating Numbers in Language Models

Published 30 Oct 2025 in cs.CL, cs.AI, cs.LG, and cs.NE | (2510.26285v1)

Abstract: Recent work has shown that different LLMs converge to similar and accurate input embedding representations for numbers. These findings conflict with the documented propensity of LLMs to produce erroneous outputs when dealing with numeric information. In this work, we aim to explain this conflict by exploring how LLMs manipulate numbers and quantify the lower bounds of accuracy of these mechanisms. We find that despite surfacing errors, different LLMs learn interchangeable representations of numbers that are systematic, highly accurate and universal across their hidden states and the types of input contexts. This allows us to create universal probes for each LLM and to trace information -- including the causes of output errors -- to specific layers. Our results lay a fundamental understanding of how pre-trained LLMs manipulate numbers and outline the potential of more accurate probing techniques in addressed refinements of LLMs' architectures.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.