Role of data coverage and model capacity in spatial map emergence

Establish whether and how training data coverage across the token vocabulary and model capacity, specifically the embedding dimension, determine the emergence of a spatial map in transformer token embeddings for tokenized coordinate prediction, and identify the conditions under which such a spatial map reliably emerges.

Background

The paper investigates why transformers trained on tokenized planetary motion fail to learn mechanistic world models, identifying the lack of an accurate spatial map in token embeddings as a key obstacle. The authors analyze how the emergence of a spatial map—in the sense of linear directions in the embedding space that correspond to true spatial coordinates—depends on tokenization choices and training setup.

In Section 2.3, they posit that both adequate coverage of the token vocabulary by the training data (data size relative to vocabulary size) and the model’s embedding capacity are crucial determinants of whether such a spatial map emerges. They then study this dependence using a simplified 1D sine-wave dataset and report empirical scaling trends, motivating a formal determination of the precise roles and conditions involved.

References

We conjecture that both data coverage and model capacity play crucial roles: (1) Data coverage: the training data must adequately cover all tokens in the vocabulary, motivating us to vary both the training size D and the vocabulary size V; (2) Model complexity: we vary the embedding dimension N while keeping other hyperparameters fixed.

From Kepler to Newton: Inductive Biases Guide Learned World Models in Transformers  (2602.06923 - Liu et al., 6 Feb 2026) in Section 2.3