Dice Question Streamline Icon: https://streamlinehq.com

Conjecture: Convergence to a Shared Platonic Representation

Establish whether representation learning algorithms trained on diverse data modalities (e.g., images and text), objectives, and tasks converge to a shared representation of the underlying latent reality variable Z, and ascertain whether increasing model size together with data scale and task diversity causally drives this convergence.

Information Square Streamline Icon: https://streamlinehq.com

Background

The paper proposes the Platonic Representation Hypothesis, arguing that neural networks across architectures, objectives, and modalities increasingly align in how they represent data and that this alignment reflects convergence toward a shared statistical model of reality. In the figure illustrating this hypothesis, images (X) and text (Y) are treated as projections of an underlying world variable Z, and the authors explicitly frame the central claim as a conjecture about convergence.

The authors present empirical evidence for representational alignment within and across modalities and discuss selective pressures toward convergence (task generality, model capacity, and simplicity bias). The conjecture encapsulates the proposed endpoint of these trends: a common, modality-agnostic representation grounded in the statistical structure of the world.

References

We conjecture that representation learning algorithms will converge on a shared representation of $Z$, and scaling model size, as well as data and task diversity, drives this convergence.

The Platonic Representation Hypothesis (2405.07987 - Huh et al., 13 May 2024) in Figure 1 caption, Section 1 (Introduction)