Do three-phase representation geometry dynamics appear beyond English and standard objectives?
Determine whether autoregressive transformer language models trained on multilingual corpora or with alternative training objectives (beyond standard cross-entropy pretraining) exhibit the same non-monotonic three-phase representation geometry sequence—Gray phase collapse, Maroon phase dimensional expansion coinciding with n-gram memorization, and BlueViolet phase anisotropic consolidation—observed in English-language models.
References
Our findings have several limitations: (i) computational constraints limited our analysis to models up to 12B parameters, though the phases persist across scales from 160M to 12B; (ii) spectral metric computation requires ∼10K samples and scales quadratically with hidden dimension (iii) our theoretical analysis assumes simplified linear feature extractors, leaving the extension to full transformer architectures as future work; (iv) we focused on English-LLMs trained with standard objectives, and whether similar phases emerge in multilingual or alternatively-trained models remains unexplored.