Attribution of LLaMa sequence-position geometry to RoPE versus attention feature convergence
Determine whether the geometric patterns observed across sequence positions in post-add latent states from intermediate blocks of LLaMa-7B arise from Rotary Positional Encodings applied within self-attention heads, from the relative convergence of features within self-attention contributions as the number of previous tokens increases, or from both. This should be established using analysis methods that can distinguish RoPE-specific effects from token-content influences, beyond PCA-based visualizations.
Sponsor
References
Due to the inherent nature of RoPE applying only within self-attention heads, it is not straightforward to fully separate the effects of the RoPE-augmented attention heads on latent states from other factors such as the content of the input tokens themselves. As such, it is not possible to determine via these visualizations alone whether the geometric patterns observed in Figure~\ref{fig:llama_pos_unit} are a result of RoPE, the relative convergence of features within contributions from self-attention heads as the number of previous tokens increases, or both. We leave interpreting the sequence-wise latent state geometric patterns of RoPE models to future research.