Disentangling FFN computation from intermediate activations without harming trainability
Establish whether feed-forward network computation in dense transformer-based language models can be disentangled from intermediate activations—including residual stream features and self-attention outputs—without significantly degrading model trainability.
References
It remains unclear if FFN computation within dense LLMs can be disentangled from any intermediate activations without significantly hurting model trainability.
— MemoryLLM: Plug-n-Play Interpretable Feed-Forward Memory for Transformers
(2602.00398 - Jaiswal et al., 30 Jan 2026) in Appendix, Background Work — Understanding Feed-Forward Networks in Transformers