Persistence of sparsity and RAM efficacy at 70B+ scale
Determine whether the sparsity of reinforcement-learning-induced task vectors and the resulting effectiveness of Reinforced Agent Merging (RAM) persist for large language models with at least 70 billion parameters. Specifically, ascertain if the sparsity hypothesis observed on 3B–7B parameter models continues to hold and whether RAM maintains its performance advantages when merging multiple RL-trained agents at massive scale.
References
Finally, our evaluation is primarily conducted on 3B and 7B parameter models; verifying whether the sparsity hypothesis and RAM's efficacy persist in massive-scale models (70B+) remains an open question for future research.
— Behavior Knowledge Merge in Reinforced Agentic Models
(2601.13572 - Yuan et al., 20 Jan 2026) in Limitations (Section*), end of main paper