Extending CSRv2 to the k=1 Extreme Sparsity Regime
Establish training and modeling techniques that enable Contrastive Sparse Representation v2 (CSRv2) to produce effective embeddings when the activation budget is k=1 (i.e., only a single latent dimension is nonzero), mitigating dead neurons and preventing sharp performance degradation in this extreme setting.
References
A key open challenge is the $k=1$ regime, where CSRv2 still suffers from severe dead neurons and sharp degradation (Appendix~\ref{sec:k_equals_1}). Since $k=1$ effectively reduces to clustering (mapping each input to a one-shot label), future work could explore clustering-inspired approaches, such as prototype or vector quantization, balanced assignment, entropy regularization, or optimal transport.
— CSRv2: Unlocking Ultra-Sparse Embeddings
(2602.05735 - Guo et al., 5 Feb 2026) in Concluding Remarks