Theoretical Foundation of the Neural Feature Ansatz (NFA)
Establish a rigorous theoretical foundation for the Neural Feature Ansatz, which posits that after training the first-layer weight Gram matrix W1^T W1 is proportional to a positive power α of the network’s Average Gradient Outer Product (AGOP) with respect to its inputs; formalize the statement and its validity beyond special cases, specifying the precise mathematical conditions under which the proportionality holds.
Sponsor
References
Developing a theoretical foundation for the NFA is however still an open question.
— On the Neural Feature Ansatz for Deep Neural Networks
(2510.15563 - Tansley et al., 17 Oct 2025) in Section 1: Introduction