Conjecture on fundamental-solution activations ensuring representability and convergence
Establish that one-hidden-layer neural networks with fixed first-layer biases and activation functions that are fundamental solutions of second-order differential operators possess good representability and convergence properties under gradient-descent training with the L2 loss.
References
These remarks, which are elaborated in Section~4, allow us to conjecture that activation functions with the property of being fundamental solutions to a second-order differential operator should give rise to networks which possess good representability and convergence properties.
— Mathematical analysis of one-layer neural network with fixed biases, a new activation function and other observations
(2604.07715 - Macià et al., 9 Apr 2026) in Introduction