2000 character limit reached
Nonlinearity Enhanced Adaptive Activation Functions (2403.19896v2)
Published 29 Mar 2024 in cs.LG, cs.CV, and cs.NE
Abstract: A general procedure for introducing parametric, learned, nonlinearity into activation functions is found to enhance the accuracy of representative neural networks without requiring significant additional computational resources. Examples are given based on the standard rectified linear unit (ReLU) as well as several other frequently employed activation functions. The associated accuracy improvement is quantified both in the context of the MNIST digit data set and a convolutional neural network (CNN) benchmark example.
Collections
Sign up for free to add this paper to one or more collections.