Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
158 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A disciplined approach to neural network hyper-parameters: Part 1 -- learning rate, batch size, momentum, and weight decay (1803.09820v2)

Published 26 Mar 2018 in cs.LG, cs.CV, cs.NE, and stat.ML

Abstract: Although deep learning has produced dazzling successes for applications of image, speech, and video processing in the past few years, most trainings are with suboptimal hyper-parameters, requiring unnecessarily long training times. Setting the hyper-parameters remains a black art that requires years of experience to acquire. This report proposes several efficient ways to set the hyper-parameters that significantly reduce training time and improves performance. Specifically, this report shows how to examine the training validation/test loss function for subtle clues of underfitting and overfitting and suggests guidelines for moving toward the optimal balance point. Then it discusses how to increase/decrease the learning rate/momentum to speed up training. Our experiments show that it is crucial to balance every manner of regularization for each dataset and architecture. Weight decay is used as a sample regularizer to show how its optimal value is tightly coupled with the learning rates and momentums. Files to help replicate the results reported here are available.

Citations (975)

Summary

  • The paper introduces a disciplined approach to tuning hyper-parameters by leveraging cyclical learning rates to achieve super-convergence.
  • It demonstrates that adjusting momentum and optimizing batch sizes can reduce training time while preventing overfitting.
  • Extensive experiments on models like ResNet and datasets such as CIFAR-10 validate the method’s ability to enhance efficiency and accuracy.

A Disciplined Approach to Neural Network Hyper-Parameters: Part 1 -- Learning Rate, Batch Size, Momentum, and Weight Decay

Leslie N. Smith's report provides a comprehensive methodology for optimizing the hyper-parameters of neural networks, with a focus on learning rate, batch size, momentum, and weight decay. The paper critiques the existing practices of hyper-parameter tuning, which often involve computationally expensive and time-consuming grid or random searches, and instead proposes more efficient techniques to achieve optimal settings. Below, I summarize and discuss the key insights, methodologies, and implications of Smith’s findings.

Key Insights and Methodologies

Smith asserts that effective hyper-parameter tuning necessitates a nuanced understanding of the training test/validation loss function, which can reveal subtle clues about underfitting or overfitting. The report advocates for a disciplined approach that takes these clues into account to optimize hyper-parameter settings efficiently.

Learning Rate and Cyclical Learning Rates

One of the cornerstone methodologies presented is the use of Cyclical Learning Rates (CLR) and the Learning Rate (LR) range test. These techniques enable one to determine a good learning rate without exhaustive search. The LR range test involves starting with a small learning rate and gradually increasing it during a pre-training run to identify the highest rate at which the model still trains stably. This upper bound, along with an appropriately set lower bound, defines the cyclical learning rate schedule. Smith also revisits and validates the concept of super-convergence, where larger learning rates can accelerate training and improve regularization, provided the learning rates are cycled appropriately.

Momentum and Cyclical Momentum

Momentum is highlighted as an integral component in conjunction with learning rates. Smith discusses the usefulness of cyclical momentum, wherein the momentum value is varied in tandem with the learning rate to stabilize training at larger learning rates. The recommendation is to start with a high momentum value and gradually decrease it as the learning rate increases.

Batch Size

Smith discusses the trade-offs associated with different batch sizes, emphasizing the need to balance computational efficiency with effective training dynamics. The recommendation is to use larger batch sizes compatible with the hardware's memory limitations to potentially allow larger learning rates.

Weight Decay

Weight decay is a form of regularization whose optimal value depends on the interplay with other hyper-parameters such as learning rate and batch size. Smith proposes that finding the right balance is crucial, and this can be achieved through a grid search or by examining early clues from test losses during training.

Strong Numerical Results and Experimental Validation

Smith substantiates the proposed methodologies through extensive experimental validation using various architectures (e.g., ResNet-56, Wide ResNets, DenseNet, Inception-ResNet) and datasets (e.g., CIFAR-10, CIFAR-100, ImageNet, MNIST). The paper provides concrete numerical results that show improved training times and enhanced model performance when employing the disciplined approach recommended.

For instance:

  • Wide32 networks demonstrated super-convergence on CIFAR-10, achieving a test accuracy of 91.9% in 100 epochs using a 1cycle learning rate policy, compared to 90.3% in 800 epochs with conventional methods.
  • In ResNet-50 and Inception-ResNet-V2 architectures trained on ImageNet, using a super-convergence method led to significant reductions in training time and higher validation accuracies.

Implications and Future Directions

The methodologies described in this report have both practical and theoretical implications. Practically, this structured approach can make neural network training more accessible and less resource-intensive, potentially democratizing advanced AI research and applications. Theoretically, these findings challenge the ad-hoc nature of traditional hyper-parameter tuning and suggest that a more disciplined and data-driven approach can yield better results.

Looking ahead, the insights from this paper open several avenues for future research:

  1. Expansion to Other Hyper-parameters: Investigating how these principles can be extended to other hyper-parameters and regularization techniques such as dropout and data augmentation.
  2. Automated Tuning Systems: Developing automated systems that can implement these disciplined approaches efficiently and adaptively across various architectures and datasets.
  3. Interdependencies in Hyper-parameters: Further exploring the interdependencies among hyper-parameters to develop a more comprehensive, interconnected tuning strategy.

Conclusion

Leslie N. Smith’s report marks a significant step towards a more systematic understanding of hyper-parameter tuning in neural networks. By leveraging cyclical learning rates, cyclical momentum, and focused methodologies for determining optimal batch sizes and weight decay, Smith provides valuable guidelines that improve both training efficiency and model performance. This disciplined approach could form the basis for more standardized and scientifically grounded methods in neural network training.

X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com