Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Going Forward-Forward in Distributed Deep Learning (2404.08573v2)

Published 30 Mar 2024 in cs.LG and cs.DC

Abstract: We introduce a new approach in distributed deep learning, utilizing Geoffrey Hinton's Forward-Forward (FF) algorithm to speed up the training of neural networks in distributed computing environments. Unlike traditional methods that rely on forward and backward passes, the FF algorithm employs a dual forward pass strategy, significantly diverging from the conventional backpropagation process. This novel method aligns more closely with the human brain's processing mechanisms, potentially offering a more efficient and biologically plausible approach to neural network training. Our research explores different implementations of the FF algorithm in distributed settings, to explore its capacity for parallelization. While the original FF algorithm focused on its ability to match the performance of the backpropagation algorithm, the parallelism aims to reduce training times and resource consumption, thereby addressing the long training times associated with the training of deep neural networks. Our evaluation shows a 3.75 times speed up on MNIST dataset without compromising accuracy when training a four-layer network with four compute nodes. The integration of the FF algorithm into distributed deep learning represents a significant step forward in the field, potentially revolutionizing the way neural networks are trained in distributed environments.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (11)
  1. Learning representations by back-propagating errors. Nature, 323(6088):533–536, 1986.
  2. Gpipe: Efficient training of giant neural networks using pipeline parallelism. In Advances in neural information processing systems, volume 32, 2019.
  3. Pipedream: generalized pipeline parallelism for dnn training. In Proceedings of the 27th ACM Symposium on Operating Systems Principles, SOSP ’19, page 1–15, New York, NY, USA, 2019. Association for Computing Machinery.
  4. Flower: A friendly federated learning research framework. 2022.
  5. Geoffrey Hinton. The forward-forward algorithm: Some preliminary investigations. arXiv preprint arXiv:2212.13345, 2022.
  6. Demystifying parallel and distributed deep learning: An in-depth concurrency analysis. ACM Computing Surveys, 52(4):Article 65, 2019.
  7. Scalable deep learning on distributed infrastructures: Challenges, techniques, and tools. ACM Computing Surveys, 53(1):Article 3, 2020.
  8. The predictive forward-forward algorithm, 2023.
  9. Gongpei Zhao et al. The cascaded forward algorithm for neural network training, 2023.
  10. Qikang Deng et al. Dff: Distributed forward-forward algorithm for large-scale model in low-performance devices. In 2023 IEEE 6th International Conference on Pattern Recognition and Artificial Intelligence (PRAI). IEEE, 2023.
  11. Sindy Löwe and Baichuan. Forward-forward. https://github.com/loeweX/Forward-Forward, 2023.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets