Peer-to-Peer Learning + Consensus with Non-IID Data (2312.13602v1)
Abstract: Peer-to-peer deep learning algorithms are enabling distributed edge devices to collaboratively train deep neural networks without exchanging raw training data or relying on a central server. Peer-to-Peer Learning (P2PL) and other algorithms based on Distributed Local-Update Stochastic/mini-batch Gradient Descent (local DSGD) rely on interleaving epochs of training with distributed consensus steps. This process leads to model parameter drift/divergence amongst participating devices in both IID and non-IID settings. We observe that model drift results in significant oscillations in test performance evaluated after local training and consensus phases. We then identify factors that amplify performance oscillations and demonstrate that our novel approach, P2PL with Affinity, dampens test performance oscillations in non-IID settings without incurring any additional communication cost.
- “XSEDE: Accelerating Scientific Discovery,” Computing in Science & Engineering, vol. 16, no. 05, pp. 62–74, Sep 2014.
- “Bridges-2: A Platform for Rapidly-Evolving and Data Intensive Research,” in Practice and Experience in Advanced Research Computing, New York, NY, USA, 2021, PEARC ’21, Association for Computing Machinery.
- “A Survey of Collaborative Machine Learning Using 5G Vehicular Communications,” IEEE Communications Surveys & Tutorials, vol. 24, no. 2, pp. 1280–1303, 2022.
- “Camaroptera: A Long-Range Image Sensor with Local Inference for Remote Sensing Applications,” ACM Trans. Embed. Comput. Syst., vol. 21, no. 3, May 2022.
- “Enable Deep Learning on Mobile Devices: Methods, Systems, and Applications,” ACM Trans. Des. Autom. Electron. Syst., vol. 27, no. 3, Mar 2022.
- “Peer-to-Peer Deep Learning for Beyond-5G IoT,” arXiv preprint arXiv:2310.18861, 2023.
- “Federated Learning With Cooperating Devices: A Consensus Approach for Massive IoT Networks,” IEEE Internet of Things Journal, vol. 7, no. 5, pp. 4641–4654, 2020.
- “Cooperative SGD: A Unified Framework for the Design and Analysis of Local-Update SGD Algorithms,” The Journal of Machine Learning Research, vol. 22, no. 1, pp. 9709–9758, 2021.
- “Communication-Efficient Learning of Deep Networks from Decentralized Data,” in Artificial Intelligence and Statistics. PMLR, 2017, pp. 1273–1282.
- “A Unified Theory of Decentralized SGD with Changing Topology and Local Updates,” in International Conference on Machine Learning. PMLR, 2020, pp. 5381–5393.
- “Collaborative Deep Learning in Fixed Topology Networks,” Advances in Neural Information Processing Systems, vol. 30, 2017.
- “Can Decentralized Algorithms Outperform Centralized Algorithms? A Case Study for Decentralized Parallel Stochastic Gradient Descent,” Advances in neural information processing systems, vol. 30, 2017.
- “Decentralized Stochastic Optimization and Machine Learning: A Unified Variance-Reduction Framework for Robust Performance and Fast Convergence,” IEEE Signal Processing Magazine, vol. 37, no. 3, pp. 102–113, 2020.
- “Federated Learning with Non-IID Data,” arXiv preprint arXiv:1806.00582, 2018.
- “Scaffold: Stochastic Controlled Averaging for Federated Learning,” in International Conference on Machine Learning. PMLR, 2020, pp. 5132–5143.
- “MNIST Handwritten Digit Database,” ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist, vol. 2, 2010.
- “PyTorch: An Imperative Style, High-Performance Deep Learning Library,” in Advances in Neural Information Processing Systems 32, pp. 8024–8035. 2019.