Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

AdaSplit: Adaptive Trade-offs for Resource-constrained Distributed Deep Learning (2112.01637v1)

Published 2 Dec 2021 in cs.LG

Abstract: Distributed deep learning frameworks like federated learning (FL) and its variants are enabling personalized experiences across a wide range of web clients and mobile/IoT devices. However, FL-based frameworks are constrained by computational resources at clients due to the exploding growth of model parameters (eg. billion parameter model). Split learning (SL), a recent framework, reduces client compute load by splitting the model training between client and server. This flexibility is extremely useful for low-compute setups but is often achieved at cost of increase in bandwidth consumption and may result in sub-optimal convergence, especially when client data is heterogeneous. In this work, we introduce AdaSplit which enables efficiently scaling SL to low resource scenarios by reducing bandwidth consumption and improving performance across heterogeneous clients. To capture and benchmark this multi-dimensional nature of distributed deep learning, we also introduce C3-Score, a metric to evaluate performance under resource budgets. We validate the effectiveness of AdaSplit under limited resources through extensive experimental comparison with strong federated and split learning baselines. We also present a sensitivity analysis of key design choices in AdaSplit which validates the ability of AdaSplit to provide adaptive trade-offs across variable resource budgets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Ayush Chopra (24 papers)
  2. Surya Kant Sahu (6 papers)
  3. Abhishek Singh (71 papers)
  4. Abhinav Java (11 papers)
  5. Praneeth Vepakomma (49 papers)
  6. Vivek Sharma (54 papers)
  7. Ramesh Raskar (123 papers)
Citations (20)

Summary

We haven't generated a summary for this paper yet.