Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enabling On-Device Training of Speech Recognition Models with Federated Dropout (2110.03634v1)

Published 7 Oct 2021 in cs.LG and cs.DC

Abstract: Federated learning can be used to train machine learning models on the edge on local data that never leave devices, providing privacy by default. This presents a challenge pertaining to the communication and computation costs associated with clients' devices. These costs are strongly correlated with the size of the model being trained, and are significant for state-of-the-art automatic speech recognition models. We propose using federated dropout to reduce the size of client models while training a full-size model server-side. We provide empirical evidence of the effectiveness of federated dropout, and propose a novel approach to vary the dropout rate applied at each layer. Furthermore, we find that federated dropout enables a set of smaller sub-models within the larger model to independently have low word error rates, making it easier to dynamically adjust the size of the model deployed for inference.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Dhruv Guliani (6 papers)
  2. Lillian Zhou (4 papers)
  3. Changwan Ryu (3 papers)
  4. Tien-Ju Yang (16 papers)
  5. Harry Zhang (37 papers)
  6. Yonghui Xiao (15 papers)
  7. Giovanni Motta (11 papers)
  8. Francoise Beaufays (8 papers)
Citations (15)

Summary

We haven't generated a summary for this paper yet.