Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Avoiding Synchronization in First-Order Methods for Sparse Convex Optimization (1712.06047v1)

Published 17 Dec 2017 in cs.DC, cs.LG, math.OC, and stat.ML

Abstract: Parallel computing has played an important role in speeding up convex optimization methods for big data analytics and large-scale ML. However, the scalability of these optimization methods is inhibited by the cost of communicating and synchronizing processors in a parallel setting. Iterative ML methods are particularly sensitive to communication cost since they often require communication every iteration. In this work, we extend well-known techniques from Communication-Avoiding Krylov subspace methods to first-order, block coordinate descent methods for Support Vector Machines and Proximal Least-Squares problems. Our Synchronization-Avoiding (SA) variants reduce the latency cost by a tunable factor of $s$ at the expense of a factor of $s$ increase in flops and bandwidth costs. We show that the SA-variants are numerically stable and can attain large speedups of up to $5.1\times$ on a Cray XC30 supercomputer.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Aditya Devarakonda (9 papers)
  2. Kimon Fountoulakis (33 papers)
  3. James Demmel (54 papers)
  4. Michael W. Mahoney (233 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.