Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LazyTensor: combining eager execution with domain-specific compilers (2102.13267v1)

Published 26 Feb 2021 in cs.PL and cs.LG

Abstract: Domain-specific optimizing compilers have demonstrated significant performance and portability benefits, but require programs to be represented in their specialized IRs. Existing frontends to these compilers suffer from the "language subset problem" where some host language features are unsupported in the subset of the user's program that interacts with the domain-specific compiler. By contrast, define-by-run ML frameworks-colloquially called "eager" mode-are popular due to their ease of use and expressivity, where the full power of the host programming language can be used. LazyTensor is a technique to target domain specific compilers without sacrificing define-by-run ergonomics. Initially developed to support PyTorch on Cloud TPUs, the technique, along with a substantially shared implementation, has been used by Swift for TensorFlow across CPUs, GPUs, and TPUs, demonstrating the generality of the approach across (1) Tensor implementations, (2) hardware accelerators, and (3) programming languages.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Alex Suhan (2 papers)
  2. Davide Libenzi (1 paper)
  3. Ailing Zhang (5 papers)
  4. Parker Schuh (6 papers)
  5. Brennan Saeta (11 papers)
  6. Jie Young Sohn (1 paper)
  7. Denys Shabalin (3 papers)
Citations (14)

Summary

We haven't generated a summary for this paper yet.