Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Hardware-Software Blueprint for Flexible Deep Learning Specialization (1807.04188v3)

Published 11 Jul 2018 in cs.LG, cs.DC, and stat.ML

Abstract: Specialized Deep Learning (DL) acceleration stacks, designed for a specific set of frameworks, model architectures, operators, and data types, offer the allure of high performance while sacrificing flexibility. Changes in algorithms, models, operators, or numerical systems threaten the viability of specialized hardware accelerators. We propose VTA, a programmable deep learning architecture template designed to be extensible in the face of evolving workloads. VTA achieves this flexibility via a parametrizable architecture, two-level ISA, and a JIT compiler. The two-level ISA is based on (1) a task-ISA that explicitly orchestrates concurrent compute and memory tasks and (2) a microcode-ISA which implements a wide variety of operators with single-cycle tensor-tensor operations. Next, we propose a runtime system equipped with a JIT compiler for flexible code-generation and heterogeneous execution that enables effective use of the VTA architecture. VTA is integrated and open-sourced into Apache TVM, a state-of-the-art deep learning compilation stack that provides flexibility for diverse models and divergent hardware backends. We propose a flow that performs design space exploration to generate a customized hardware architecture and software operator library that can be leveraged by mainstream learning frameworks. We demonstrate our approach by deploying optimized deep learning models used for object classification and style transfer on edge-class FPGAs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Thierry Moreau (11 papers)
  2. Tianqi Chen (77 papers)
  3. Luis Vega (60 papers)
  4. Jared Roesch (8 papers)
  5. Eddie Yan (5 papers)
  6. Lianmin Zheng (34 papers)
  7. Josh Fromm (7 papers)
  8. Ziheng Jiang (23 papers)
  9. Luis Ceze (38 papers)
  10. Carlos Guestrin (58 papers)
  11. Arvind Krishnamurthy (37 papers)
Citations (69)

Summary

We haven't generated a summary for this paper yet.