Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Apollo: Transferable Architecture Exploration (2102.01723v1)

Published 2 Feb 2021 in cs.LG and cs.AR

Abstract: The looming end of Moore's Law and ascending use of deep learning drives the design of custom accelerators that are optimized for specific neural architectures. Architecture exploration for such accelerators forms a challenging constrained optimization problem over a complex, high-dimensional, and structured input space with a costly to evaluate objective function. Existing approaches for accelerator design are sample-inefficient and do not transfer knowledge between related optimizations tasks with different design constraints, such as area and/or latency budget, or neural architecture configurations. In this work, we propose a transferable architecture exploration framework, dubbed Apollo, that leverages recent advances in black-box function optimization for sample-efficient accelerator design. We use this framework to optimize accelerator configurations of a diverse set of neural architectures with alternative design constraints. We show that our framework finds high reward design configurations (up to 24.6% speedup) more sample-efficiently than a baseline black-box optimization approach. We further show that by transferring knowledge between target architectures with different design constraints, Apollo is able to find optimal configurations faster and often with better objective value (up to 25% improvements). This encouraging outcome portrays a promising path forward to facilitate generating higher quality accelerators.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Amir Yazdanbakhsh (38 papers)
  2. Christof Angermueller (6 papers)
  3. Berkin Akin (10 papers)
  4. Yanqi Zhou (30 papers)
  5. Albin Jones (1 paper)
  6. Milad Hashemi (17 papers)
  7. Kevin Swersky (51 papers)
  8. Satrajit Chatterjee (11 papers)
  9. Ravi Narayanaswami (6 papers)
  10. James Laudon (13 papers)
Citations (23)

Summary

We haven't generated a summary for this paper yet.