Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RHNAS: Realizable Hardware and Neural Architecture Search (2106.09180v1)

Published 17 Jun 2021 in cs.LG, cs.AR, and cs.NE

Abstract: The rapidly evolving field of Artificial Intelligence necessitates automated approaches to co-design neural network architecture and neural accelerators to maximize system efficiency and address productivity challenges. To enable joint optimization of this vast space, there has been growing interest in differentiable NN-HW co-design. Fully differentiable co-design has reduced the resource requirements for discovering optimized NN-HW configurations, but fail to adapt to general hardware accelerator search spaces. This is due to the existence of non-synthesizable (invalid) designs in the search space of many hardware accelerators. To enable efficient and realizable co-design of configurable hardware accelerators with arbitrary neural network search spaces, we introduce RHNAS. RHNAS is a method that combines reinforcement learning for hardware optimization with differentiable neural architecture search. RHNAS discovers realizable NN-HW designs with 1.84x lower latency and 1.86x lower energy-delay product (EDP) on ImageNet and 2.81x lower latency and 3.30x lower EDP on CIFAR-10 over the default hardware accelerator design.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Yash Akhauri (20 papers)
  2. Adithya Niranjan (1 paper)
  3. J. Pablo Muñoz (14 papers)
  4. Suvadeep Banerjee (4 papers)
  5. Abhijit Davare (2 papers)
  6. Pasquale Cocchini (2 papers)
  7. Anton A. Sorokin (1 paper)
  8. Ravi Iyer (14 papers)
  9. Nilesh Jain (18 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.