Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Design Space Exploration of Approximate Computing Techniques with a Reinforcement Learning Approach (2312.17525v1)

Published 29 Dec 2023 in cs.AR, cs.LG, and cs.PF

Abstract: Approximate Computing (AxC) techniques have become increasingly popular in trading off accuracy for performance gains in various applications. Selecting the best AxC techniques for a given application is challenging. Among proposed approaches for exploring the design space, Machine Learning approaches such as Reinforcement Learning (RL) show promising results. In this paper, we proposed an RL-based multi-objective Design Space Exploration strategy to find the approximate versions of the application that balance accuracy degradation and power and computation time reduction. Our experimental results show a good trade-off between accuracy degradation and decreased power and computation time for some benchmarks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (11)
  1. W. Hu et al., “Exploring the design space of approximate arithmetic circuits using reinforcement learning,” in 2019 Design, Automation & Test in Europe Conference & Exhibition (DATE).   IEEE, 2019, pp. 442–447.
  2. M. Rizakis et al., “Approximate fpga-based lstms under computation time constraints,” in Applied Reconfigurable Computing. Architectures, Tools, and Applications, N. Voros et al., Eds.   Cham: Springer International Publishing, 2018, pp. 3–15.
  3. Y. Wu et al., “Ironman: Reinforcement learning based design space exploration for approximate computing,” in 2021 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).   IEEE, 2021, pp. 1–8.
  4. Q. Gautier et al., “Sherlock: A multi-objective design space exploration framework,” ACM Transactions on Design Automation of Electronic Systems (TODAES), vol. 27, no. 4, pp. 1–20, 2022.
  5. L. P. Kaelbling, M. L. Littman, and A. W. Moore, “Reinforcement learning: A survey,” Journal of artificial intelligence research, vol. 4, pp. 237–285, 1996.
  6. A. Savino et al., “Approximate computing design exploration through data lifetime metrics,” in 2019 IEEE European Test Symposium (ETS), 2019, pp. 1–7.
  7. V. Mrazek et al., “Evoapprox8b: Library of approximate adders and multipliers for circuit design and benchmarking of approximation methods,” in Design, Automation & Test in Europe Conference & Exhibition (DATE), 2017, 2017, pp. 258–261.
  8. M. Češka et al., “Approximating complex arithmetic circuits with formal error guarantees: 32-bit multipliers accomplished,” in 2017 IEEE/ACM International Conference on Computer-Aided Design (ICCAD), 2017, pp. 416–423.
  9. F. S. Melo, “Convergence of q-learning: A simple proof,” Institute Of Systems and Robotics, Tech. Rep, pp. 1–4, 2001.
  10. S. V. Swaroop et al., “Gymnasium,” https://github.com/s-vineet/gymnasium, 2021, accessed: 2023-04-07.
  11. G. Brockman et al., “Openai gym,” arXiv preprint arXiv:1606.01540, 2016.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets