Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On the consistency of hyper-parameter selection in value-based deep reinforcement learning (2406.17523v3)

Published 25 Jun 2024 in cs.LG and cs.AI

Abstract: Deep reinforcement learning (deep RL) has achieved tremendous success on various domains through a combination of algorithmic design and careful selection of hyper-parameters. Algorithmic improvements are often the result of iterative enhancements built upon prior approaches, while hyper-parameter choices are typically inherited from previous methods or fine-tuned specifically for the proposed technique. Despite their crucial impact on performance, hyper-parameter choices are frequently overshadowed by algorithmic advancements. This paper conducts an extensive empirical study focusing on the reliability of hyper-parameter selection for value-based deep reinforcement learning agents, including the introduction of a new score to quantify the consistency and reliability of various hyper-parameters. Our findings not only help establish which hyper-parameters are most critical to tune, but also help clarify which tunings remain consistent across different training regimes.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Johan Obando-Ceron (18 papers)
  2. João G. M. Araújo (8 papers)
  3. Aaron Courville (201 papers)
  4. Pablo Samuel Castro (54 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.