Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ORL: Reinforcement Learning Benchmarks for Online Stochastic Optimization Problems (1911.10641v2)

Published 24 Nov 2019 in cs.LG, cs.AI, and math.OC

Abstract: Reinforcement Learning (RL) has achieved state-of-the-art results in domains such as robotics and games. We build on this previous work by applying RL algorithms to a selection of canonical online stochastic optimization problems with a range of practical applications: Bin Packing, Newsvendor, and Vehicle Routing. While there is a nascent literature that applies RL to these problems, there are no commonly accepted benchmarks which can be used to compare proposed approaches rigorously in terms of performance, scale, or generalizability. This paper aims to fill that gap. For each problem we apply both standard approaches as well as newer RL algorithms and analyze results. In each case, the performance of the trained RL policy is competitive with or superior to the corresponding baselines, while not requiring much in the way of domain knowledge. This highlights the potential of RL in real-world dynamic resource allocation problems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Bharathan Balaji (19 papers)
  2. Jordan Bell-Masterson (1 paper)
  3. Enes Bilgin (1 paper)
  4. Andreas Damianou (28 papers)
  5. Pablo Moreno Garcia (1 paper)
  6. Arpit Jain (9 papers)
  7. Runfei Luo (2 papers)
  8. Alvaro Maggiar (2 papers)
  9. Balakrishnan Narayanaswamy (11 papers)
  10. Chun Ye (8 papers)
Citations (31)

Summary

We haven't generated a summary for this paper yet.