Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Measuring Interventional Robustness in Reinforcement Learning (2209.09058v1)

Published 19 Sep 2022 in cs.LG and cs.AI

Abstract: Recent work in reinforcement learning has focused on several characteristics of learned policies that go beyond maximizing reward. These properties include fairness, explainability, generalization, and robustness. In this paper, we define interventional robustness (IR), a measure of how much variability is introduced into learned policies by incidental aspects of the training procedure, such as the order of training data or the particular exploratory actions taken by agents. A training procedure has high IR when the agents it produces take very similar actions under intervention, despite variation in these incidental aspects of the training procedure. We develop an intuitive, quantitative measure of IR and calculate it for eight algorithms in three Atari environments across dozens of interventions and states. From these experiments, we find that IR varies with the amount of training and type of algorithm and that high performance does not imply high IR, as one might expect.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Katherine Avery (1 paper)
  2. Jack Kenney (2 papers)
  3. Pracheta Amaranath (2 papers)
  4. Erica Cai (8 papers)
  5. David Jensen (66 papers)

Summary

We haven't generated a summary for this paper yet.