Ensure Retrainability of Primal IRL Rewards
Establish a practical method to ensure that the final reward function returned by primal inverse reinforcement learning algorithms that alternate small updates to the discriminator and the policy (such as GAIL-style approaches solving min_π max_f J(π_E, f) − J(π, f)) permits effective retraining from scratch; specifically, when the recovered reward is optimised via reinforcement learning in the original environment to completion, the resulting policy should match the expert’s performance.
References
We therefore are left with our first open question to ponder: Challenge 1: How do we ensure the final reward function returned by primal IRL methods permits effective (even if not efficient) retraining from scratch?
— EvIL: Evolution Strategies for Generalisable Imitation Learning
(2406.11905 - Sapora et al., 15 Jun 2024) in Section “Reward-Centric Challenges of Efficient IRL”, Challenge 1