REACT: Revealing Evolutionary Action Consequence Trajectories for Interpretable Reinforcement Learning (2404.03359v1)
Abstract: To enhance the interpretability of Reinforcement Learning (RL), we propose Revealing Evolutionary Action Consequence Trajectories (REACT). In contrast to the prevalent practice of validating RL models based on their optimal behavior learned during training, we posit that considering a range of edge-case trajectories provides a more comprehensive understanding of their inherent behavior. To induce such scenarios, we introduce a disturbance to the initial state, optimizing it through an evolutionary algorithm to generate a diverse population of demonstrations. To evaluate the fitness of trajectories, REACT incorporates a joint fitness function that encourages both local and global diversity in the encountered states and chosen actions. Through assessments with policies trained for varying durations in discrete and continuous environments, we demonstrate the descriptive power of REACT. Our results highlight its effectiveness in revealing nuanced aspects of RL models' behavior beyond optimal performance, thereby contributing to improved interpretability.
- Reinforcement Learning Interpretation Methods: A Survey. IEEE Access 8 (2020), 171058–171077. https://doi.org/10.1109/ACCESS.2020.3023394
- Philipp Altmann. 2023. hyphi gym. https://github.com/philippaltmann/hyphi-gym/
- Dan Amir and Ofra Amir. 2018. HIGHLIGHTS: Summarizing Agent Behavior to People. In Adaptive Agents and Multi-Agent Systems. https://api.semanticscholar.org/CorpusID:21755369
- Fatigue and human performance: an updated framework. Sports medicine 53, 1 (2023), 7–31.
- Leveraging procedural generation to benchmark reinforcement learning. In International conference on machine learning. PMLR, 2048–2056.
- Gymnasium Robotics. http://github.com/Farama-Foundation/Gymnasium-Robotics
- David B Fogel. 2006. Evolutionary computation: toward a new philosophy of machine intelligence. John Wiley & Sons.
- D4rl: Datasets for deep data-driven reinforcement learning. arXiv preprint arXiv:2004.07219 (2020).
- Inheritance-based diversity measures for explicit convergence control in evolutionary algorithms. In Proceedings of the Genetic and Evolutionary Computation Conference. 841–848.
- Scenario co-evolution for reinforcement learning on a grid world smart factory domain. In Proceedings of the Genetic and Evolutionary Computation Conference. 898–906.
- Gymnasium-Robotics Contributors. 2022. Gymnasium-Robotics: A a collection of robotics simulation environments for Reinforcement Learning.
- Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor. CoRR abs/1801.01290 (2018). arXiv:1801.01290 http://arxiv.org/abs/1801.01290
- Explainability in deep reinforcement learning. Knowledge-Based Systems 214 (2021), 106685. https://doi.org/10.1016/j.knosys.2020.106685
- Establishing Appropriate Trust via Critical States. CoRR abs/1810.08174 (2018). arXiv:1810.08174 http://arxiv.org/abs/1810.08174
- Enabling Robots to Communicate their Objectives. CoRR abs/1702.03465 (2017). arXiv:1702.03465 http://arxiv.org/abs/1702.03465
- Evolutionary many-objective optimization: A short review. In 2008 IEEE congress on evolutionary computation (IEEE world congress on computational intelligence). IEEE, 2419–2426.
- Shauharda Khadka and Kagan Tumer. 2018. Evolutionary Reinforcement Learning. CoRR abs/1805.07917 (2018). arXiv:1805.07917 http://arxiv.org/abs/1805.07917
- Pang Wei Koh and Percy Liang. 2017. Understanding black-box predictions via influence functions. In International conference on machine learning. PMLR, 1885–1894.
- Exploring Computational User Models for Agent Policy Summarization. CoRR abs/1905.13271 (2019). arXiv:1905.13271 http://arxiv.org/abs/1905.13271
- Interpretable deep learning: interpretation, interpretability, trustworthiness, and beyond. Knowledge and Information Systems (2022).
- Bin Lin and Jianwen Su. 2008. One Way Distance: For Shape Based Similarity Search of Moving Object Trajectories. GeoInformatica 12 (2008), 117–142. https://api.semanticscholar.org/CorpusID:5279325
- Scott M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. Advances in neural information processing systems 30 (2017).
- Evolutionary diversity optimization using multi-objective indicators. In Proceedings of the Genetic and Evolutionary Computation Conference. 837–845.
- Effective Diversity in Population-Based Reinforcement Learning. CoRR abs/2002.00632 (2020). arXiv:2002.00632 https://arxiv.org/abs/2002.00632
- Identifying mislabeled data using the area under the margin ranking. Advances in Neural Information Processing Systems 33 (2020), 17044–17056.
- Martin L Puterman. 1990. Markov decision processes. Handbooks in operations research and management science 2 (1990), 331–434.
- Stable-Baselines3: Reliable Reinforcement Learning Implementations. Journal of Machine Learning Research 22, 268 (2021), 1–8. http://jmlr.org/papers/v22/20-1364.html
- ” Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining. 1135–1144.
- Andrew G. Barto Richard S. Sutton. 2014, 2015. Reinforcement Learning: An Introduction (2 ed.). The MIT Press, Cambridge, Massachusetts, London, England.
- Proximal Policy Optimization Algorithms. arXiv:1707.06347 [cs.LG]
- Pedro Sequeira and Melinda Gervasio. 2020. Interestingness elements for explainable reinforcement learning: Understanding agents’ capabilities and limitations. Artificial Intelligence 288 (2020), 103367. https://doi.org/10.1016/j.artint.2020.103367
- Pirkko Vartiainen. 2002. On the principles of comparative evaluation. Evaluation 8, 3 (2002), 359–371.
- Evolutionary Diversity Optimization with Clustering-based Selection for Reinforcement Learning. In International Conference on Learning Representations. https://openreview.net/forum?id=74x5BXs4bWD
- Mark Wineberg and Franz Oppacher. 2003. The underlying similarity of diversity measures used in evolutionary computation. In Genetic and evolutionary computation conference. Springer, 1493–1504.
- Quality-Similar Diversity via Population Based Reinforcement Learning. In The Eleventh International Conference on Learning Representations. https://openreview.net/forum?id=bLmSMXbqXr
- Outracing champion Gran Turismo drivers with deep reinforcement learning. Nature 602, 7896 (2022), 223–228.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.