Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
158 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A User Study on Explainable Online Reinforcement Learning for Adaptive Systems (2307.04098v1)

Published 9 Jul 2023 in cs.SE and cs.AI

Abstract: Online reinforcement learning (RL) is increasingly used for realizing adaptive systems in the presence of design time uncertainty. Online RL facilitates learning from actual operational data and thereby leverages feedback only available at runtime. However, Online RL requires the definition of an effective and correct reward function, which quantifies the feedback to the RL algorithm and thereby guides learning. With Deep RL gaining interest, the learned knowledge is no longer explicitly represented, but is represented as a neural network. For a human, it becomes practically impossible to relate the parametrization of the neural network to concrete RL decisions. Deep RL thus essentially appears as a black box, which severely limits the debugging of adaptive systems. We previously introduced the explainable RL technique XRL-DINE, which provides visual insights into why certain decisions were made at important time points. Here, we introduce an empirical user study involving 54 software engineers from academia and industry to assess (1) the performance of software engineers when performing different tasks using XRL-DINE and (2) the perceived usefulness and ease of use of XRL-DINE.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (61)
  1. RUDDER: Return Decomposition for Delayed Rewards. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d’Alché-Buc, Emily B. Fox, and Roman Garnett (Eds.). 13544–13555.
  2. Self-Explanation in Adaptive Systems.. In 17th IEEE Intl Conf on Eng. of Complex Computer Systems, ICECCS 2012.
  3. Understanding Uncertainty in Self-adaptive Systems. In ACSOS 2020. IEEE, 242–251.
  4. Reinforcement Learning Techniques for Decentralized Self-adaptive Service Assembly. In Service-Oriented and Cloud Computing - 5th IFIP WG 2.14 European Conference, ESOCC 2016, Vienna, Austria, September 5-7, 2016, Proceedings (Lecture Notes in Computer Science, Vol. 9846), Marco Aiello, Einar Broch Johnsen, Schahram Dustdar, and Ilche Georgievski (Eds.). Springer, 53–68.
  5. Do Automatically Generated Test Cases Make Debugging Easier? An Experimental Assessment of Debugging Effectiveness and Efficiency. ACM Trans. Softw. Eng. Methodol. 25, 1 (2015), 5:1–5:38.
  6. On learning in collective self-adaptive systems: state of practice and a 3D framework. In 14th Intl Symp. on Softw. Eng. for Adaptive & Self-Managing Systems, SEAMS@ICSE 2019. ACM.
  7. Reliability of self-rated experience and confidence as predictors for students’ performance in software engineering. Empir. Softw. Eng. 26, 4 (2021), 80.
  8. Fred D Davis. 1989. Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS quarterly (1989), 319–340.
  9. Fred D Davis and Viswanath Venkatesh. 2004. Toward preprototype user acceptance testing of new information systems: implications for software project management. IEEE Transactions on Engineering management 51, 1 (2004), 31–46.
  10. Daniel Dewey. 2014. Reinforcement Learning and the Reward Engineering Principle. In 2014 AAAI Spring Symposia, Stanford University, Palo Alto, California, USA, March 24-26, 2014. AAAI Press.
  11. Finale Doshi-Velez and Been Kim. 2017. Towards A Rigorous Science of Interpretable Machine Learning. CoRR abs/1702.08608 (2017). http://arxiv.org/abs/1702.08608
  12. Designing for situation awareness: An approach to user-centered design. CRC press.
  13. Explaining Online Reinforcement Learning Decisions of Self-Adaptive Systems. In IEEE International Conference on Autonomic Computing and Self-Organizing Systems, ACSOS 2022, Virtual, CA, USA, September 19-23, 2022, Roberto Casadei, Elisabetta Di Nitto, Ilias Gerostathopoulos, Danilo Pianini, Ivana Dusparic, Timothy Wood, Phyllis R. Nelson, Evangelos Pournaras, Nelly Bencomo, Sebastian Götz, Christian Krupitzer, and Claudia Raibulet (Eds.). IEEE, 51–60.
  14. DevOps for Trustworthy Smart IoT Systems. Now Publishers.
  15. Jasmina Gajcin and Ivana Dusparic. 2022. ReCCoVER: Detecting Causal Confusion for Explainable Reinforcement Learning. In 4th Intl Workshop on EXplainable and TRAnsparent AI and Multi-Agent Systems (EXTRAAMAS@AAAMAS 2022).
  16. Chat-XAI: A New Chatbot to Explain Artificial Intelligence. In Intelligent Systems and Applications - Proceedings of the 2021 Intelligent Systems Conference, IntelliSys 2021, Amsterdam, The Netherlands, 2-3 September, 2021, Volume 3 (Lecture Notes in Networks and Systems, Vol. 296), Kohei Arai (Ed.). Springer, 125–134.
  17. Towards History-Aware Self-Adaptation with Explanation Capabilities.. In IEEE 4th Intl Workshops on Foundations and Applications of Self* Systems, FAS*W@SASO/ICCAC 2019.
  18. A Survey of Methods for Explaining Black Box Models. ACM Comput. Surv. 51, 5 (2019).
  19. Deep Reinforcement Learning with Double Q-Learning. In Thirtieth AAAI Conf on Artificial Intelligence.
  20. Explainability in deep reinforcement learning. Knowl. Based Syst. 214 (2021).
  21. Explainable reinforcement learning via reward decomposition. In IJCAI/ECAI Workshop on Explainable Artificial Intelligence.
  22. Daniel A. Keim. 2001. Visual exploration of large data sets. Commun. ACM 44, 8 (2001).
  23. William R King and Jun He. 2006. A meta-analysis of the technology acceptance model. Information & management 43, 6 (2006), 740–755.
  24. Oliver Laitenberger and Horst M Dreyer. 1998. Evaluating the usefulness and the ease of use of a web-based inspection data collection tool. In Proceedings Fifth International Software Metrics Symposium. Metrics (Cat. No. 98TB100262). IEEE, 122–132.
  25. Hey! Preparing Humans to do Tasks in Self-adaptive Systems.. In 16th International Symp. on Softw. Eng. for Adaptive and Self-Managing Systems, SEAMS@ICSE 2021.
  26. Nikola Marangunić and Andrina Granić. 2015. Technology acceptance model: a literature review from 1986 to 2013. Universal access in the information society 14, 1 (2015), 81–95.
  27. Triggering Proactive Business Process Adaptations via Online Reinforcement Learning. In Business Process Management - 18th International Conference, BPM 2020, Seville, Spain, September 13-18, 2020, Proceedings (Lecture Notes in Computer Science, Vol. 12168), Dirk Fahland, Chiara Ghidini, Jörg Becker, and Marlon Dumas (Eds.). Springer, 273–290. https://doi.org/10.1007/978-3-030-58666-9_16
  28. Triggering Proactive Business Process Adaptations via Online Reinforcement Learning. In Business Process Management - 18th International Conference, BPM 2020, Seville, Spain, September 13-18, 2020, Proceedings (Lecture Notes in Computer Science, Vol. 12168), Dirk Fahland, Chiara Ghidini, Jörg Becker, and Marlon Dumas (Eds.). Springer, 273–290.
  29. Andreas Metzger and Clarissa Cassales Marquezan. 2011. Future Internet Apps: The Next Wave of Adaptive Service-Oriented Systems?. In Towards a Service-Based Internet - 4th European Conference, ServiceWave 2011, Poznan, Poland, October 26-28, 2011. Proceedings (Lecture Notes in Computer Science, Vol. 6994), Witold Abramowicz, Ignacio Martín Llorente, Mike Surridge, Andrea Zisman, and Julien Vayssière (Eds.). Springer, 230–241.
  30. Realizing self-adaptive systems via online reinforcement learning and feature-model-guided exploration. Computing (March, 2022).
  31. Tim Miller. 2019. Explanation in artificial intelligence: Insights from the social sciences. Artif. Intell. 267 (2019).
  32. Human-level control through deep reinforcement learning. Nature 518, 7540 (2015).
  33. Comparing Model-Based Predictive Approaches to Self-Adaptation: CobRA and PLA. In 12th IEEE/ACM International Symposium on Software Engineering for Adaptive and Self-Managing Systems, SEAMS@ICSE 2017, Buenos Aires, Argentina, May 22-23, 2017. IEEE Computer Society, 42–53.
  34. SWIM: an exemplar for evaluation and comparison of self-adaptation approaches for web applications. In 13th Intl Conf on Software Engineering for Adaptive and Self-Managing Systems, SEAMS@ICSE 2018. ACM.
  35. Michael G Morris and Andrew Dillon. 1997. How user perceptions influence software use. IEEE software 14, 4 (1997), 58–65.
  36. Ahmed Moustafa and Takayuki Ito. 2018. A deep reinforcement learning approach for large-scale service composition. In Intl. Conf. on Principles and Practice of Multi-Agent Systems.
  37. Ahmed Moustafa and Minjie Zhang. 2014. Learning Efficient Compositions for QoS-Aware Service Provisioning. In 2014 IEEE International Conference on Web Services, ICWS, 2014, Anchorage, AK, USA, June 27 - July 2, 2014. IEEE Computer Society, 185–192.
  38. Bridging the Gap Between Value and Policy Based Reinforcement Learning. In Advances in Neural Information Processing Systems 12 (NIPS 2017). 2772–2782.
  39. Online Reinforcement Learning for Self-adaptive Information Systems. In Advanced Information Systems Engineering - 32nd International Conference, CAiSE 2020, Grenoble, France, June 8-12, 2020, Proceedings (Lecture Notes in Computer Science, Vol. 12127), Schahram Dustdar, Eric Yu, Camille Salinesi, Dominique Rieu, and Vik Pant (Eds.). Springer, 169–184. https://doi.org/10.1007/978-3-030-49435-3_11
  40. Online Reinforcement Learning for Self-Adaptive Information Systems. In Int’l Conference on Advanced Information Systems Engineering, CAiSE’20, Eric Yu and Schahram Dustdar (Eds.).
  41. Curiosity-driven Exploration by Self-supervised Prediction. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017 (Proceedings of Machine Learning Research, Vol. 70), Doina Precup and Yee Whye Teh (Eds.). PMLR, 2778–2787.
  42. The impact of time controlled reading on software inspection effectiveness and efficiency: a controlled experiment. In Proceedings of the Second International Symposium on Empirical Software Engineering and Measurement, ESEM 2008, October 9-10, 2008, Kaiserslautern, Germany, H. Dieter Rombach, Sebastian G. Elbaum, and Jürgen Münch (Eds.). ACM, 139–148.
  43. A Survey of Methodology in Self-Adaptive Systems Research. In ACSOS. IEEE.
  44. Erika Puiutta and Eric M. S. P. Veith. 2020. Explainable Reinforcement Learning: A Survey. In Machine Learning and Knowledge Extraction - 4th IFIP Cross-Domain Conference, CD-MAKE 2020 (LNCS, Vol. 12279). Springer.
  45. Decentralized Self-Adaptive Systems: A Mapping Study. In 16th International Symposium on Software Engineering for Adaptive and Self-Managing Systems, SEAMS@ICSE 2021, Madrid, Spain, May 18-24, 2021. IEEE, 18–29.
  46. Automated provenance graphs for [email protected].. In MODELS ’20: ACM/IEEE 23rd Intl Conf on Model Driven Engineering Languages and Systems.
  47. ”Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, August 13-17, 2016, Balaji Krishnapuram, Mohak Shah, Alexander J. Smola, Charu C. Aggarwal, Dou Shen, and Rajeev Rastogi (Eds.). ACM, 1135–1144.
  48. Cynthia Rudin. 2019. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1, 5 (2019), 206–215.
  49. Wojciech Samek and Klaus-Robert Müller. 2019. Towards Explainable Artificial Intelligence. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning, Wojciech Samek, Grégoire Montavon, Andrea Vedaldi, Lars Kai Hansen, and Klaus-Robert Müller (Eds.). Lecture Notes in Computer Science, Vol. 11700. Springer, 5–22.
  50. Pedro Sequeira and Melinda T. Gervasio. 2020. Interestingness elements for explainable reinforcement learning: Understanding agents’ capabilities and limitations. Artif. Intell. 288 (2020).
  51. Richard S Sutton and Andrew G Barto. 2018. Reinforcement learning: An introduction. MIT press.
  52. Policy Gradient Methods for Reinforcement Learning with Function Approximation. In Advances in Neural Information Processing Systems 12 (NIPS 1999). 1057–1063.
  53. Event-driven temporal models for explanations - ETeMoX: explaining reinforcement learning. Softw. Syst. Model. 21, 3 (2022).
  54. Adaptive and large-scale service composition based on deep reinforcement learning. Knowledge-Based Systems 180 (2019).
  55. Daniel S. Weld and Gagan Bansal. 2019. The challenge of crafting intelligible intelligence. Commun. ACM 62, 6 (2019), 70–79.
  56. Self-Explanation in Adaptive Systems Based on Runtime Goal-Based Models. Trans. Comput. Collect. Intell. 16 (2014).
  57. Danny Weyns. 2020. An Introduction to Self-adaptive Systems: A Contemporary Software Engineering Perspective. John Wiley & Sons.
  58. Perpetual Assurances for Self-Adaptive Systems. In Software Engineering for Self-Adaptive Systems III. Assurances (LNCS, Vol. 9640). Springer.
  59. Preliminary Results of a Survey on the Use of Self-Adaptation in Industry. In 17th Intl Symp. on Software Engineering for Adaptive and Self-Managing Systems, SEAMS@ICSE 2022.
  60. A Meta Reinforcement Learning-based Approach for Self-Adaptive System. In IEEE International Conference on Autonomic Computing and Self-Organizing Systems, ACSOS 2021, Washington, DC, USA, September 27 - Oct. 1, 2021, Esam El-Araby, Vana Kalogeraki, Danilo Pianini, Frédéric Lassabe, Barry Porter, Sona Ghahremani, Ingrid Nunes, Mohamed Bakhouya, and Sven Tomforde (Eds.). IEEE, 1–10.
  61. Anomaly Detection and Classification to enable Self-Explainability of Autonomous Systems. In Design, Automation & Test in Europe Conference & Exhibition, DATE 2021, Grenoble, France, February 1-5, 2021. IEEE, 1304–1309.
Citations (1)

Summary

We haven't generated a summary for this paper yet.