Evaluating the Impact of Personalized Value Alignment in Human-Robot Interaction: Insights into Trust and Team Performance Outcomes (2311.16051v1)
Abstract: This paper examines the effect of real-time, personalized alignment of a robot's reward function to the human's values on trust and team performance. We present and compare three distinct robot interaction strategies: a non-learner strategy where the robot presumes the human's reward function mirrors its own, a non-adaptive-learner strategy in which the robot learns the human's reward function for trust estimation and human behavior modeling, but still optimizes its own reward function, and an adaptive-learner strategy in which the robot learns the human's reward function and adopts it as its own. Two human-subject experiments with a total number of 54 participants were conducted. In both experiments, the human-robot team searches for potential threats in a town. The team sequentially goes through search sites to look for threats. We model the interaction between the human and the robot as a trust-aware Markov Decision Process (trust-aware MDP) and use Bayesian Inverse Reinforcement Learning (IRL) to estimate the reward weights of the human as they interact with the robot. In Experiment 1, we start our learning algorithm with an informed prior of the human's values/goals. In Experiment 2, we start the learning algorithm with an uninformed prior. Results indicate that when starting with a good informed prior, personalized value alignment does not seem to benefit trust or team performance. On the other hand, when an informed prior is unavailable, alignment to the human's values leads to high trust and higher perceived performance while maintaining the same objective team performance.
- Improving Human-Machine Collaboration Through Transparency-based Feedback – Part I: Human Trust and Workload Model. IFAC-PapersOnLine 51, 34 (2019), 315–321. 2nd IFAC Conference on Cyber-Physical and Human Systems CPHS 2018.
- Improving Human-Machine Collaboration Through Transparency-based Feedback – Part II: Control Design and Synthesis. IFAC-PapersOnLine 51, 34 (2019), 322–328. 2nd IFAC Conference on Cyber-Physical and Human Systems CPHS 2018.
- Saurabh Arora and Prashant Doshi. 2021. A survey of inverse reinforcement learning: Challenges, methods and progress. Artificial Intelligence 297 (Aug. 2021), 103500. https://doi.org/10.1016/j.artint.2021.103500
- Context-Adaptive Management of Drivers’ Trust in Automated Vehicles. IEEE Robotics and Automation Letters 5, 4 (2020), 6908–6915. https://doi.org/10.1109/LRA.2020.3025736
- Clustering Trust Dynamics in a Human-Robot Sequential Decision-Making Task. IEEE Robotics and Automation Letters 7, 4 (2022), 8815–8822. https://doi.org/10.1109/LRA.2022.3188902
- Human-robot interaction: Developing trust in robots. In 2012 7th ACM/IEEE International Conference on Human-Robot Interaction (HRI). 109–110. https://doi.org/10.1145/2157689.2157709
- Erdem Bıyık and Dorsa Sadigh. 2018. Batch Active Preference-Based Learning of Reward Functions. arXiv:1810.04303 [cs.LG]
- Planning with Trust for Human-Robot Collaboration. In Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction (Chicago, IL, USA) (HRI ’18). Association for Computing Machinery, New York, NY, USA, 307–315. https://doi.org/10.1145/3171221.3171264
- Trust-Aware Decision Making for Human-Robot Collaboration: Model Learning and Planning. J. Hum.-Robot Interact. 9, 2, Article 9 (jan 2020), 23 pages. https://doi.org/10.1145/3359616
- Erin K. Chiou and John D. Lee. 2023. Trusting Automation: Designing for Responsivity and Resilience. Human Factors 65, 1 (2023), 137–165. https://doi.org/10.1177/00187208211009995 arXiv:https://doi.org/10.1177/00187208211009995 PMID: 33906505.
- Deep reinforcement learning from human preferences. arXiv:1706.03741 [stat.ML]
- The Dynamics of Trust and Verbal Anthropomorphism in Human-Autonomy Teaming. In 2021 IEEE 2nd International Conference on Human-Machine Systems (ICHMS). 1–6. https://doi.org/10.1109/ICHMS53169.2021.9582655
- Interactive Team Cognition - Cooke - 2013 - Cognitive Science - Wiley Online Library. 37, 2 (2013), 255–285. https://onlinelibrary.wiley.com/doi/full/10.1111/cogs.12009
- Look who’s talking now: Implications of AV’s explanations on driver’s trust, AV preference, anxiety and mental workload. Transportation Research Part C: Emerging Technologies 104 (July 2019), 428–442. https://doi.org/10.1016/j.trc.2019.05.025
- Connor Esterwood and Lionel P. Robert Jr. 2023. Three Strikes and you are out!: The impacts of multiple human–robot trust violations and repairs on robot trustworthiness. Computers in Human Behavior 142 (May 2023), 107658. https://doi.org/10.1016/j.chb.2023.107658
- Pragmatic-Pedagogic Value Alignment. In Robotics Research, Nancy M. Amato, Greg Hager, Shawna Thomas, and Miguel Torres-Torriti (Eds.). Springer International Publishing, Cham, 49–57.
- Reverse Psychology in Trust-Aware Human-Robot Interaction. IEEE Robotics and Automation Letters 6, 3 (2021), 4851–4858. https://doi.org/10.1109/LRA.2021.3067626
- Enabling Team of Teams: A Trust Inference and Propagation (TIP) Model in Multi-Human Multi-Robot Teams. In Robotics: Science and Systems XIX. Robotics: Science and Systems Foundation. https://doi.org/10.15607/RSS.2023.XIX.003
- Yaohui Guo and X. Jessie Yang. 2021. Modeling and Predicting Trust Dynamics in Human-Robot Teaming: A Bayesian Inference Approach. International Journal of Social Robotics (12 2021). https://doi.org/10.1007/s12369-020-00703-3
- Cooperative Inverse Reinforcement Learning. https://doi.org/10.48550/ARXIV.1606.03137
- Sandra G. Hart and Lowell E. Staveland. 1988. Development of NASA-TLX (Task Load Index): Results of Empirical and Theoretical Research. In Human Mental Workload, Peter A. Hancock and Najmedin Meshkati (Eds.). Advances in Psychology, Vol. 52. North-Holland, 139–183.
- ”What If It Is Wrong”: Effects of Power Dynamics and Trust Repair Strategy on Trust and Compliance in HRI. In Proceedings of the 2023 ACM/IEEE International Conference on Human-Robot Interaction (HRI ’23). Association for Computing Machinery, New York, NY, USA, 271–280. https://doi.org/10.1145/3568162.3576964
- Ten challenges for making automation a ”team player” in joint human-agent activity. IEEE Intelligent Systems 19, 6 (Nov. 2004), 91–95. https://doi.org/10.1109/MIS.2004.74 Conference Name: IEEE Intelligent Systems.
- Moral psychology of nursing robots: Exploring the role of robots in dilemmas of patient autonomy. European Journal of Social Psychology 53, 1 (2023), 108–128. https://doi.org/10.1002/ejsp.2890 arXiv:https://onlinelibrary.wiley.com/doi/pdf/10.1002/ejsp.2890
- Individualized Mutual Adaptation in Human-Agent Teams. IEEE Transactions on Human-Machine Systems 51, 6 (2021), 706–714. https://doi.org/10.1109/THMS.2021.3107675
- A workload adaptive haptic shared control scheme for semi-autonomous driving. Accident Analysis & Prevention 152 (2021), 105968. https://doi.org/10.1016/j.aap.2020.105968
- Joseph B. Lyons and Svyatoslav Y. Guznov. 2019. Individual differences in human–machine trust: A multi-study look at the perfect automation schema. Theoretical Issues in Ergonomics Science 20, 4 (2019), 440–458. https://doi.org/10.1080/1463922X.2018.1491071 arXiv:https://doi.org/10.1080/1463922X.2018.1491071
- Explanations and trust: What happens to trust when a robot partner does something unexpected? Computers in Human Behavior 138 (Jan. 2023), 107473. https://doi.org/10.1016/j.chb.2022.107473
- Trusting Autonomous Security Robots: The Role of Reliability and Stated Social Intent. Human Factors 63, 4 (2021), 603–618. https://doi.org/10.1177/0018720820901629 arXiv:https://doi.org/10.1177/0018720820901629 PMID: 32027537.
- Should Robots be Obedient? http://arxiv.org/abs/1705.09990 arXiv:1705.09990 [cs].
- Bonnie Muir and Neville Moray. 1996. Trust in automation. Part II. Experimental studies of trust and human intervention in a process control simulation. Ergonomics 39 (04 1996), 429–60. https://doi.org/10.1080/00140139608964474
- Andrew Y. Ng and Stuart Russell. 2000. Algorithms for Inverse Reinforcement Learning. In in Proc. 17th International Conf. on Machine Learning. Morgan Kaufmann, 663–670.
- Charles Pippin and Henrik Christensen. 2014. Trust modeling in multi-robot patrolling. Proceedings - IEEE International Conference on Robotics and Automation, 59–66. https://doi.org/10.1109/ICRA.2014.6906590
- Lindsay Sanneman and Julie A. Shah. 2023. Validating metrics for reward alignment in human-autonomy teaming. Computers in Human Behavior 146 (Sept. 2023), 107809. https://doi.org/10.1016/j.chb.2023.107809
- Thomas B. Sheridan. 2016. Human–Robot Interaction: Status and Challenges. Human Factors 58, 4 (2016), 525–532. https://doi.org/10.1177/0018720816644364 Publisher: SAGE Publications Inc.
- Multi-task trust transfer for human–robot interaction. The International Journal of Robotics Research 39, 2-3 (2020), 233–249. https://doi.org/10.1177/0278364919866905 arXiv:https://doi.org/10.1177/0278364919866905
- Trust calibration within a human-robot team: Comparing automatically generated explanations. In 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI). 109–116. https://doi.org/10.1109/HRI.2016.7451741
- Building Trust in a Human-Robot Team with Automatically Generated Explanations. Proceedings of the Interservice/Industry Training, Simulation and Education Conference (I/ITSEC) 15315 (2015), 1–12.
- Anqi Xu and Gregory Dudek. 2015. OPTIMo: Online Probabilistic Trust Inference Model for Asymmetric Human-Robot Collaborations. In 2015 10th ACM/IEEE International Conference on Human-Robot Interaction (HRI). 221–228.
- From Trust to Trust Dynamics: Combining Empirical and Computational Approaches to Model and Predict Trust Dynamics in Human-Autonomy Interaction. In Human-Automation Interaction: Transportation, Vincent G. Duffy, Steven J. Landry, John D. Lee, and Neville A. Stanton (Eds.). 253–265.
- Toward Quantifying Trust Dynamics: How People Adjust Their Trust After Moment-to-Moment Interaction With Automation. Human Factors 65, 5 (2023), 862–878. https://doi.org/10.1177/00187208211034716
- Evaluating Effects of User Experience and System Transparency on Trust in Automation. In Proceedings of the 2017 ACM/IEEE International Conference on Human-Robot Interaction - HRI ’17. ACM, New York, NY, USA, 408–416. https://doi.org/10.1145/2909824.3020230
- In situ bidirectional human-robot value alignment. Science Robotics 7, 68 (2022), eabm4183. https://doi.org/10.1126/scirobotics.abm4183
- Trust-Aware Planning: Modeling Trust Evolution in Iterated Human-Robot Interaction. In Proceedings of the 2023 ACM/IEEE International Conference on Human-Robot Interaction. ACM, Stockholm Sweden, 281–289. https://doi.org/10.1145/3568162.3578628
- Modeling Interaction via the Principle of Maximum Causal Entropy. In Proceedings of the 27th International Conference on International Conference on Machine Learning (Haifa, Israel) (ICML’10). Omnipress, Madison, WI, USA, 1255–1262.
- Shreyas Bhat (6 papers)
- Joseph B. Lyons (6 papers)
- Cong Shi (36 papers)
- X. Jessie Yang (38 papers)