Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Evaluating the Impact of Personalized Value Alignment in Human-Robot Interaction: Insights into Trust and Team Performance Outcomes (2311.16051v1)

Published 27 Nov 2023 in cs.RO

Abstract: This paper examines the effect of real-time, personalized alignment of a robot's reward function to the human's values on trust and team performance. We present and compare three distinct robot interaction strategies: a non-learner strategy where the robot presumes the human's reward function mirrors its own, a non-adaptive-learner strategy in which the robot learns the human's reward function for trust estimation and human behavior modeling, but still optimizes its own reward function, and an adaptive-learner strategy in which the robot learns the human's reward function and adopts it as its own. Two human-subject experiments with a total number of 54 participants were conducted. In both experiments, the human-robot team searches for potential threats in a town. The team sequentially goes through search sites to look for threats. We model the interaction between the human and the robot as a trust-aware Markov Decision Process (trust-aware MDP) and use Bayesian Inverse Reinforcement Learning (IRL) to estimate the reward weights of the human as they interact with the robot. In Experiment 1, we start our learning algorithm with an informed prior of the human's values/goals. In Experiment 2, we start the learning algorithm with an uninformed prior. Results indicate that when starting with a good informed prior, personalized value alignment does not seem to benefit trust or team performance. On the other hand, when an informed prior is unavailable, alignment to the human's values leads to high trust and higher perceived performance while maintaining the same objective team performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (45)
  1. Improving Human-Machine Collaboration Through Transparency-based Feedback – Part I: Human Trust and Workload Model. IFAC-PapersOnLine 51, 34 (2019), 315–321. 2nd IFAC Conference on Cyber-Physical and Human Systems CPHS 2018.
  2. Improving Human-Machine Collaboration Through Transparency-based Feedback – Part II: Control Design and Synthesis. IFAC-PapersOnLine 51, 34 (2019), 322–328. 2nd IFAC Conference on Cyber-Physical and Human Systems CPHS 2018.
  3. Saurabh Arora and Prashant Doshi. 2021. A survey of inverse reinforcement learning: Challenges, methods and progress. Artificial Intelligence 297 (Aug. 2021), 103500. https://doi.org/10.1016/j.artint.2021.103500
  4. Context-Adaptive Management of Drivers’ Trust in Automated Vehicles. IEEE Robotics and Automation Letters 5, 4 (2020), 6908–6915. https://doi.org/10.1109/LRA.2020.3025736
  5. Clustering Trust Dynamics in a Human-Robot Sequential Decision-Making Task. IEEE Robotics and Automation Letters 7, 4 (2022), 8815–8822. https://doi.org/10.1109/LRA.2022.3188902
  6. Human-robot interaction: Developing trust in robots. In 2012 7th ACM/IEEE International Conference on Human-Robot Interaction (HRI). 109–110. https://doi.org/10.1145/2157689.2157709
  7. Erdem Bıyık and Dorsa Sadigh. 2018. Batch Active Preference-Based Learning of Reward Functions. arXiv:1810.04303 [cs.LG]
  8. Planning with Trust for Human-Robot Collaboration. In Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction (Chicago, IL, USA) (HRI ’18). Association for Computing Machinery, New York, NY, USA, 307–315. https://doi.org/10.1145/3171221.3171264
  9. Trust-Aware Decision Making for Human-Robot Collaboration: Model Learning and Planning. J. Hum.-Robot Interact. 9, 2, Article 9 (jan 2020), 23 pages. https://doi.org/10.1145/3359616
  10. Erin K. Chiou and John D. Lee. 2023. Trusting Automation: Designing for Responsivity and Resilience. Human Factors 65, 1 (2023), 137–165. https://doi.org/10.1177/00187208211009995 arXiv:https://doi.org/10.1177/00187208211009995 PMID: 33906505.
  11. Deep reinforcement learning from human preferences. arXiv:1706.03741 [stat.ML]
  12. The Dynamics of Trust and Verbal Anthropomorphism in Human-Autonomy Teaming. In 2021 IEEE 2nd International Conference on Human-Machine Systems (ICHMS). 1–6. https://doi.org/10.1109/ICHMS53169.2021.9582655
  13. Interactive Team Cognition - Cooke - 2013 - Cognitive Science - Wiley Online Library. 37, 2 (2013), 255–285. https://onlinelibrary.wiley.com/doi/full/10.1111/cogs.12009
  14. Look who’s talking now: Implications of AV’s explanations on driver’s trust, AV preference, anxiety and mental workload. Transportation Research Part C: Emerging Technologies 104 (July 2019), 428–442. https://doi.org/10.1016/j.trc.2019.05.025
  15. Connor Esterwood and Lionel P. Robert Jr. 2023. Three Strikes and you are out!: The impacts of multiple human–robot trust violations and repairs on robot trustworthiness. Computers in Human Behavior 142 (May 2023), 107658. https://doi.org/10.1016/j.chb.2023.107658
  16. Pragmatic-Pedagogic Value Alignment. In Robotics Research, Nancy M. Amato, Greg Hager, Shawna Thomas, and Miguel Torres-Torriti (Eds.). Springer International Publishing, Cham, 49–57.
  17. Reverse Psychology in Trust-Aware Human-Robot Interaction. IEEE Robotics and Automation Letters 6, 3 (2021), 4851–4858. https://doi.org/10.1109/LRA.2021.3067626
  18. Enabling Team of Teams: A Trust Inference and Propagation (TIP) Model in Multi-Human Multi-Robot Teams. In Robotics: Science and Systems XIX. Robotics: Science and Systems Foundation. https://doi.org/10.15607/RSS.2023.XIX.003
  19. Yaohui Guo and X. Jessie Yang. 2021. Modeling and Predicting Trust Dynamics in Human-Robot Teaming: A Bayesian Inference Approach. International Journal of Social Robotics (12 2021). https://doi.org/10.1007/s12369-020-00703-3
  20. Cooperative Inverse Reinforcement Learning. https://doi.org/10.48550/ARXIV.1606.03137
  21. Sandra G. Hart and Lowell E. Staveland. 1988. Development of NASA-TLX (Task Load Index): Results of Empirical and Theoretical Research. In Human Mental Workload, Peter A. Hancock and Najmedin Meshkati (Eds.). Advances in Psychology, Vol. 52. North-Holland, 139–183.
  22. ”What If It Is Wrong”: Effects of Power Dynamics and Trust Repair Strategy on Trust and Compliance in HRI. In Proceedings of the 2023 ACM/IEEE International Conference on Human-Robot Interaction (HRI ’23). Association for Computing Machinery, New York, NY, USA, 271–280. https://doi.org/10.1145/3568162.3576964
  23. Ten challenges for making automation a ”team player” in joint human-agent activity. IEEE Intelligent Systems 19, 6 (Nov. 2004), 91–95. https://doi.org/10.1109/MIS.2004.74 Conference Name: IEEE Intelligent Systems.
  24. Moral psychology of nursing robots: Exploring the role of robots in dilemmas of patient autonomy. European Journal of Social Psychology 53, 1 (2023), 108–128. https://doi.org/10.1002/ejsp.2890 arXiv:https://onlinelibrary.wiley.com/doi/pdf/10.1002/ejsp.2890
  25. Individualized Mutual Adaptation in Human-Agent Teams. IEEE Transactions on Human-Machine Systems 51, 6 (2021), 706–714. https://doi.org/10.1109/THMS.2021.3107675
  26. A workload adaptive haptic shared control scheme for semi-autonomous driving. Accident Analysis & Prevention 152 (2021), 105968. https://doi.org/10.1016/j.aap.2020.105968
  27. Joseph B. Lyons and Svyatoslav Y. Guznov. 2019. Individual differences in human–machine trust: A multi-study look at the perfect automation schema. Theoretical Issues in Ergonomics Science 20, 4 (2019), 440–458. https://doi.org/10.1080/1463922X.2018.1491071 arXiv:https://doi.org/10.1080/1463922X.2018.1491071
  28. Explanations and trust: What happens to trust when a robot partner does something unexpected? Computers in Human Behavior 138 (Jan. 2023), 107473. https://doi.org/10.1016/j.chb.2022.107473
  29. Trusting Autonomous Security Robots: The Role of Reliability and Stated Social Intent. Human Factors 63, 4 (2021), 603–618. https://doi.org/10.1177/0018720820901629 arXiv:https://doi.org/10.1177/0018720820901629 PMID: 32027537.
  30. Should Robots be Obedient? http://arxiv.org/abs/1705.09990 arXiv:1705.09990 [cs].
  31. Bonnie Muir and Neville Moray. 1996. Trust in automation. Part II. Experimental studies of trust and human intervention in a process control simulation. Ergonomics 39 (04 1996), 429–60. https://doi.org/10.1080/00140139608964474
  32. Andrew Y. Ng and Stuart Russell. 2000. Algorithms for Inverse Reinforcement Learning. In in Proc. 17th International Conf. on Machine Learning. Morgan Kaufmann, 663–670.
  33. Charles Pippin and Henrik Christensen. 2014. Trust modeling in multi-robot patrolling. Proceedings - IEEE International Conference on Robotics and Automation, 59–66. https://doi.org/10.1109/ICRA.2014.6906590
  34. Lindsay Sanneman and Julie A. Shah. 2023. Validating metrics for reward alignment in human-autonomy teaming. Computers in Human Behavior 146 (Sept. 2023), 107809. https://doi.org/10.1016/j.chb.2023.107809
  35. Thomas B. Sheridan. 2016. Human–Robot Interaction: Status and Challenges. Human Factors 58, 4 (2016), 525–532. https://doi.org/10.1177/0018720816644364 Publisher: SAGE Publications Inc.
  36. Multi-task trust transfer for human–robot interaction. The International Journal of Robotics Research 39, 2-3 (2020), 233–249. https://doi.org/10.1177/0278364919866905 arXiv:https://doi.org/10.1177/0278364919866905
  37. Trust calibration within a human-robot team: Comparing automatically generated explanations. In 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI). 109–116. https://doi.org/10.1109/HRI.2016.7451741
  38. Building Trust in a Human-Robot Team with Automatically Generated Explanations. Proceedings of the Interservice/Industry Training, Simulation and Education Conference (I/ITSEC) 15315 (2015), 1–12.
  39. Anqi Xu and Gregory Dudek. 2015. OPTIMo: Online Probabilistic Trust Inference Model for Asymmetric Human-Robot Collaborations. In 2015 10th ACM/IEEE International Conference on Human-Robot Interaction (HRI). 221–228.
  40. From Trust to Trust Dynamics: Combining Empirical and Computational Approaches to Model and Predict Trust Dynamics in Human-Autonomy Interaction. In Human-Automation Interaction: Transportation, Vincent G. Duffy, Steven J. Landry, John D. Lee, and Neville A. Stanton (Eds.). 253–265.
  41. Toward Quantifying Trust Dynamics: How People Adjust Their Trust After Moment-to-Moment Interaction With Automation. Human Factors 65, 5 (2023), 862–878. https://doi.org/10.1177/00187208211034716
  42. Evaluating Effects of User Experience and System Transparency on Trust in Automation. In Proceedings of the 2017 ACM/IEEE International Conference on Human-Robot Interaction - HRI ’17. ACM, New York, NY, USA, 408–416. https://doi.org/10.1145/2909824.3020230
  43. In situ bidirectional human-robot value alignment. Science Robotics 7, 68 (2022), eabm4183. https://doi.org/10.1126/scirobotics.abm4183
  44. Trust-Aware Planning: Modeling Trust Evolution in Iterated Human-Robot Interaction. In Proceedings of the 2023 ACM/IEEE International Conference on Human-Robot Interaction. ACM, Stockholm Sweden, 281–289. https://doi.org/10.1145/3568162.3578628
  45. Modeling Interaction via the Principle of Maximum Causal Entropy. In Proceedings of the 27th International Conference on International Conference on Machine Learning (Haifa, Israel) (ICML’10). Omnipress, Madison, WI, USA, 1255–1262.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Shreyas Bhat (6 papers)
  2. Joseph B. Lyons (6 papers)
  3. Cong Shi (36 papers)
  4. X. Jessie Yang (38 papers)
Citations (8)

Summary

  • The paper demonstrates that adaptive learner robots significantly enhance trust and reliance when they adjust to human values through IRL.
  • It employs controlled experiments comparing informed and uninformed priors to assess the effects of value alignment on team performance.
  • Findings suggest that robots adopting human values can optimize trust dynamics without compromising objective task outcomes.

Understanding Human-Robot Value Alignment

Introduction to Value Alignment in HRI

Robots are entering various facets of human society, assuming roles that require them to work in tandem with humans. As such, the ability for robots to align their values with human teammates becomes crucial, particularly to foster trust and enhance team performance. This paper sheds light on the dynamic relationship between human-robot value alignment and its impact on trust and performance within a team setting.

The Experiment and Its Design

The paper explores the interaction between humans and robots using three distinct robot strategies:

  • The Non-Learner strategy, where the robot assumes that the human shares its values without learning or adapting.
  • The Non-Adaptive Learner strategy, where the robot learns the human's values but still prioritizes its own during decision-making.
  • The Adaptive Learner strategy, where the robot not only learns the human's values but also adopts them as its own.

Two separate experiments were conducted involving participants tasked with finding threats in a virtual environment under a risk of exposure. Different from prior studies, which usually presuppose value alignment as beneficial, this paper presents an empirical evaluation of its effects. Through the lens of a trust-aware Markov Decision Process (MDP) and Bayesian Inverse Reinforcement Learning (IRL), the adaptive strategies were compared against a backdrop of informed and uninformed priors on human values.

Results and Insights

  • Experiment 1 (Informed Prior): When the IRL learning algorithm was initiated with a priori knowledge about human reward weights, no notable benefits were seen across the board for the learning and adaptive strategies. The only exception was a higher stated reliance intention for the non-learner approach compared to the adaptive learner. Surprisingly, a more predictable (non-adaptive) robot inclined users toward increased reliance.
  • Experiment 2 (Uninformed Prior): Without an informed prior, the advantages of value alignment became evident. The adaptive-learner strategy exhibited a significant increase in trust, agreement with robot recommendations, reliance intentions, and perceived team performance, without sacrificing objective performance.

Implications and Concluding Thoughts

The findings are pivotal for real-world applications where robots interact with humans, especially in scenarios where a robot's initial understanding of human values is not accessible. Ensuring that robots can adapt their values to those of their human counterparts can lead to improved trust dynamics and team outcomes. This research contributes to understanding how human-robot partnerships may function optimally, emphasizing the need for robots that can understand and align with human values.

This paper is also mindful of its scope and limitations, encouraging further research to generalize findings across diverse demographics and complex decision-making situations. The necessity for adaptable and trust-conscious robots in a collaborative future is clear; how we mold these artificial entities to reflect human moral and ethical systems will be the subject of continued and necessary debate and investigation.

X Twitter Logo Streamline Icon: https://streamlinehq.com