Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3687272.3688319acmconferencesArticle/Chapter ViewAbstractPublication PageshaiConference Proceedingsconference-collections
research-article
Open access

GPT-4 as a Moral Reasoner for Robot Command Rejection

Published: 24 November 2024 Publication History

Abstract

To support positive, ethical human-robot interactions, robots need to be able to respond to unexpected situations in which societal norms are violated, including rejecting unethical commands. Implementing robust communication for robots is inherently difficult due to the variability of context in real-world settings and the risks of unintended influence during robots’ communication. HRI researchers have begun exploring the potential use of LLMs as a solution for language-based communication, which will require an in-depth understanding and evaluation of LLM applications in different contexts. In this work, we explore how an existing LLM responds to and reasons about a set of norm-violating requests in HRI contexts. We ask human participants to assess the performance of a hypothetical GPT-4-based robot on moral reasoning and explanatory language selection as it compares to human intuitions. Our findings suggest that while GPT-4 performs well at identifying norm violation requests and suggesting non-compliant responses, its flaws in not matching the linguistic preferences and context sensitivity of humans prevent it from being a comprehensive solution for moral communication between humans and robots. Based on our results, we provide a four-point recommendation for the community in incorporating LLMs into HRI systems.

References

[1]
Dan Amir and Ofra Amir. 2018. Highlights: Summarizing agent behavior to people. In Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems. 1168–1176.
[2]
Ofra Amir, Finale Doshi-Velez, and David Sarne. 2018. Agent strategy summarization. In Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems. 1203–1207.
[3]
Sule Anjomshoae, Amro Najjar, Davide Calvaresi, and Kary Främling. 2019. Explainable agents and robots: Results from a systematic literature review. In 18th International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2019), Montreal, Canada, May 13–17, 2019. International Foundation for Autonomous Agents and Multiagent Systems, 1078–1088.
[4]
Christoph Bartneck, Timo Bleeker, Jeroen Bun, Pepijn Fens, and Lynyrd Riet. 2010. The influence of robot anthropomorphism on the feelings of embarrassment when interacting with robots. Paladyn 1, 2 (2010), 109–115.
[5]
Andrea Bonarini. 2020. Communication in human-robot interaction. Current Robotics Reports 1, 4 (2020), 279–285.
[6]
Cynthia Breazeal. 2004. Designing sociable robots. MIT press.
[7]
Gordon Briggs, Tom Williams, Ryan Blake Jackson, and Matthias Scheutz. 2021. Why and How Robots Should Say ‘No’. International Journal of Social Robotics (2021), 1–17.
[8]
Gordon Michael Briggs and Matthias Scheutz. 2015. “Sorry, I can’t do that”: Developing Mechanisms to Appropriately Reject Directives in Human-Robot Interactions. In 2015 AAAI fall symposium series.
[9]
Penelope Brown, Stephen C Levinson, and Stephen C Levinson. 1987. Politeness: Some universals in language usage. Vol. 4. Cambridge university press.
[10]
Derek Cormier, Gem Newman, Masayuki Nakane, James E Young, and Stephane Durocher. 2013. Would you do as a robot commands? An obedience study for human-robot interaction. In The 1st international conference on human–agent interaction.
[11]
Devleena Das, Siddhartha Banerjee, and Sonia Chernova. 2021. Explainable ai for robot failures: Generating explanations that improve user assistance in fault recovery. In Proceedings of the 2021 ACM/IEEE International Conference on Human-Robot Interaction. 351–360.
[12]
Terrence Fong, Illah Nourbakhsh, and Kerstin Dautenhahn. 2003. A survey of socially interactive robots. Robotics and autonomous systems 42, 3-4 (2003), 143–166.
[13]
Norina Gasteiger, Mehdi Hellou, and Ho Seok Ahn. 2023. Factors for personalization and localization to optimize human–robot interaction: A literature review. International Journal of Social Robotics 15, 4 (2023), 689–701.
[14]
Yoyo Tsung-Yu Hou, EunJeong Cheon, and Malte F Jung. 2024. Power in Human-Robot Interaction. In Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction. 269–282.
[15]
Ryan Blake Jackson, Sihui Li, Santosh Balajee Banisetty, Sriram Siva, Hao Zhang, Neil Dantam, and Tom Williams. 2021. An Integrated Approach to Context-Sensitive Moral Cognition in Robot Cognitive Architectures. In IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[16]
Ryan Blake Jackson, Ruchen Wen, and Tom Williams. 2019. Tact in noncompliance: The need for pragmatically apt responses to unethical commands. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society. 499–505.
[17]
Ryan Blake Jackson and Tom Williams. 2018. Robot: Asker of questions and changer of norms. Proceedings of ICRES (2018).
[18]
Ryan Blake Jackson and Tom Williams. 2019. Language-capable robots may inadvertently weaken human moral norms. In 2019 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, 401–410.
[19]
Ryan Blake Jackson and Tom Williams. 2021. A theory of social agency for human-robot interaction. Frontiers in Robotics and AI 8 (2021), 687726.
[20]
Ryan Blake Jackson, Tom Williams, and Nicole Smith. 2020. Exploring the role of gender in perceptions of robotic noncompliance. In Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction. 559–567.
[21]
Malte F Jung, Nikolas Martelaro, and Pamela J Hinds. 2015. Using robots to moderate team conflict: the case of repairing violations. In Proceedings of the tenth annual ACM/IEEE international conference on human-robot interaction. 229–236.
[22]
Ulas Berk Karli, Juo-Tung Chen, Victor Nikhil Antony, and Chien-Ming Huang. 2024. Alchemist: LLM-Aided End-User Development of Robot Applications. In Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction. 361–370.
[23]
Daniel Kasenberg, Antonio Roque, Ravenna Thielstrom, Meia Chita-Tegmark, and Matthias Scheutz. 2019. Generating justifications for norm-related agent decisions. In Proceedings of the 12th International Conference on Natural Language Generation. 484–493.
[24]
Boyoung Kim, Ruchen Wen, Ewart J de Visser, Qin Zhu, Tom Williams, and Elizabeth Phillips. 2021. Investigating Robot Moral Advice to Deter Cheating Behavior. In RO-MAN TSAR Workshop.
[25]
Boyoung Kim, Ruchen Wen, Qin Zhu, Tom Williams, and Elizabeth Phillips. 2021. Robots as Moral Advisors: The Effects of Deontological, Virtue, and Confucian Role Ethics on Encouraging Honest Behavior. In Companion of the 2021 ACM/IEEE International Conference on Human-Robot Interaction. 10–18.
[26]
Callie Y Kim, Christine P Lee, and Bilge Mutlu. 2024. Understanding Large-Language Model (LLM)-powered Human-Robot Interaction. In Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction. 371–380.
[27]
Benjamin Kuipers. 2016. Human-like morality and ethics for robots. In Workshops at the Thirtieth AAAI Conference on Artificial Intelligence.
[28]
Gregory LeMasurier, Alvika Gautam, Zhao Han, Jacob W Crandall, and Holly A Yanco. 2024. Reactive or proactive? how robots should explain failures. In Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction. 413–422.
[29]
Meghann Lomas, Robert Chevalier, Ernest Vincent Cross, Robert Christopher Garrett, John Hoare, and Michael Kopack. 2012. Explaining robot actions. In Proceedings of the seventh annual ACM/IEEE international conference on Human-Robot Interaction. 187–188.
[30]
Karthik Mahadevan, Jonathan Chien, Noah Brown, Zhuo Xu, Carolina Parada, Fei Xia, Andy Zeng, Leila Takayama, and Dorsa Sadigh. 2024. Generative expressive robot behaviors using large language models. In Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction. 482–491.
[31]
Bertram F Malle. 2016. Integrating Robot Ethics and Machine Morality: The Study and Design of Moral Competence in Robots. Ethics and Info. Tech. (2016).
[32]
Bertram F Malle and Matthias Scheutz. 2014. Moral competence in social robots. In 2014 IEEE international symposium on ethics in science, technology and engineering. IEEE.
[33]
Matthew Marge, Carol Espy-Wilson, Nigel G. Ward, Abeer Alwan, Yoav Artzi, Mohit Bansal, Gil Blankenship, Joyce Chai, Hal Daumé, Debadeepta Dey, Mary Harper, Thomas Howard, Casey Kennington, Ivana Kruijff-Korbayová, Dinesh Manocha, Cynthia Matuszek, Ross Mead, Raymond Mooney, Roger K. Moore, Mari Ostendorf, Heather Pon-Barry, Alexander I. Rudnicky, Matthias Scheutz, Robert St. Amant, Tong Sun, Stefanie Tellex, David Traum, and Zhou Yu. 2022. Spoken language interaction with robots: Recommendations for future research. Computer Speech & Language 71 (2022), 101255. https://www.sciencedirect.com/science/article/pii/S0885230821000620
[34]
Cynthia Matuszek, Nick Depalma, Ross Mead, Tom Williams, and Ruchen Wen. 2024. Scarecrows in Oz: Large Language Models in HRI. In Companion of the 2024 ACM/IEEE International Conference on Human-Robot Interaction. 1338–1340.
[35]
Terran Mott, Aaron Fanganello, and Tom Williams. 2024. What a Thing to Say! Which Linguistic Politeness Strategies Should Robots Use in Noncompliance Interactions?. In Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction. 501–510.
[36]
Terran Mott and Tom Williams. 2024. Hidden Scarecrows: Potential Consequences of Inaccurate Assumptions About LLMs in Robotic Moral Reasoning. In Proceedings of the HRI Workshop on Scarecrows in Oz: Large Language Models in HRI.
[37]
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, and Niko Suenderhauf. 2023. Sayplan: Grounding large language models using 3d scene graphs for scalable task planning. arXiv preprint arXiv:2307.06135 (2023).
[38]
Daniel J Rea, Denise Geiskkovitch, and James E Young. 2017. Wizard of Awwws: Exploring psychological impact on the researchers in social HRI experiments. In Proceedings of the Companion of the 2017 ACM/IEEE International Conference on Human-Robot Interaction.
[39]
Avi Rosenfeld and Ariella Richardson. 2019. Explainability in human–agent systems. Autonomous agents and multi-agent systems 33 (2019), 673–705.
[40]
Fatai Sado, Chu Kiong Loo, Wei Shiung Liew, Matthias Kerzel, and Stefan Wermter. 2023. Explainable goal-driven agents and robots-a comprehensive review. Comput. Surveys 55, 10 (2023), 1–41.
[41]
Tatsuya Sakai and Takayuki Nagai. 2022. Explainable autonomous robots: a survey and perspective. Advanced Robotics 36, 5-6 (2022), 219–238.
[42]
Sarah Sebo, Brett Stoll, Brian Scassellati, and Malte F Jung. 2020. Robots in groups and teams: a literature review. Proceedings of the ACM on Human-Computer Interaction 4, CSCW2 (2020), 1–36.
[43]
Rossitza Setchi, Maryam Banitalebi Dehkordi, and Juwairiya Siraj Khan. 2020. Explainable robotics in human-robot interactions. Procedia Computer Science 176 (2020), 3057–3066.
[44]
SP Sharan, Francesco Pittaluga, Manmohan Chandraker, 2023. Llm-assist: Enhancing closed-loop planning with language-based reasoning. arXiv preprint arXiv:2401.00125 (2023).
[45]
Andrew Silva, Pradyumna Tambwekar, Mariah Schrum, and Matthew Gombolay. 2024. Towards Balancing Preference and Performance through Adaptive Personalized Explainability. In Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction. 658–668.
[46]
Chan Hee Song, Jiaman Wu, Clayton Washington, Brian M Sadler, Wei-Lun Chao, and Yu Su. 2023. Llm-planner: Few-shot grounded planning for embodied agents with large language models. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 2998–3009.
[47]
Sonja Stange and Stefan Kopp. 2020. Effects of a Social Robot’s Self-Explanations on How Humans Understand and Evaluate Its Behavior. In Proceedings of the 2020 ACM/IEEE international conference on human-robot interaction. 619–627.
[48]
Gabriele Trovato, Massimiliano Zecca, Salvatore Sessa, Lorenzo Jamone, Jaap Ham, Kenji Hashimoto, and Atsuo Takanishi. 2013. Cross-cultural study on human-robot greeting interaction: acceptance and discomfort by Egyptians and Japanese. Paladyn, Journal of Behavioral Robotics 4, 2 (2013), 83–93.
[49]
Lennart Wachowiak, Andrew Fenn, Haris Kamran, Andrew Coles, Oya Celiktutan, and Gerard Canal. 2024. When Do People Want an Explanation from a Robot?. In Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction. 752–761.
[50]
Eric-Jan Wagenmakers, Jonathon Love, Maarten Marsman, Tahira Jamil, Alexander Ly, Josine Verhagen, Ravi Selker, Quentin F Gronau, Damian Dropmann, Bruno Boutin, 2018. Bayesian inference for psychology. Part II: Example applications with JASP. Psychonomic bulletin & review 25 (2018), 58–76.
[51]
Lei Wang, Chen Ma, Xueyang Feng, Zeyu Zhang, Hao Yang, Jingsen Zhang, Zhiyuan Chen, Jiakai Tang, Xu Chen, Yankai Lin, 2024. A survey on large language model based autonomous agents. Frontiers of Computer Science 18, 6 (2024), 1–26.
[52]
Yiqi Wang, Wentao Chen, Xiaotian Han, Xudong Lin, Haiteng Zhao, Yongfei Liu, Bohan Zhai, Jianbo Yuan, Quanzeng You, and Hongxia Yang. 2024. Exploring the reasoning abilities of multimodal large language models (mllms): A comprehensive survey on emerging trends in multimodal reasoning. arXiv preprint arXiv:2401.06805 (2024).
[53]
Alexander Wei, Nika Haghtalab, and Jacob Steinhardt. 2024. Jailbroken: How does llm safety training fail?Advances in Neural Information Processing Systems 36 (2024).
[54]
Ruchen Wen. 2021. Toward Hybrid Relational-Normative Models of Robot Cognition. In Companion of the 2021 ACM/IEEE International Conference on Human-Robot Interaction. 568–570.
[55]
Ruchen Wen, Zhao Han, and Tom Williams. 2022. Teacher, Teammate, Subordinate, Friend: Generating Norm Violation Responses Grounded in Role-based Relational Norms. In Proceedings of the 17th ACM/IEEE International Conference on Human-Robot Interaction (HRI). 24.8% acceptance rate.
[56]
Ruchen Wen, Ryan Blake Jackson, Tom Williams, and Qin Zhu. 2019. Towards a role ethics approach to command rejection. In HRI Workshop on the Dark Side of Human-Robot Interaction.
[57]
Ruchen Wen, Boyoung Kim, Elizabeth Phillips, Qin Zhu, and Tom Williams. 2021. Comparing Strategies for Robot Communication of Role-Grounded Moral Norms. In Companion of the 2021 ACM/IEEE International Conference on Human-Robot Interaction. 323–327.
[58]
Ruchen Wen, Boyoung Kim, Elizabeth Phillips, Qin Zhu, and Tom Williams. 2022. Comparing Norm-Based and Role-Based Strategies for Robot Communication of Role-Grounded Moral Norms. ACM Transactions on Human-Robot Interaction (T-HRI) (2022).
[59]
Ruchen Wen, Boyoung Kim, Elizabeth Phillips, Qin Zhu, and Tom Williams. 2023. On Further Reflection... Moral Reflections Enhance Robotic Moral Persuasive Capability. In International Conference on Persuasive Technology. Springer, 290–304.
[60]
Tom Williams, Ryan Blake Jackson, and Jane Lockshin. 2018. A Bayesian Analysis of Moral Norm Malleability during Clarification Dialogues. In CogSci.
[61]
Tom Williams, Cynthia Matuszek, Kristiina Jokinen, Raj Korpan, James Pustejovsky, and Brian Scassellati. 2023. Voice in the Machine: Ethical Considerations for Language-Capable Robots. Commun. ACM 66, 8 (2023), 20–23. https://doi.org/10.1145/3604632
[62]
Tom Williams, Cynthia Matuszek, Ross Mead, and Nick Depalma. 2024. Scarecrows in Oz: The Use of Large Language Models in HRI., 11 pages.
[63]
Tom Williams, Qin Zhu, Ruchen Wen, and Ewart J de Visser. 2020. The Confucian Matador: three defenses against the mechanical bull. In Companion of the 2020 ACM/IEEE International Conference on Human-Robot Interaction.
[64]
Katie Winkle, Gaspar Isaac Melsión, Donald McMillan, and Iolanda Leite. 2021. Boosting Robot Credibility and Challenging Gender Norms in Responding to Abusive Behaviour: A Case for Feminist Robots. In Companion of the 2021 ACM/IEEE International Conference on Human-Robot Interaction. 29–37.
[65]
Jingfeng Yang, Hongye Jin, Ruixiang Tang, Xiaotian Han, Qizhang Feng, Haoming Jiang, Shaochen Zhong, Bing Yin, and Xia Hu. 2024. Harnessing the power of llms in practice: A survey on chatgpt and beyond. ACM Transactions on Knowledge Discovery from Data 18, 6 (2024), 1–32.
[66]
Ziyi Yang, Shreyas S Raman, Ankit Shah, and Stefanie Tellex. 2023. Plug in the Safety Chip: Enforcing Temporal Constraints for LLM Agents. (2023).
[67]
Lixiao Zhu and Thomas Williams. 2020. Effects of proactive explanations by robots on human-robot trust. In Social Robotics: 12th International Conference, ICSR 2020, Golden, CO, USA, November 14–18, 2020, Proceedings 12. Springer, 85–95.
[68]
Qin Zhu, Tom Williams, and Ruchen Wen. 2021. Role-based Morality, Ethical Pluralism, and Morally Capable Robots. Journal of Contemporary Eastern Asia 20, 1 (2021), 134–150.

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
HAI '24: Proceedings of the 12th International Conference on Human-Agent Interaction
November 2024
502 pages
ISBN:9798400711787
DOI:10.1145/3687272
This work is licensed under a Creative Commons Attribution-NoDerivatives International 4.0 License.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 24 November 2024

Check for updates

Author Tags

  1. command rejection
  2. large language models in HRI
  3. moral communication
  4. moral reasoning
  5. robot explanation

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Funding Sources

Conference

HAI '24
Sponsor:
HAI '24: International Conference on Human-Agent Interaction
November 24 - 27, 2024
Swansea, United Kingdom

Acceptance Rates

Overall Acceptance Rate 121 of 404 submissions, 30%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 82
    Total Downloads
  • Downloads (Last 12 months)82
  • Downloads (Last 6 weeks)39
Reflects downloads up to 20 Jan 2025

Other Metrics

Citations

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media