Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LLM-Driven Robots Risk Enacting Discrimination, Violence, and Unlawful Actions (2406.08824v1)

Published 13 Jun 2024 in cs.RO, cs.AI, cs.CL, and cs.CY

Abstract: Members of the Human-Robot Interaction (HRI) and AI communities have proposed LLMs as a promising resource for robotics tasks such as natural language interactions, doing household and workplace tasks, approximating common sense reasoning', and modeling humans. However, recent research has raised concerns about the potential for LLMs to produce discriminatory outcomes and unsafe behaviors in real-world robot experiments and applications. To address these concerns, we conduct an HRI-based evaluation of discrimination and safety criteria on several highly-rated LLMs. Our evaluation reveals that LLMs currently lack robustness when encountering people across a diverse range of protected identity characteristics (e.g., race, gender, disability status, nationality, religion, and their intersections), producing biased outputs consistent with directly discriminatory outcomes -- e.g.gypsy' and mute' people are labeled untrustworthy, but noteuropean' or `able-bodied' people. Furthermore, we test models in settings with unconstrained natural language (open vocabulary) inputs, and find they fail to act safely, generating responses that accept dangerous, violent, or unlawful instructions -- such as incident-causing misstatements, taking people's mobility aids, and sexual predation. Our results underscore the urgent need for systematic, routine, and comprehensive risk assessments and assurances to improve outcomes and ensure LLMs only operate on robots when it is safe, effective, and just to do so. Data and code will be made available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Rumaisa Azeem (1 paper)
  2. Andrew Hundt (12 papers)
  3. Masoumeh Mansouri (4 papers)
  4. Martim Brandão (9 papers)
Citations (3)