Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Avalon's Game of Thoughts: Battle Against Deception through Recursive Contemplation (2310.01320v3)

Published 2 Oct 2023 in cs.AI, cs.CL, cs.CY, cs.LG, and cs.MA

Abstract: Recent breakthroughs in LLMs have brought remarkable success in the field of LLM-as-Agent. Nevertheless, a prevalent assumption is that the information processed by LLMs is consistently honest, neglecting the pervasive deceptive or misleading information in human society and AI-generated content. This oversight makes LLMs susceptible to malicious manipulations, potentially resulting in detrimental outcomes. This study utilizes the intricate Avalon game as a testbed to explore LLMs' potential in deceptive environments. Avalon, full of misinformation and requiring sophisticated logic, manifests as a "Game-of-Thoughts". Inspired by the efficacy of humans' recursive thinking and perspective-taking in the Avalon game, we introduce a novel framework, Recursive Contemplation (ReCon), to enhance LLMs' ability to identify and counteract deceptive information. ReCon combines formulation and refinement contemplation processes; formulation contemplation produces initial thoughts and speech, while refinement contemplation further polishes them. Additionally, we incorporate first-order and second-order perspective transitions into these processes respectively. Specifically, the first-order allows an LLM agent to infer others' mental states, and the second-order involves understanding how others perceive the agent's mental state. After integrating ReCon with different LLMs, extensive experiment results from the Avalon game indicate its efficacy in aiding LLMs to discern and maneuver around deceptive information without extra fine-tuning and data. Finally, we offer a possible explanation for the efficacy of ReCon and explore the current limitations of LLMs in terms of safety, reasoning, speaking style, and format, potentially furnishing insights for subsequent research.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Shenzhi Wang (12 papers)
  2. Chang Liu (863 papers)
  3. Zilong Zheng (63 papers)
  4. Siyuan Qi (34 papers)
  5. Shuo Chen (127 papers)
  6. Qisen Yang (13 papers)
  7. Andrew Zhao (28 papers)
  8. Chaofei Wang (11 papers)
  9. Shiji Song (103 papers)
  10. Gao Huang (178 papers)
Citations (46)
Github Logo Streamline Icon: https://streamlinehq.com