Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Encouraging Divergent Thinking in Large Language Models through Multi-Agent Debate (2305.19118v4)

Published 30 May 2023 in cs.CL
Encouraging Divergent Thinking in Large Language Models through Multi-Agent Debate

Abstract: Modern LLMs like ChatGPT have shown remarkable performance on general language tasks but still struggle on complex reasoning tasks, which drives the research on cognitive behaviors of LLMs to explore human-like problem-solving strategies. Along this direction, one representative strategy is self-reflection, which asks an LLM to refine the solution with the feedback generated by itself iteratively. However, our study shows that such reflection-style methods suffer from the Degeneration-of-Thought (DoT) problem: once the LLM has established confidence in its solutions, it is unable to generate novel thoughts later through reflection even if its initial stance is incorrect. To address the DoT problem, we propose a Multi-Agent Debate (MAD) framework, in which multiple agents express their arguments in the state of "tit for tat" and a judge manages the debate process to obtain a final solution. Clearly, our MAD framework encourages divergent thinking in LLMs which would be helpful for tasks that require deep levels of contemplation. Experiment results on two challenging datasets, commonsense machine translation and counter-intuitive arithmetic reasoning, demonstrate the effectiveness of our MAD framework. Extensive analyses suggest that the adaptive break of debate and the modest level of "tit for tat" state are required for MAD to obtain good performance. Moreover, we find that LLMs might not be a fair judge if different LLMs are used for agents. Code is available at https://github.com/Skytliang/Multi-Agents-Debate.

Encouraging Divergent Thinking in LLMs Through Multi-Agent Debate

The paper under review explores a novel framework, termed Multi-Agent Debate (MAD), designed to enhance the reasoning capabilities of LLMs by encouraging divergent thinking, thereby addressing the Degeneration-of-Thought (DoT) problem observed in self-reflection methodologies. The DoT issue manifests when LLMs become excessively confident in their initial solutions, inhibiting their ability to generate innovative thoughts, even in the event of errors in their preliminary estimations.

The authors identify the susceptibility of reflection-style approaches, such as self-refinement, to DoT, understanding that through iteration, these methods yield increasingly homogenous and less innovative solutions. To overcome these limitations, the MAD framework leverages a "tit for tat" multi-agent setup, wherein multiple LLMs engage in structured debates, overseen by a judge LLM that adjudicates final solutions across iterative cycles. This setup is proposed as a means of inducing variance in the reasoning chains of LLMs by leveraging disagreement as a productive cognitive force.

The paper evaluates the MAD framework using two challenging datasets: commonsense machine translation and counter-intuitive arithmetic reasoning. The findings demonstrate MAD's capacity to improve performance significantly, with results from the GPT-3.5-Turbo backbone surpassing those of GPT-4 on certain translations tasks. This is a telling result, reflecting the efficacy of MAD in leveraging disagreements to correct biases and rigid thought patterns inherent to LLMs.

Critical findings from this paper highlight that MAD's success is contingent upon two key dynamics: the adaptive termination of debate rounds and maintaining a moderate "tit for tat" state. Whereas forcing debates to continue indefinitely can lead to diminishing returns, an adaptive break strategy seemingly capitalizes on when sufficient divergent thinking has been achieved, thereby enhancing overall efficacy. Additionally, while some level of disagreement fosters innovation, excessive discord without convergence may impede the discovery of correct solutions, suggesting the necessity of a balanced interplay between divergence and agreement.

Moreover, the analysis reveals an inherent bias in the judge's decisions, suggesting that LLMs may not serve as impartial adjudicators if diverse models are tasked as debating agents. This indicates a complex dynamic in leveraging LLMs for task-oriented dispute resolution and suggests an area ripe for future exploration—finding the right balance between diversity and homogeneity in multi-agent tasks.

The practical implications of MAD are worth noting, especially in the spheres of automated translation and complex problem-solving, where broadening the solution space through agent-induced dissent can improve solution accuracy and reliability. Theoretically, MAD represents an intriguing foray into the exploration of artificial cognitive processes, striving to emulate more authentically the disagreement-driven innovation seen in human deliberation.

Future developments may further refine this multi-agent approach, potentially by incorporating more nuanced models of disagreement or evaluating dynamics across more extensive agent ecosystems. Additionally, exploration into the applicability of MAD across various architectures and its integration into real-world AI systems remains a promising direction for ongoing research. Ultimately, the paper contributes a thought-provoking empirical framework, illustrating the impactful role of structured dissent in enriching the cognitive abilities of LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Tian Liang (50 papers)
  2. Zhiwei He (42 papers)
  3. Wenxiang Jiao (44 papers)
  4. Xing Wang (191 papers)
  5. Yan Wang (733 papers)
  6. Rui Wang (996 papers)
  7. Yujiu Yang (155 papers)
  8. Zhaopeng Tu (135 papers)
  9. Shuming Shi (126 papers)
Citations (258)