Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MedAgents: Large Language Models as Collaborators for Zero-shot Medical Reasoning (2311.10537v4)

Published 16 Nov 2023 in cs.CL and cs.AI
MedAgents: Large Language Models as Collaborators for Zero-shot Medical Reasoning

Abstract: LLMs, despite their remarkable progress across various general domains, encounter significant barriers in medicine and healthcare. This field faces unique challenges such as domain-specific terminologies and reasoning over specialized knowledge. To address these issues, we propose MedAgents, a novel multi-disciplinary collaboration framework for the medical domain. MedAgents leverages LLM-based agents in a role-playing setting that participate in a collaborative multi-round discussion, thereby enhancing LLM proficiency and reasoning capabilities. This training-free framework encompasses five critical steps: gathering domain experts, proposing individual analyses, summarising these analyses into a report, iterating over discussions until a consensus is reached, and ultimately making a decision. Our work focuses on the zero-shot setting, which is applicable in real-world scenarios. Experimental results on nine datasets (MedQA, MedMCQA, PubMedQA, and six subtasks from MMLU) establish that our proposed MedAgents framework excels at mining and harnessing the medical expertise within LLMs, as well as extending its reasoning abilities. Our code can be found at https://github.com/gersteinlab/MedAgents.

MedAgents: LLMs as Collaborators for Zero-shot Medical Reasoning

The exploration of LLMs in specialized domains, such as the medical field, presents inherent challenges due to the domain-specific terminologies and the requirement for complex reasoning. The paper "MedAgents: LLMs as Collaborators for Zero-shot Medical Reasoning" introduces an innovative approach, termed as Multi-disciplinary Collaboration (MC) framework, which leverages LLM-based agents in a role-playing scenario to augment their reasoning capabilities without the need for additional training. This structure is designed to mimic a collaborative, multi-round discussion among different domain experts, improving the capacity of LLMs in medical applications, especially for zero-shot scenarios.

The MC framework is articulated through five critical stages:

  1. Expert Gathering: This involves assembling domain experts to provide multiple perspectives on the medical question at hand, enhancing the depth and breadth of analysis.
  2. Analysis Proposition: Expert agents propose individual analyses derived from their specialized knowledge, representing nuanced perspectives on the medical query.
  3. Report Summarization: The individual analyses are consolidated into a cohesive report, which serves as a basis for further discussion.
  4. Collaborative Consultation: Multiple rounds of discussion among agents refine the report until a consensus is reached, ensuring each expert's perspective is incorporated and vetted.
  5. Decision Making: The final, consensus-derived report forms the basis for answering the initial medical question.

Empirical evaluation on nine datasets, including MedQA, MedMCQA, and PubMedQA, demonstrates that this MC framework excels over existing zero-shot methods like chain-of-thought (CoT) and self-consistency (SC) prompting approaches. Notably, the MC framework achieves superior accuracy in the zero-shot setting compared to 5-shot baselines, showcasing its enhanced ability to mine and apply intrinsic medical knowledge embedded within LLMs.

In examining the configuration of the MC framework, a pivotal finding is the influence of role-playing on LLM performance. Each agent's domain-specific knowledge becomes pivotal in uncovering and debating the diverse aspects of medical queries, which is a sophisticated alternative to simple prompting. The MC framework addresses a significant drawback of LLMs: the typical inadequacy in handling tasks that demand specialized knowledge and reasoning beyond their general training.

The paper also elaborates on error analysis, categorizing common failure types into lack of domain knowledge, mis-retrieval of domain expertise, consistency issues, and chain-of-thought errors. It proposes refinements targeting these errors as a pathway to enhance model reliability in future iterations.

Overall, this paper posits that leveraging multidisciplinary collaboration among LLM-based agents can effectively uncover domain-specific knowledge, floodlighting paths for their application in real-world, training-agnostic medical reasoning tasks. This framework marks a significant step forward in optimizing LLMs for healthcare applications without necessitating vast specialized training, hinting at future research directions that could involve integrating hybrid models or optimized prompting strategies to further bolster the intrinsic utility and precision of LLMs in specialized domains.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Xiangru Tang (62 papers)
  2. Anni Zou (6 papers)
  3. Zhuosheng Zhang (125 papers)
  4. Yilun Zhao (59 papers)
  5. Xingyao Zhang (17 papers)
  6. Arman Cohan (121 papers)
  7. Mark Gerstein (25 papers)
  8. Ziming Li (44 papers)
Citations (96)
Github Logo Streamline Icon: https://streamlinehq.com

GitHub