Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

AgentClinic: a multimodal agent benchmark to evaluate AI in simulated clinical environments (2405.07960v4)

Published 13 May 2024 in cs.HC and cs.CL
AgentClinic: a multimodal agent benchmark to evaluate AI in simulated clinical environments

Abstract: Evaluating LLMs (LLM) in clinical scenarios is crucial to assessing their potential clinical utility. Existing benchmarks rely heavily on static question-answering, which does not accurately depict the complex, sequential nature of clinical decision-making. Here, we introduce AgentClinic, a multimodal agent benchmark for evaluating LLMs in simulated clinical environments that include patient interactions, multimodal data collection under incomplete information, and the usage of various tools, resulting in an in-depth evaluation across nine medical specialties and seven languages. We find that solving MedQA problems in the sequential decision-making format of AgentClinic is considerably more challenging, resulting in diagnostic accuracies that can drop to below a tenth of the original accuracy. Overall, we observe that agents sourced from Claude-3.5 outperform other LLM backbones in most settings. Nevertheless, we see stark differences in the LLMs' ability to make use of tools, such as experiential learning, adaptive retrieval, and reflection cycles. Strikingly, Llama-3 shows up to 92% relative improvements with the notebook tool that allows for writing and editing notes that persist across cases. To further scrutinize our clinical simulations, we leverage real-world electronic health records, perform a clinical reader study, perturb agents with biases, and explore novel patient-centric metrics that this interactive environment firstly enables.

Exploring the AgentClinic Benchmark: Advancing AI in Simulated Clinical Environments

Introduction to AgentClinic

AgentClinic is an open-source multimodal agent benchmark designed to simulate real-world clinical environments using language agents. This benchmark platform introduces unique features like multimodal interactions, bias incorporation, and complex agent roles to create a comprehensive and challenging environment for testing LLMs within a medical context.

Insights into Language Agents

The Roles of Different Agents

AgentClinic utilizes four types of language agents to drive its simulated medical platform:

  • Patient Agent: This agent mimics a real patient by presenting symptoms and medical history without knowing the actual diagnosis.
  • Doctor Agent: Acts as the primary evaluation target, diagnosing the patient based on the provided symptoms, history, and results from diagnostic tests requested from the Measurement Agent.
  • Measurement Agent: Delivers test results like blood pressure and EKG readings upon the Doctor Agent’s request, simulating real-life medical test scenarios.
  • Moderator Agent: This agent assesses the accuracy of the diagnosis made by the Doctor Agent, parsing through potentially unstructured response data.

This diversity in agent roles allows AgentClinic to mimic the flow of real medical consultations more closely than previous benchmarks which mainly relied on static Q&A formats.

Implementing and Assessing Biases

A critical part of the AgentClinic benchmark is its focus on biases known to affect medical diagnostics, like cognitive and implicit biases. These biases are intentionally introduced to both Doctor and Patient Agents to paper their influence on diagnostic accuracy and patient trust. The intriguing findings include:

  • Influence on Diagnostic Accuracy: Biases led to varying impacts on accuracy, with cognitive biases causing a noticeable reduction, especially when affecting the Doctor Agent.
  • Patient Perceptions: Despite minimal impacts on diagnostic accuracy, biased interactions led to significant changes in patient compliance, confidence in the doctor, and willingness to return for follow-up consultations.

Testing and Results

Diagnostic Accuracy Across Models

AgentClinic tested different LLMs including GPT-4, Mixtral-8x7B, and Llama 2 70B-chat. The detailed results illustrated notable variations in diagnostic accuracy, highlighting the varying capabilities of these models within a complex simulated clinical environment.

Dynamic Interactions and Their Impact

The benchmark goes beyond static evaluations to explore the effect of dynamic factors like the number of allowable interaction turns and the choice of patient LLM on diagnostic outcomes. Findings revealed that:

  • Interaction Turns: Both limited and excessive interaction turns adversely affected diagnostic accuracy, underscoring the importance of optimal information exchange in clinical decision-making.
  • Patient LLM: Using the same model for both Doctor and Patient Agents generally resulted in higher diagnostic accuracy, suggesting model-specific communication advantages.

The Multimodal Challenge

AgentClinic introduced multimodal capabilities by incorporating diagnostic imaging into the simulations. Two multimodal LLMs were tested, exploring their ability to integrate visual data with textual patient dialogue for diagnosis. The experiments shed light on the models' understanding and utilization of complex multimodal information but also highlighted the challenges these models face in clinical scenarios requiring integrated data interpretation.

Implications and Future Directions

The development of AgentClinic represents a substantial step forward in creating more realistic and dynamic AI testing environments for medical applications. The introduction of realistic biases, multimodal interactions, and varied agent dynamics promotes a deeper understanding of AI capabilities and limitations in healthcare.

Looking ahead, the framework could benefit from inclusion of additional clinical roles and broader medical conditions. Enhancing the realism and complexity of interactions in future iterations could help bridge the gap between AI capabilities and practical medical needs.

By continuing to evolve and expand benchmarks like AgentClinic, the field can ensure that AI tools are rigorously tested and refined, supporting their role as valuable aides in clinical settings rather than as replacements for human expertise.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Samuel Schmidgall (27 papers)
  2. Rojin Ziaei (4 papers)
  3. Carl Harris (3 papers)
  4. Eduardo Reis (9 papers)
  5. Jeffrey Jopling (6 papers)
  6. Michael Moor (20 papers)
Citations (23)
Github Logo Streamline Icon: https://streamlinehq.com