Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
173 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Advancing Healthcare Automation: Multi-Agent System for Medical Necessity Justification (2404.17977v2)

Published 27 Apr 2024 in cs.AI and cs.MA

Abstract: Prior Authorization delivers safe, appropriate, and cost-effective care that is medically justified with evidence-based guidelines. However, the process often requires labor-intensive manual comparisons between patient medical records and clinical guidelines, that is both repetitive and time-consuming. Recent developments in LLMs have shown potential in addressing complex medical NLP tasks with minimal supervision. This paper explores the application of Multi-Agent System (MAS) that utilize specialized LLM agents to automate Prior Authorization task by breaking them down into simpler and manageable sub-tasks. Our study systematically investigates the effects of various prompting strategies on these agents and benchmarks the performance of different LLMs. We demonstrate that GPT-4 achieves an accuracy of 86.2% in predicting checklist item-level judgments with evidence, and 95.6% in determining overall checklist judgment. Additionally, we explore how these agents can contribute to explainability of steps taken in the process, thereby enhancing trust and transparency in the system.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (8)
  1. Mapping us commercial payers’ coverage policies for medical interventions. The American journal of managed care, 22:e323–e328.
  2. Retrieval-augmented generation for large language models: A survey.
  3. Hey ai, can you solve complex tasks by talking to agents?
  4. Xiaodong Li and Maurice Clerc. 2019. Swarm Intelligence, pages 353–384. Springer International Publishing, Cham.
  5. Is time spent on prior authorizations associated with approval? J Nurse Pract, 19(2):104479. Epub 2022 Nov 10.
  6. Rethinking the role of demonstrations: What makes in-context learning work?
  7. In-Context Retrieval-Augmented Language Models. Transactions of the Association for Computational Linguistics, 11:1316–1331.
  8. Chain-of-thought prompting elicits reasoning in large language models. In Advances in Neural Information Processing Systems, volume 35, pages 24824–24837. Curran Associates, Inc.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com