Papers
Topics
Authors
Recent
2000 character limit reached

Eigen-1: Adaptive Multi-Agent Refinement with Monitor-Based RAG for Scientific Reasoning (2509.21193v1)

Published 25 Sep 2025 in cs.CL and cs.AI

Abstract: LLMs have recently shown strong progress on scientific reasoning, yet two major bottlenecks remain. First, explicit retrieval fragments reasoning, imposing a hidden "tool tax" of extra tokens and steps. Second, multi-agent pipelines often dilute strong solutions by averaging across all candidates. We address these challenges with a unified framework that combines implicit retrieval and structured collaboration. At its foundation, a Monitor-based retrieval module operates at the token level, integrating external knowledge with minimal disruption to reasoning. On top of this substrate, Hierarchical Solution Refinement (HSR) iteratively designates each candidate as an anchor to be repaired by its peers, while Quality-Aware Iterative Reasoning (QAIR) adapts refinement to solution quality. On Humanity's Last Exam (HLE) Bio/Chem Gold, our framework achieves 48.3\% accuracy -- the highest reported to date, surpassing the strongest agent baseline by 13.4 points and leading frontier LLMs by up to 18.1 points, while simultaneously reducing token usage by 53.5\% and agent steps by 43.7\%. Results on SuperGPQA and TRQA confirm robustness across domains. Error analysis shows that reasoning failures and knowledge gaps co-occur in over 85\% of cases, while diversity analysis reveals a clear dichotomy: retrieval tasks benefit from solution variety, whereas reasoning tasks favor consensus. Together, these findings demonstrate how implicit augmentation and structured refinement overcome the inefficiencies of explicit tool use and uniform aggregation. Code is available at: https://github.com/tangxiangru/Eigen-1.

Summary

  • The paper introduces a monitor-based RAG framework that dynamically refines candidate solutions during scientific reasoning.
  • The framework achieved 48.3% accuracy on HLE Bio/Chem Gold, outperforming baselines by up to +18.1 points and reducing retrieval overhead by over 50%.
  • It integrates Hierarchical Solution Refinement and Quality-Aware Iterative Reasoning to iteratively enhance logic, accuracy, and computational efficiency.

Adaptive Multi-Agent Refinement for Scientific Reasoning

Introduction

The paper "Eigen-1: Adaptive Multi-Agent Refinement with Monitor-Based RAG for Scientific Reasoning" presents a structured multi-agent framework that leverages Monitor-based Retrieval-Augmented Generation (RAG) to address challenges faced by LLMs in scientific reasoning. The primary issues identified by the authors are the fragmentation of reasoning through explicit retrieval and the inefficiencies introduced by uniform candidate aggregation in multi-agent setups.

Monitor-Based RAG

The proposed framework, Eigen-1, incorporates Monitor-based RAG, which functions seamlessly at the token level to detect reasoning insufficiencies. This setup implicitly augments the reasoning process by generating contextual queries and injecting retrieved evidence with minimal disruption.

  • Implicit Retrieval: Unlike traditional RAG paradigms that pause the reasoning process for external retrieval, the Monitor-based RAG operates continuously, preserving the flow.
  • Experimental Evidence: It achieves a 48.3% accuracy on the Humanity's Last Exam (HLE) Bio/Chem Gold, outperforming the best LLM by considerable margins and reducing retrieval overhead by over 50%. Figure 1

    Figure 1: HLE Bio/Chem Gold overall accuracy. On the 149-problem split, our system attains 48.3% accuracy, exceeding the strongest agent baseline by +13.4 points and leading frontier LLMs by up to +18.1 points.

Hierarchical Solution Refinement

The Hierarchical Solution Refinement (HSR) framework rotates each candidate solution as an anchor to be refined by its peers, facilitating structured, cross-solution improvement.

  • Anchor-Reference Paradigm: Each candidate solution is treated as an anchor while peers provide reference-based improvements, preventing premature consensus and enhancing the final solution.
  • Mechanism: Targeted improvements include logic completion and numerical correction, among others. HSR moves beyond simple averaging, allowing solutions to converge based on quality.

Quality-Aware Iterative Reasoning

Quality-Aware Iterative Reasoning (QAIR) evaluates intermediate solution quality and iteratively refines candidates using quality-driven metrics.

  • Evaluation and Iteration: Using a scoring rubric that assesses logic, answer correctness, and explanation, QAIR ensures that only substantive quality improvements are pursued.
  • Adaptive Workflow: The adaptive workflow allows it to stop when solutions converge on optimal quality, ensuring computational efficiency. Figure 2

    Figure 2: Framework overview. The Monitor detects insufficiency, and integrated modules ensure iterative solution refinement with minimal reasoning disruption.

Experimental Results

  • Data and Metrics: Experiments were conducted on diverse benchmarks including SuperGPQA and TRQA, confirming robustness across domains.
  • Key Findings: The framework reduced the "tool tax" from retrieval by a substantial margin and significantly improved reasoning efficiency and accuracy.
  • Error Analysis: It highlighted two major overlapping failure modes—reasoning process errors and knowledge gaps—each addressed by the framework's components.

Implications and Future Work

The integration of implicit augmentation and structured refinement paves the way for optimized reasoning frameworks capable of more natural knowledge integration, critical for domains requiring complex problem-solving.

  • Future Directions: While focused on scientific reasoning, these principles could generalize to other high-stakes areas involving multistep logical inference and dynamic knowledge needs.

Conclusion

Eigen-1's innovations in the integration of external knowledge reveal pathways to enhance reasoning performance and efficiency through structured agent frameworks and adaptive refinement processes. Further exploration will aim to extend this model's applicability to broader domains in AI-driven scientific inquiry.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Github Logo Streamline Icon: https://streamlinehq.com
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.