Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 148 tok/s
Gemini 2.5 Pro 48 tok/s Pro
GPT-5 Medium 34 tok/s Pro
GPT-5 High 40 tok/s Pro
GPT-4o 101 tok/s Pro
Kimi K2 183 tok/s Pro
GPT OSS 120B 443 tok/s Pro
Claude Sonnet 4.5 35 tok/s Pro
2000 character limit reached

Agentic AutoSurvey: Let LLMs Survey LLMs (2509.18661v1)

Published 23 Sep 2025 in cs.IR, cs.CL, and cs.HC

Abstract: The exponential growth of scientific literature poses unprecedented challenges for researchers attempting to synthesize knowledge across rapidly evolving fields. We present \textbf{Agentic AutoSurvey}, a multi-agent framework for automated survey generation that addresses fundamental limitations in existing approaches. Our system employs four specialized agents (Paper Search Specialist, Topic Mining & Clustering, Academic Survey Writer, and Quality Evaluator) working in concert to generate comprehensive literature surveys with superior synthesis quality. Through experiments on six representative LLM research topics from COLM 2024 categories, we demonstrate that our multi-agent approach achieves significant improvements over existing baselines, scoring 8.18/10 compared to AutoSurvey's 4.77/10. The multi-agent architecture processes 75--443 papers per topic (847 total across six topics) while targeting high citation coverage (often $\geq$80\% on 75--100-paper sets; lower on very large sets such as RLHF) through specialized agent orchestration. Our 12-dimension evaluation captures organization, synthesis integration, and critical analysis beyond basic metrics. These findings demonstrate that multi-agent architectures represent a meaningful advancement for automated literature survey generation in rapidly evolving scientific domains.

Summary

  • The paper introduces a multi-agent system that automates literature surveys with significant improvements, achieving a performance score of 8.18/10.
  • The methodology employs advanced techniques like sentence-transformer embeddings and K-means clustering to organize and synthesize thematic clusters.
  • The system’s 12-dimensional evaluation framework underscores its potential to deliver comprehensive and efficient literature synthesis in rapidly evolving fields.

Agentic AutoSurvey: Let LLMs Survey LLMs

Agentic AutoSurvey introduces a multi-agent system designed to automate the generation of literature surveys in rapidly evolving scientific domains, particularly focusing on LLMs. By leveraging specialized agents, including Paper Search Specialist, Topic Mining {content} Clustering, Academic Survey Writer, and Quality Evaluator, this framework aims to overcome limitations of prior approaches such as AutoSurvey. The system demonstrates improvements in synthesis quality, citation coverage, and evaluation metrics, achieving a performance score of 8.18/10 against AutoSurvey's 4.77/10.

System Architecture and Methodology

Multi-Agent Framework

Agentic AutoSurvey employs a multi-agent architecture to efficiently process and synthesize literature. Each agent is tailored for specific tasks:

  • Paper Search Specialist: Executes advanced queries across multiple sources, ensuring high citation coverage and comprehensive retrieval.
  • Topic Mining {content} Clustering: Utilizes sentence-transformer embeddings and K-means clustering to organize papers into coherent themes.
  • Academic Survey Writer: Focuses on integrating findings across clusters for a synthesized narrative.
  • Quality Evaluator: Performs a 12-dimensional assessment to guarantee academic rigor and synthesis quality. Figure 1

    Figure 1: Enhanced Agentic Framework Architecture.

Clustering Analysis

The Topic Mining {content} Clustering Agent effectively organizes literature into thematic clusters based on semantic embeddings. Optimal clustering is achieved through silhouette score maximization, and inter-cluster relationship identification further enriches the synthesis process. Figure 2

Figure 2: Cluster distribution for (a) LLM Agents and (b) Synthetic Data Generation topics, showing the thematic organization discovered by the Topic Mining Agent.

Technical Implementation

Paper Retrieval and Clustering

Embedding generation is performed using the all-MiniLM-L6-v2 model, mapping text to a dense vector space. The Paper Search Specialist generates diverse queries, maximizing coverage through integration with APIs like Semantic Scholar and arXiv. Subsequent clustering applies advanced techniques to ensure optimal thematic organization.

Survey Generation

The Academic Survey Writer generates content that moves beyond simple enumeration, emphasizing cross-cluster integration and identifying methodological patterns and trends. Emphasis is placed on producing publication-quality narrative that achieves substantial citation coverage.

Evaluation Framework

The Quality Evaluator agent implements a nuanced 12-dimensional assessment, including dimensions like synthesis quality, citation coverage, and critical analysis, providing a comprehensive evaluation of the survey's quality.

Experimental Evaluation

Performance Metrics

Agentic AutoSurvey achieves substantial improvements across all dimensions compared to AutoSurvey. Notably, the system excels in core quality dimensions such as synthesis quality and citation coverage due to specialized agent orchestration. Figure 3

Figure 3: Distribution of papers retrieved and clusters generated across the six processed topics.

Clustering Quality

The clustering exhibits significant thematic separation, capturing diverse research threads within each topic, contributing to the high synthesis quality scores. Figure 4

Figure 4: Cluster size distribution heatmap showing the number of papers in each cluster across four survey topics.

Discussion and Limitations

While Agentic AutoSurvey represents a significant leap forward in automated synthesis, limitations remain, such as handling large corpora with reduced citation coverage. Future enhancements may involve hierarchical writing strategies to incorporate a broader range of literature.

Conclusion

Agentic AutoSurvey advances automated survey generation through its multi-agent architecture, achieving superior synthesis quality and citation coverage. This approach offers significant potential benefits to researchers in rapidly evolving scientific domains, facilitating comprehensive understanding and integration of vast literatures. Figure 5

Figure 5: Clustering statistics across survey topics, indicating the clustering quality across evaluated dimensions.

In summary, Agentic AutoSurvey represents a meaningful advancement in AI-driven survey generation, providing researchers with a powerful tool to synthesize knowledge efficiently in the ever-growing landscape of scientific literature. Future work could focus on overcoming scalability challenges and optimizing agent collaboration for enhanced synthesis quality.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.