Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 24 tok/s Pro
GPT-5 High 25 tok/s Pro
GPT-4o 113 tok/s Pro
Kimi K2 216 tok/s Pro
GPT OSS 120B 428 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Executable Knowledge Graphs for Replicating AI Research (2510.17795v1)

Published 20 Oct 2025 in cs.CL, cs.AI, cs.LG, cs.MA, and cs.SE

Abstract: Replicating AI research is a crucial yet challenging task for LLM agents. Existing approaches often struggle to generate executable code, primarily due to insufficient background knowledge and the limitations of retrieval-augmented generation (RAG) methods, which fail to capture latent technical details hidden in referenced papers. Furthermore, previous approaches tend to overlook valuable implementation-level code signals and lack structured knowledge representations that support multi-granular retrieval and reuse. To overcome these challenges, we propose Executable Knowledge Graphs (xKG), a modular and pluggable knowledge base that automatically integrates technical insights, code snippets, and domain-specific knowledge extracted from scientific literature. When integrated into three agent frameworks with two different LLMs, xKG shows substantial performance gains (10.9% with o3-mini) on PaperBench, demonstrating its effectiveness as a general and extensible solution for automated AI research replication. Code will released at https://github.com/zjunlp/xKG.

Summary

  • The paper proposes Executable Knowledge Graphs (xKG) that integrate research literature with runnable code to tackle replication challenges in AI.
  • It details a modular, hierarchical graph construction method that curates literature and extracts techniques to generate executable code modules.
  • Experimental results show performance gains, with up to a 10.90% improvement when paired with state-of-the-art agent frameworks.

Overview of Executable Knowledge Graphs for Replicating AI Research

This paper addresses the replication issue in AI research by introducing Executable Knowledge Graphs ({xKG}). The authors propose a new approach to facilitate the reproduction of AI research by creating a structured knowledge base that combines textual insights with executable code from scientific literature. This approach aims to address existing challenges such as insufficient code execution capabilities and limited retrieval methods that overlook practical implementation signals.

Introduction

AI research generates numerous publications annually, but replicating these studies remains a challenge due to incomplete code repositories and scattered background information. Existing methods rely heavily on retrieval-augmented generation (RAG), which often lacks the insight needed for effective implementation. Executable Knowledge Graphs are proposed as a solution, offering a structured knowledge representation that integrates detailed code snippets and technical insights extracted from literature.

Methodology

Design of {xKG}

The Executable Knowledge Graph ({xKG}) is a modular, hierarchical graph designed to facilitate AI research replication. It comprises nodes representing papers, techniques, and executable code modules. Nodes are interlinked via structural and implementation edges, encapsulating both theoretical concepts and their corresponding runnable code. Figure 1

Figure 1: The structure of {xKG}, integrating papers, techniques, and executable components.

Graph Construction

The {xKG} is constructed using a multi-step process:

  1. Corpus Curation: Automating the identification of relevant literature and corresponding code repositories using o4-mini.
  2. Technique Extraction: Extracting academic concepts from papers and linking them to executable components through RAG methods.
  3. Code Modularization: Synthesizing code snippets with techniques to produce executable modules, checked for executability.

The final {xKG} comprises 42 curated papers encompassing 591,145 tokens, enabling scalable knowledge integration across various research domains.

Experimental Results

Evaluation

The efficacy of {xKG} is evaluated using the PaperBench Code-Dev benchmark integrated within multiple agent frameworks. The notable gains are demonstrated, with improvements of up to 10.90% when integrated with PaperCoder and o3-mini.

Performance Analysis

A detailed analysis reveals {xKG}'s benefits across multiple facets:

  • Reusability of Structured Knowledge: As shown in ablation studies, the absence of code nodes results in significant performance drops, highlighting the value of executable knowledge in AI research replication.
  • Paper-Specific Performance: The impact of {xKG} is heavily influenced by the nature of the target paper. Analytical papers show more substantial gains due to their component-based structures that align well with {xKG}. Figure 2

    Figure 2: Performance impact across different papers, indicating the value of high-quality code retrieval.

Case Study

A deep dive into the MU-DPO paper showcases how {xKG} augments an agent's ability to plan and implement complex research tasks more effectively by offering modular, verified code components. Figure 3

Figure 3: MU-DPO case paper showing enhanced implementation through {xKG} integration.

Conclusion

Executable Knowledge Graphs offer a robust framework for replicating AI research by bridging the gap between theoretical concepts and executable implementations. The modularity and reusability inherent to {xKG} make it a valuable tool for enhancing AI research replication. This work outlines the potential for scaling knowledge in AI, reducing noise in current retrieval methods, and providing streamlined support for emerging research domains. Future work will focus on expanding {xKG} to broader research areas and refining its applicability across diverse scientific inquiries.

Dice Question Streamline Icon: https://streamlinehq.com
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Github Logo Streamline Icon: https://streamlinehq.com