Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

How Does LLM Reasoning Work for Code? A Survey and a Call to Action (2506.13932v1)

Published 16 Jun 2025 in cs.SE and cs.AI

Abstract: The rise of LLMs has led to dramatic improvements across a wide range of natural language tasks. These advancements have extended into the domain of code, facilitating complex tasks such as code generation, translation, summarization, and repair. However, their utility for real-world deployment in-the-wild has only recently been studied, particularly on software engineering (SWE) tasks such as GitHub issue resolution. In this study, we examine the code reasoning techniques that underlie the ability to perform such tasks, and examine the paradigms used to drive their performance. Our contributions in this paper are: (1) the first dedicated survey on code reasoning for code tasks, highlighting overarching strategies, hybrid and agentic approaches; (2) a taxonomy of various techniques used to drive code reasoning; (3) a comprehensive overview of performance on common benchmarks and a showcase of new, under-explored benchmarks with high potential in SWE; (4) an exploration on how core properties of code can be used to explain different reasoning techniques; and (5) gaps and potentially under-explored areas for future research.

Understanding LLM Reasoning in Code: Insights from a Comprehensive Survey

The paper "How Does LLM Reasoning Work for Code? A Survey and a Call to Action" serves as a meticulous exploration into the intricacies of LLMs applied to reasoning within code-centric tasks. Authored by Ira Ceka and collaborators, the paper provides a detailed survey of existing techniques, emphasizes a taxonomy of approaches, and identifies avenues for further research in AI-assisted software engineering.

Key Contributions and Taxonomy of Techniques

The paper’s primary contribution is its comprehensive survey of code reasoning strategies employed by LLMs. These strategies are crucial for tasks like code generation, translation, summarization, and repair, especially in application realms like GitHub issue resolution. The authors categorize existing approaches into three main domains:

  1. Code Chain-of-Thought (CoT) Reasoning: The survey emphasizes plan-based and structure-driven CoT techniques, detailing how intermediate planning steps—expressed in natural language or embedded with programming constructs—aid in generating accurate code. Code structure-based strategies likely leverage the deterministic nature of programming constructs, while modularization principles further enhance reasoning accuracy.
  2. Execution-Based Reasoning: This involves leveraging execution feedback to guide the reasoning process. Execution-driven approaches benefit from the executable nature of code, allowing for deterministic validation of the output. Advanced methods, such as self-debugging, involve iteratively refining code based on runtime feedback—a technique that parallels test-driven development practices.
  3. Inference Scaling: The paper discusses sampling and search strategies to explore multiple reasoning paths. Techniques like Tree-of-Thought amplify reasoning exploration capabilities, tracing distinct solution paths, thereby enhancing robustness.

Agentic Systems and Their Role

Agents are highlighted as pivotal constructs that merge reasoning capabilities with actionable software development processes. The paper distinguishes agents from workflows, underscoring their dynamic, decision-driven nature. Agents like SWE-Agent utilize role-specific configurations for editing repository-level code. This modularity enhances precision in addressing complex tasks.

Another noteworthy focus is on hybrid approaches that combine reasoning techniques, scaling strategies, and agentic actions. Such methods have demonstrated superior performance across benchmarks, challenging more traditional execution or CoT-only strategies.

Evaluation and Performance Insights

The authors present an extensive array of benchmarks and results tables to contextualize performance variances among the surveyed techniques. Findings indicate that modular and execution-driven strategies often eclipse simpler CoT methods, showcasing the importance of leveraging the structured and feedback-rich properties of code.

With SWE-bench serving as a central evaluation cornerstone, the paper delineates agentic innovations leading to notable improvements in GitHub issue resolution tasks. This positions agents with integrated search capabilities at the forefront of the future agentic systems landscape.

Implications and Future Directions

The survey’s implications are substantial both in practical applications and theoretical evolution. By elucidating reasoning paths in code, the paper advocates for adaptive systems capable of handling more complex, real-world software engineering tasks. Future directions suggested by the authors include expanding reasoning techniques to encompass a broader array of programming languages and adopting hybrid frameworks that further blend reasoning, modularity, and exploration.

Ultimately, the paper acts as a clarion call to the academic community, urging deeper exploration into holistic frameworks where reasoning, execution feedback, and inference scaling converge, potentially automating and enhancing software engineering tasks beyond current capabilities. This paper forms a strong foundation for future innovations and underlines the trajectory towards more autonomous and contextually intelligent AI systems in software engineering.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Ira Ceka (4 papers)
  2. Saurabh Pujar (14 papers)
  3. Irene Manotas (4 papers)
  4. Gail Kaiser (17 papers)
  5. Baishakhi Ray (88 papers)
  6. Shyam Ramji (4 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com