Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 162 tok/s
Gemini 2.5 Pro 56 tok/s Pro
GPT-5 Medium 38 tok/s Pro
GPT-5 High 35 tok/s Pro
GPT-4o 104 tok/s Pro
Kimi K2 164 tok/s Pro
GPT OSS 120B 426 tok/s Pro
Claude Sonnet 4.5 35 tok/s Pro
2000 character limit reached

Autellix: An Efficient Serving Engine for LLM Agents as General Programs (2502.13965v1)

Published 19 Feb 2025 in cs.LG, cs.AI, and cs.DC

Abstract: LLM applications are evolving beyond simple chatbots into dynamic, general-purpose agentic programs, which scale LLM calls and output tokens to help AI agents reason, explore, and solve complex tasks. However, existing LLM serving systems ignore dependencies between programs and calls, missing significant opportunities for optimization. Our analysis reveals that programs submitted to LLM serving engines experience long cumulative wait times, primarily due to head-of-line blocking at both the individual LLM request and the program. To address this, we introduce Autellix, an LLM serving system that treats programs as first-class citizens to minimize their end-to-end latencies. Autellix intercepts LLM calls submitted by programs, enriching schedulers with program-level context. We propose two scheduling algorithms-for single-threaded and distributed programs-that preempt and prioritize LLM calls based on their programs' previously completed calls. Our evaluation demonstrates that across diverse LLMs and agentic workloads, Autellix improves throughput of programs by 4-15x at the same latency compared to state-of-the-art systems, such as vLLM.

Summary

  • The paper introduces Autellix, treating entire LLM programs as first-class scheduling units to minimize wait times and head-of-line blocking.
  • The system incorporates PLAS and ATLAS metrics for adaptive scheduling, achieving 4-15x throughput improvements over traditional methods.
  • Autellix integrates with existing LLM serving engines to enhance latency reduction and scalability across complex, agentic workloads.

Autellix: An Efficient Serving Engine for LLM Agents as General Programs

Introduction

The paper "Autellix: An Efficient Serving Engine for LLM Agents as General Programs" addresses the increasing complexity and scale of LLM applications, which require efficient serving solutions beyond individual LLM calls. Autellix proposes a novel serving system that considers LLM programs as first-class entities, optimizing end-to-end latencies by minimizing cumulative wait times and enhancing scheduling strategies for both single-threaded and distributed agentic programs.

Problem Statement

Current LLM serving systems often lack program-level context, leading to significant inefficiencies in handling dynamic and complex agentic applications. As highlighted, these systems face challenges like head-of-line blocking, where LLM requests experience long delays due to dependency-induced wait times. Autellix is designed to treat entire programs as fundamental units, thereby addressing these bottlenecks through enriched scheduling algorithms.

Proposed Solution

Autellix introduces a comprehensive framework that incorporates global program-level statistics into the scheduling process:

  • Program-Level Attained Service (PLAS): This metric prioritizes LLM calls based on previously completed program-level service times, effectively reducing both call-level and program-level blocking.
  • Adaptive Thread-Level Attained Service (ATLAS): Specifically for distributed programs, ATLAS optimizes scheduling by estimating critical paths and prioritizing threads accordingly. Figure 1

    Figure 1: \smallAI Agent Infrastructure. Top: Developers and users build and execute agentic programs that orchestrate execution and persist global, cumulative history across agents, tools, and humans. Bottom: LLM serving systems process agents' LLM calls and route calls across one or more LLM engines.

System Architecture

Autellix integrates seamlessly as a layer on top of existing LLM serving engines, offering a stateful API for persistent session management. The architecture is divided into:

  • Front-End: Simplifies the user's interface with Autellix, allowing programmatic LLM call management.
  • Back-End Scheduler: Leverages PLAS and ATLAS for effective request scheduling.
  • Load Balancer: Ensures efficient engine utilization by considering data locality, balancing short and long LLM call assignments accordingly.

Evaluation and Results

The evaluation demonstrates Autellix's significant improvement in throughput and latency metrics across various workloads, outperforming baseline systems like vLLM and MLFQ:

  • Throughput Improvements: Autellix enhances throughput by 4-15x over state-of-the-art systems by reducing wait times and optimizing resource allocation effectively.
  • Latency Reduction: The system shows considerable improvements in reducing average and tail latencies, emphasizing its potential for high-demand agentic tasks. Figure 2

    Figure 2: Multi-engine, Main Results. Latencies (Avg., P95/99) w.r.t. different load balancing policies.

Implementation Considerations

Autellix's deployment involves several intricate aspects:

  • Scheduling Complexity: The introduction of PLAS and ATLAS necessitates efficient process tracking and priority management, but these are essential for reducing bottlenecks.
  • Data Locality Optimizations: By optimizing LLM call routing based on program data sharing characteristics, Autellix minimizes redundant computations across engines.
  • Scalability Potential: The system is designed for expansive LLM infrastructure, with demonstrated scalability across multiple engines and workloads.

Conclusion

Autellix represents a pivotal advancement in LLM serving systems by effectively treating entire programs as atomic units for scheduling, achieving remarkable improvements in efficiency and resource usage. The system's ability to dynamically adapt to agentic workloads highlights its relevance for future AI applications, ensuring sustained performance under increasing computational demands.

Future Directions

Future developments could expand Autellix's capabilities with speculative execution, incorporating insights from compiler optimizations to anticipate program paths further. Additionally, exploration into post-training LLM optimizations could complement Autellix's robust serving infrastructure, enhancing the overall capability of LLM-driven applications.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 6 tweets and received 17 likes.

Upgrade to Pro to view all of the tweets about this paper: