Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
88 tokens/sec
Gemini 2.5 Pro Premium
46 tokens/sec
GPT-5 Medium
16 tokens/sec
GPT-5 High Premium
17 tokens/sec
GPT-4o
95 tokens/sec
DeepSeek R1 via Azure Premium
90 tokens/sec
GPT OSS 120B via Groq Premium
461 tokens/sec
Kimi K2 via Groq Premium
212 tokens/sec
2000 character limit reached

LLM-based Multi-Agent Systems

Updated 19 August 2025
  • LLM-based Multi-Agent Systems are frameworks that integrate large language models with the MAPE-K loop to enable adaptive, context-aware decision-making.
  • These systems utilize rich natural language processing to synthesize sensor data, history, and communication for dynamic strategy formation and emergent behaviors.
  • Practical implementations, such as online marketplace simulations, demonstrate enhanced negotiation and coordination despite challenges like context window limitations and computational overhead.

LLM-based Multi-Agent Systems (MASs) are artificial intelligence frameworks where each agent is a software entity endowed with language understanding and reasoning capabilities derived from large-scale pretrained models, often of the GPT-x family. In these systems, agents independently monitor their environment, interpret input, reason about their own state and that of others, plan actions, and communicate with peers using expressive natural language. By integrating LLMs into the traditional autonomic computing paradigm, these MASs achieve self-adaptation and enhanced inter-agent communication that supports both individual autonomy and global group objectives. Recent methodologies, notably those built on the MAPE-K (Monitoring, Analyzing, Planning, Executing, Knowledge) loop, embed LLMs deep within agent control flows to realize robust, adaptive, and emergent behaviors in complex, dynamic environments.

1. LLM Integration within Multi-Agent Control Loops

In LLM-based MASs, each agent extends the classic MAPE-K loop by using an LLM (e.g., GPT-4) to handle the analysis and planning phases. The workflow is as follows: sensor data and inter-agent messages are aggregated during the monitor phase and encoded as a natural language prompt, which is then supplied to the LLM engine. The LLM’s output, constituting both the environmental analysis and the corresponding action plan, is interpreted by the agent and executed. This results in agents that can interpret sophisticated, context-rich environmental states and formulate adaptive, context-sensitive strategies in real time.

Key workflow schematic:

Action=E(LLM(M(Data)History))\text{Action} = E\big( \text{LLM}(M(\text{Data}) \oplus \text{History}) \big)

Where M(Data)M(\text{Data}) means the monitored state, \oplus represents the concatenation with interaction history, and E()E(\cdot) denotes the agent's action execution function.

This tight integration allows each agent to not only synthesize historical context but also exhibit behaviors—including emergent ones such as cross-checking between similar roles or even “self-messaging”—which are not explicitly programmed but arise from the agent’s reasoning and communicative capacity.

2. Self-Adaptation and the MAPE-K Paradigm

Self-adaptation in LLM-based MASs is achieved by continuously closing a feedback loop between agent perception and action. The LLM component is critical in the analyze+plan stages, allowing for natural language-based synthesis of observations and dynamic formation of adaptive strategies. After each execution, the agent’s new state is monitored, providing fresh data for the subsequent iteration.

The MAPE-K structure decomposes the self-adaptation as follows:

  • Monitoring: Aggregation of sensor data and inter-agent messages.
  • Analyzing and Planning: Use of LLM for reasoning over concatenated prompt+history to select the next plan.
  • Executing: Translation of LLM plans into concrete system actions.
  • Knowledge: Storage and reuse of both static pretrained model weights and short-term prompt/context history.

This approach generates agents that autonomously negotiate behavioral updates, navigate dynamic scenarios, and evolve novel coordination protocols at runtime, reinforcing adaptability in open systems.

3. Communication Expressiveness and Coordination

LLMs substantially elevate the expressiveness of agent communication. Unlike traditional MASs that rigidly use hand-designed message formats, LLM-based agents exchange richly structured natural language utterances—enabling nuanced articulation of intent, justification of actions, negotiation, and contextual reporting at a human-readable level. Consequently, agents dynamically invent new negotiation mechanisms, report internal deliberations transparently, and adapt their communication style to evolving environmental and peer contexts.

Challenges mitigated via LLM-based communication include:

  • Overly limited “classical” protocol expressiveness.
  • High coordination and synchronization overhead for evolving protocols.
  • Poor scalability in the face of unforeseen, rapidly shifting communication needs.

A principal limitation that remains is history management, especially given the finite context window of transformer-based LLMs. Current deployments rely on rudimentary local memory (concatenated prompt histories), though the paper identifies the integration of more advanced memory modules as a critical future direction.

4. Practical Implementation: Case Study on Online Market MAS

A concrete evaluation is presented through the implementation of a simplified online book marketplace using the JADE agent framework. Here, agents play the roles of dynamic sellers (setting prices) and buyers (optimizing for lowest price), interacting through a loop in which the key steps are:

  • Agent encapsulates state and received messages into a prompt.
  • Prompt is supplied to GPT-4 (temperature = 0.7) via API.
  • LLM output is parsed and mapped to concrete economic actions (e.g., price updates, negotiation offers).
  • Conversation and decision flows are governed via successive concatenations into the next prompt, simulating rudimentary local history.

Empirical findings:

  • Agents demonstrated sophisticated negotiation and decision-making, sometimes outperforming hand-coded strategies.
  • Emergent phenomena, such as inventive self-messaging and inter-seller cross-checking, surfaced, underscoring both the creativity and unpredictability that LLMs introduce to MAS behavior.
  • System performance was strongly influenced by the quality of local history management and the contextual awareness designed into the prompt construction.
Feature Classic MAS LLM-based MAS (paper)
Communication Format Rigid, static Rich natural language
Adaptivity Hand-coded strategies Dynamic, LLM-driven
Memory Management Explicit buffers Prompt/history concatenation
Emergent Behaviors Limited Readily observed

5. Tradeoffs, Performance, and Real-World Constraints

Performance and Resource Considerations

LLM integration introduces notable computational overhead and latency, particularly where real-time responsiveness is critical. Each agent’s control loop includes at least one LLM invocation per cycle, with inference costs scaling linearly in the number of active agents and conversation rounds. For laboratory settings and moderate agent populations, this overhead is manageable, but for high-frequency or time-critical environments, minimization strategies are essential (e.g., prompt engineering, context pruning, model distillation).

Limitations and Opportunities

While embedding LLMs into MASs dramatically enhances adaptability and expressiveness, several open challenges remain:

  • History Management: Synthesizing long-term memory efficiently while preserving LLM context windows.
  • Interpretability: Deciphering LLM-generated reasoning in safety- or audit-sensitive deployments.
  • Ethics and Safety: Ensuring responsible operation when agents autonomously adapt goals or invent new communication protocols.
  • Scalability and Latency: Exploring local caching, model quantization, or deploying dedicated models per agent to alleviate overhead.

Suggested deployment strategies include reserving LLM-empowered MASs for contexts that benefit from high communication expressiveness and runtime adaptability (e.g., human-in-the-loop decision support, edge device orchestration), and introducing regular human validation checkpoints for safety-critical tasks.

6. Implications, Extensions, and Future Directions

The architecture presented in this work demonstrates that LLM-powered MASs substantially advance autonomic system design by fusing linguistic proficiency with traditional self-adaptive frameworks. Looking ahead:

  • Unique per-agent LLMs: Deploying distinct LLM instances or non-shared generative models to curtail inadvertent knowledge sharing and foster agent differentiation.
  • Integration with other domains: Extending beyond simulated marketplaces to evolutionary robotics, IoT coordination, and distributed problem-solving under uncertainty.
  • Human-agent collaborative systems: Embedding conversational interfaces to blend human guidance with agent-driven adaptation in closed-loop feedback.
  • Advanced planning and memory: Supplementing LLMs with richer, possibly symbolic, memory modules and local planning stacks to strengthen reliability in complex environments.

This paradigm signals a transition from static, hand-crafted agent protocols toward highly autonomous, flexible, and communicative agent societies capable of responding creatively and robustly to open-ended, real-world challenges.