MorphAgent: Self-Adaptive Multi-Agent System
- MorphAgent is a decentralized, self-adaptive multi-agent system that enables autonomous collaboration in LLM-based environments using evolving agent profiles.
- It leverages three metrics—Role Clarity, Role Differentiation, and Task–Role Alignment—to optimize agent roles and maintain adaptive task alignment.
- The system employs a two-phase workflow of dynamic profile updates and iterative task execution, ensuring robust performance even under high failure probabilities.
MorphAgent is an autonomous, self-organizing, and self-adaptive multi-agent system (MAS) for decentralized agent collaboration, specifically designed for LLM-based environments. It enables agents to dynamically evolve their roles and capabilities via self-evolving agent profiles, guided by three complementary metrics. Distinct from traditional MAS architectures, MorphAgent abolishes predefined roles and centralized coordination, instead achieving robust team-level performance and adaptability to task requirements through decentralized, metric-driven optimization and continuous feedback loops (Lu et al., 2024).
1. Motivation and Problem Framework
LLM-based MAS have historically relied on three paradigms: centralized coordination via a manager agent, predefined workflows or static topologies with fixed roles, and rigid SOP-based pipelines lacking runtime flexibility. These designs constrain individual agent autonomy, inhibit self-organization, and preclude self-adaptability to changing environments or task objectives. MorphAgent is motivated by principles observed in natural swarms, such as ants and birds, seeking to instantiate emergent intelligence in artificial systems through locally-governed, adaptive behaviors. The goals are:
- Individual autonomy: Agents independently select actions using only locally available information.
- Self-organization: Team structures and communication patterns emerge without central control.
- Self-adaptability: Agent roles and skill profiles adapt online to evolving task or environmental demands (Lu et al., 2024).
2. Self-Evolving Agent Profiles
Each agent maintains a dynamic profile , comprising:
- : functional capabilities (e.g., "Python coding," "legal analysis"),
- : contextual parameters (domain-specific preferences or adjustments),
- : interaction rules (policies for peer communication).
Profile updates utilize the LLM prompting mechanism, formalized as
where are local observations, are recent actions of neighboring agents, is the LLM-driven update function, and denotes the dynamic neighborhood.
Profile refinement is guided by three metrics:
- Role Clarity Score (RCS):
Where : syntactic depth, : lexical entropy, : cosine similarity with "skill prototype" and count of skill-bearing nouns ().
- Role Differentiation Score (RDS):
Promoting separation in profile embedding space; is a sigmoid normalizer, is the sentence embedding.
- Task–Role Alignment Score (TRAS):
Where
and are task complexity and agent capability measures, computed via cosine similarities to reference vectors.
3. Two-Phase Decentralized Workflow
MorphAgent employs a two-phase process:
Phase 1: Profile Update Phase
- Agents are initialized with minimal or identical profiles.
- For up to rounds (or until all metrics increase by at least 0.1), each agent:
- Computes its current RCS, RDS, and TRAS.
- Compares scores to previous iterations.
- Selects a targeted feedback prompt and uses LLM to update .
Outputs a set of optimized agent profiles.
Phase 2: Task Execution Phase
- Agents iterate through an Observe–Think–Act loop:
- Observe shared workspace (problem statement, peer outputs).
- Decide (via LLM policy ) whether to EXECUTE (contribute a result) or SKIP (yield turn).
- Act on decision.
Consensus is reached when all agents consecutively issue SKIP.
- Upon execution failure or disruptive events, agents may invoke a lightweight (single-round) profile update and resume.
4. Team Dynamics and Metric Interplay
Role overlap is explicitly penalized through the RDS metric: agents with highly similar profiles are prompted to differentiate. The interplay among the three core metrics directs profile optimization:
- Overemphasis on RCS can result in homogenous, but well-articulated, agent profiles.
- Overemphasis on RDS leads to maximally separated—but potentially irrelevant—roles.
- TRAS tempers both by aligning profile features with task requirements.
The global objective function formalizing team profile quality is:
where are hyperparameters. Optimization proceeds by decentralized, gradient-free updates driven by iterative LLM prompting.
5. Empirical Evaluation and Quantitative Results
Experiments incorporate the following evaluation settings:
- Benchmarks: BigCodeBench (code generation; 1140 functions), BigBenchHard (general reasoning), MATH (mathematical reasoning).
- Base LLMs: GPT-4o-mini, DeepSeek-V3, GPT-3.5-Turbo (ablation studies).
- Baselines: GPTSwarm, AgentVerse, AFLOW.
- Dynamic environment: Each agent has per-round failure probability (0 to 0.8) to skip their turn, simulating robustness to node failures.
Key findings:
- On BigCodeBench with no failures, MorphAgent obtains around 52–54% accuracy versus 47–50% for all baselines.
- Under high failure probability (up to 0.8), baselines drop to 1–20% accuracy, while MorphAgent retains 40–54%.
- Across all tested backbones, MorphAgent consistently outperforms existing MAS frameworks.
Ablation results (Section 4.2) demonstrate the necessity of all three metrics:
| Setting | GPT-4o-mini | GPT-3.5-Turbo |
|---|---|---|
| + None | 50.67% | 38.33% |
| + RCS only | 50.00% | 39.33% |
| + RDS only | 41.66% | 37.00% |
| + TRAS only | 49.66% | 35.33% |
| MorphAgent (all) | 52.00% | 43.33% |
6. Insights, Limitations, and Prospective Directions
Dynamic, self-evolving profiles enable robust self-organization into task-appropriate, complementary roles. Effective MAS operation requires balancing role clarity, inter-agent differentiation, and task-role alignment; reliance on any single metric degrades overall quality. Decentralized, LLM-driven profile updates provide resilience against agent dropouts or execution failures.
Limitations include computational overhead from multiple LLM calls during profile optimization (up to 5 initial rounds) and increasing overhead with large agent counts—albeit with sublinear growth in practice (Appendix A.7).
Prospective directions include the development of more efficient peer-to-peer communication protocols, lightweight embedding-based metric estimators to reduce LLM call volume, adaptive meta-learning of metric weights () for task specificity, support for heterogeneous tool-using agents, and formal game-theoretic analysis of profile co-evolution (Lu et al., 2024).