Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 177 tok/s
Gemini 2.5 Pro 50 tok/s Pro
GPT-5 Medium 31 tok/s Pro
GPT-5 High 32 tok/s Pro
GPT-4o 93 tok/s Pro
Kimi K2 183 tok/s Pro
GPT OSS 120B 447 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

AgentPoison: Backdoor Attacks in AI Systems

Updated 11 November 2025
  • AgentPoison is a framework for backdoor attacks in AI, characterized by dynamic trigger generation, stealth, and robust attack methodologies.
  • It employs various techniques such as GAN-based shrinkage and attention mask search to craft triggers across modalities like vision, graphs, and federated systems.
  • Optimization frameworks balance high attack success rates (often >90%) while preserving clean data performance and evading conventional detection defenses.

AgentPoison refers to the systematic methodology, architecture, and empirical frameworks for implementing and analyzing backdoor attacks and defenses across a variety of AI model pipelines, including dynamic graphs, code search, computer vision, diffusion models, hashing-based retrieval, and federated systems. The concept encompasses both general strategies for poisoning agent-based learning systems (where “agent” may denote models, clients, code processes, or graph entities) and specific domain instantiations—characterized by trigger design, stealth, transferability, and interactions with robust learning rules and model defenses.

1. Conceptual Foundations and Threat Model

AgentPoison attacks are typified by the injection of maliciously crafted triggers or patterns into a dataset, model, or system during training. The injected artifact (“trigger”) is designed such that, when presented at inference time, the trained agent (model) is forced to mispredict or take specific adversary-chosen actions, while normal behavior is retained on clean data.

Key principles include:

  • Trigger Construction: Spatial, syntactic, semantic, or temporal triggers; static (fixed location/pattern) or dynamic (distributional over pattern/placement).
  • Stealth: Minimal visible change to benign prediction or retrieval metrics; triggers are often shrunken via gradient-based selection, hidden via steganography, or distributed in frequency or graph space.
  • Agent Generality: Applicable to neural networks, graph models, code representation systems, diffusion/denoising architectures, and federated/ensemble settings.
  • Poison Ratio: Fraction of the training data modified (typically ≤10%).

Underlying threat models assume adversarial access to the training process, ranging from full white-box (model and gradients accessible) to practical gray-/black-box scenarios (only data-level or limited parameter control).

2. Trigger Generation Techniques

Trigger design in AgentPoison systems varies by modality, but unified strategies are evident:

Approach Modality/Domain Key Mechanism
GAN-based shrinkage Dynamic link prediction Encoder–LSTM–Decoder generates time-varying subgraph sequences, links are pruned by attack-gradient magnitude (Chen et al., 2021)
Attention mask search DNNs/Vision Transformers Residual attention maps optimize mask location/shape for maximal activation (Gong et al., 9 Dec 2024)
Frequency perturbation Vision SSMs (ViM) Distributed triggers in highly sensitive frequency bins, reconstructed via inverse DFT (Wu et al., 1 Jul 2025)
Steganography Diffusion models (image-to-image) Per-sample DCT-based trigger hiding, leveraging mid-frequency coefficient substitution (Chen et al., 8 Apr 2025)
Shadow centroid Deep-hashing systems Features of poisoned queries are aligned to anchor centroids from surrogate data (Zhou et al., 9 Oct 2025)

Dynamic/physical-world variants randomize trigger location, scale, and appearance during training to maintain robustness against real-world transformation and occlusion (Li et al., 2021). Triggers may exist in the graph topological space, spiking event timings, or as code grammar insertions.

3. Optimization Frameworks

AgentPoison attacks employ composite objectives balancing attack success and stealth/utility:

  • Backdoor Losses: LattackL_{attack} terms enforce misprediction on triggered data.
  • Clean Performance Preservation: LcleanL_{clean} or LbenignL_{benign} retains accuracy on non-triggered inputs.
  • Visibility Regularization: LvisL_{vis} (e.g., SSIM, \ell_\infty norm) constrains trigger perceptibility.
  • Gradient-based Link Selection: For graph models, trigger links are ranked by absolute attack-gradient magnitude for maximal impact sparsity.
  • Alternating Retraining: Trigger and model are alternately co-optimized—retraining on mixed or clean samples per iteration for improved accuracy and trigger embedding (Gong et al., 9 Dec 2024).
  • Knowledge Distillation: In federated settings, a distillation regularizer aligns malicious clients’ model outputs with the benign global model to evade outlier-based aggregation defenses (Wang et al., 2022).

Pseudocode reflects these steps, with variants specialized for modality such as input batching, poisoning ratio scheduling, adversarial sample generation, and hyperparameter tuning for stealth/attack strength trade-off.

4. Empirical Performance

Performance evaluation leverages domain-specific metrics:

Metric Description
Attack Success Rate (ASR) Fraction of test inputs with triggers causing misprediction
Clean Data Accuracy (CDA) Accuracy on unmodified test set
Stealth/Perceptual Metrics SSIM, LPIPS, grammar error (ΔG\Delta G), perplexity (ΔP\Delta P)
Robustness to Defenses ASR measured after patch/pruning/distillation defenses, visibility tests, and clustering-based filters

Empirical findings across domains indicate:

  • ASR > 90% typical for dynamic graph, code search, SSM, and diffusion agents at ≤10% poison fraction (Chen et al., 2021, Qi et al., 2023, Wu et al., 1 Jul 2025, Truong et al., 26 Feb 2025, Chen et al., 8 Apr 2025).
  • Clean accuracy drops are generally <2%.
  • Triggers are undetectable by outlier, pruning, or entropy-based defenses in advanced frameworks.
  • Stealth triggers (gradient-shrunken, DCT-hidden, syntactic/semantic text, shadow-feature alignment) evade classical detection pipelines (STRIP, Neural Cleanse, clustering).

5. Countermeasures and Defense Frameworks

Defensive research against AgentPoison includes:

  • Clustering-based Data Filtering: Density-based algorithms (HDBSCAN/UMAP) identify tight clusters of representation outliers (CUBE) (Cui et al., 2022).
  • Influence Graph Extraction: Maximum-average subgraph heuristics isolate mutually influential (likely poisoned) samples via influence function approximations (Sun et al., 2021).
  • Fine-tuning and pruning: Adaptive retraining fails to fully erase stealth triggers (especially in deep-hashing or dynamic-link settings).
  • Backdoor Vector Arithmetic: Weight-space subtraction of pre-computed attack vectors (IBVS) cancels unknown triggers in model-merging agents (Pawlak et al., 9 Oct 2025).
  • Trigger inversion via multi-step loss: Diffusion model defenses recover and invert hidden triggers by analyzing timestep distribution shift and denoising consistency (PureDiffusion) (Truong et al., 26 Feb 2025).

However, advanced AgentPoison attacks typically withstand mainstream defenses, especially when triggers are dynamic, distributed, and model-aware.

6. Limitations, Open Problems, and Future Directions

Current limitations include:

  • White-box assumption (gradient/model access) in trigger optimization.
  • Scaling bottlenecks for large graphs or batch sizes; cubic-time influence graph construction may be prohibitive (Sun et al., 2021).
  • Black-box attacks remain challenging in absence of model gradients or hidden states.
  • Lack of formal guarantees for transferability of surrogate-trigger attacks beyond empirical evidence.

Open problems/future research:

  • Provable robust training against arbitrary small-subgraph or stealth-trigger poisoning.
  • Physical-world trigger detection and defense under uncertain transformations.
  • Adaptive multi-trigger attack and defense for large-scale model-merging or federated agents.
  • Cross-domain generalization theory for hashing-based and code-based agent backdoors.
  • Extensions to spiking neural networks and multimodal architectures.

AgentPoison, as a broad class of attack and defense strategies, constitutes a primary security challenge for AI systems—necessitating continued research into dynamic, stealthy trigger generation, robust optimization, empirical validation, and scalable defense.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to AgentPoison.