Papers
Topics
Authors
Recent
2000 character limit reached

Agentic AI in Cybersecurity

Updated 14 December 2025
  • Agentic AI in cybersecurity is an approach where autonomous agents perform end-to-end perception, reasoning, and decision-making for digital defense.
  • Modular architectures and multi-agent coordination enable scalable detection, adaptive response, and policy enforcement across network, cloud, and IoT environments.
  • Empirical evaluations demonstrate high accuracy, reduced response times, and robust performance in both offensive and defensive cybersecurity operations.

Agentic AI in Cybersecurity refers to the use of autonomous or near-autonomous AI agents—typically capable of end-to-end perception, reasoning, decision, and tool use—to defend, monitor, and attack digital infrastructure. These agentic systems demonstrate adaptive, memory-augmented, and often cooperative behaviors operating at machine speed across a range of cybersecurity domains, including network defense, adversarial simulations, threat intelligence, and policy enforcement. Their application spans decentralized monitoring, multi-modal detection, incident response, training scenario generation, and both offensive and defensive operations, underpinned by modern AI advances in reinforcement learning, LLMs, and multi-agent system design.

1. Architectures and Design Principles

Agentic cybersecurity frameworks commonly employ modular or layered architectures distributing agents across nodes, domains, or tasks. In NetMoniAI—a canonical two-tier agentic AI architecture for network security—lightweight micro-agents are deployed to each network node, featuring multi-layer logic (service, agent, model, application) for local feature engineering, Mahalanobis-distance anomaly scoring, and BERT/LLM-based semantic classification. A Central Controller aggregates only pertinent, threshold-exceeding alerts, computes a weighted global risk metric, and performs time-windowed correlation for coordinated attack detection. Communication between components is event-driven, asynchronous, and resource-minimizing, leveraging REST/JSON, dynamic sampling, and heartbeats to maintain scalability and resilience even on IoT-class nodes. The approach demonstrates bounded computation, sublinear scaling (per-report processing <<30ms), and sub-5s detection-classification pipelines (Zambare et al., 12 Aug 2025).

Multi-agent frameworks such as AgenticCyber extend this paradigm to multimodal threat detection scenarios, fusing log, vision, and audio agents each with specialized anomaly detectors and LLM-driven reasoning. Fusion and policy selection occur via orchestration layers (LangChain) and large-multimodal models (e.g., Gemini), supporting cross-modal attention mechanisms and Q-learning-based adaptive response. Agents run as containerized microservices, facilitating elastic deployment (Kubernetes, Kafka) and modular extensibility to new modalities or analysis tools (Roy, 6 Dec 2025).

Recent architectures generalize to full-stack environments (cloud, edge, API, IoT), assigning agents as sidecars or gateways to critical nodes and enforcing Markov decision process (MDP)-based policy updates. Behavioral baselining, federated intelligence sharing, decentralized risk scoring, and dynamic policy adjustment underpin these adaptive frameworks. Zero-trust principles, least-privilege enforcement, and schema-constrained tool invocation are systematically embedded in such adaptive security blueprints (Olayinka et al., 25 Sep 2025).

2. Methodologies for Detection, Response, and Adaptation

Agentic agents perform local feature extraction and anomaly scoring (e.g., Mahalanobis distances over vectorized traffic metrics), followed by semantic threat classification utilizing local or remote LLMs. When thresholds are exceeded, only then are heavier inference and reporting invoked. At the system level, controllers synthesize multi-source agent reports by balancing statistical confidence and semantic classification probabilities in weighted aggregations: Sglobal=i=1Nwi[αSi+(1α)(ln(1pi))]S_\mathrm{global} = \sum_{i=1}^N w_i[\alpha S_i + (1-\alpha)(-\ln(1-p_i))] where wiw_i encodes node trust, α\alpha adjusts emphasis between statistical anomaly and LLM-provided semantic certainty (Zambare et al., 12 Aug 2025).

Adaptive engagement paradigms displace static incident-response kill chains. Defensive agents continuously monitor for precursory signals, adjust internal thresholds (τ\tau), and modulate response actions (atπ(atst)a_t \sim \pi(a_t|s_t)) in real time, orchestrating containment (isolation, honeypots), and updating model parameters with reinforcement-based feedback loops. Closed-loop architectures incorporating contextual scoring, multi-phase clustering (e.g., DBSCAN), and human-in-the-loop oversight yield “defense web” ecosystems—prototyping perpetual learning and response cycles (Tallam, 28 Feb 2025). Cross-modality detection employs attention-based fusion of multi-agent outputs; Q-learners select remediation policies optimized for reward metrics balancing neutralization speed and false positive rates (Roy, 6 Dec 2025).

Training/Evaluation regimes emphasize reproducibility and quantification. Empirical detection metrics include accuracy, precision, recall, F1, detection latency, and mean time to respond (MTTR), with NS-3 simulations for benchmarking scalability. Domain-specific evaluation extends to continuous risk scoring, federated learning convergence, and adaptation/acquisition curves over extended deployments (Zambare et al., 12 Aug 2025, Olayinka et al., 25 Sep 2025).

3. Security Risks, Threat Models, and Defensive Strategies

Agentic AI systems introduce security risks distinct from conventional AI and software automation. Expanded attack surfaces derive from agentic autonomy, persistent memory, tool-chaining, and inter-agent coordination. Threats are taxonomized as follows (Datta et al., 27 Oct 2025, Khan et al., 16 Oct 2024, Ghosh et al., 27 Nov 2025):

  • Prompt injection and jailbreaks: Direct and indirect manipulations exploit input vectors, leading agents to override constraints or propagate malicious instructions—obfuscated or multi-modal attack surfaces exacerbate risk.
  • Unauthorized tool use: Agents with over-broad capabilities can autonomously discover and execute web exploits (XSS, SQLi), one-day/zero-day attacks, or escalate privileges through emergent tool compositions.
  • Multi-agent protocol-level attacks: Spoofing, replay, denial-of-service, credential leakage, or knowledge-poisoning propagate through MCP/A2A communication standards, undermining trust and auditability.
  • Interface/environment fragility: Brittle observation-action mappings (e.g., GUI scraping agents) and non-deterministic perception raise the risk of misexecution and accidental data exfiltration.
  • Governance and autonomy concerns: Full autonomy can, without robust policy frameworks, result in human-out-of-the-loop errors, ethical violations, or misaligned persistent behaviors.

Defensive strategies are thus layered:

  • Architectural controls involve strong separation of agent containers, minimal privilege (RBAC), and secure sandboxes for tool execution and memory storage.
  • Design-time application of formal methods, static analysis, schema-constrained tool invocation, and adversarial fine-tuning (including circuits-breakers for suspicious sequences) limits attack propagation.
  • Runtime mitigations include policy enforcement agents, anomaly detection (statistical and behavioral analytics), evidence-based plan validation, and enforced audit trails. Shadow-monitoring agents and zero-knowledge audit proofs (e.g., GROTH16) create cryptographic accountability at scale (Huang et al., 29 Oct 2025).
  • Continuous governance leverages internal red teaming, risk assessment via auxiliary agents, and prompt/plan repair agents with formal policy gates and human approval for high-impact actions (Ghosh et al., 27 Nov 2025).

Risk analysis frameworks (e.g., MAESTRO) span layered stacks: from model, data, and inference, through deployment, observability, and multi-agent ecosystems. Operational risk is quantified as R=P×I×ER = P \times I \times E (likelihood, impact, exploitability), with resilience functions measuring mitigated risk against system capacity (Zambare et al., 12 Aug 2025).

4. Evaluation, Comparative Studies, and Benchmarks

Performance of agentic AI in cybersecurity is empirically validated across multiple benchmarks:

  • Detection performance: In large-scale simulation, agentic micro-agent architectures achieve detection accuracy up to 0.96, precision 0.95, recall 0.97, and F1 scores of 0.96 within a mean latency of 4.2s, maintaining <<4% FPR and <<3% FNR under resource constraints (Zambare et al., 12 Aug 2025).
  • Multimodal systems: AgenticCyber demonstrates a 96.2% F1, 420ms average latency, 65% reduction in MTTR, and improved situational awareness (0.92 Endsley score), outperforming static MAS and classic IDS baselines by significant margins (Roy, 6 Dec 2025).
  • CTF/attack-defense settings: RedTeamLLM outperforms classical LLM-based agents on offensive tasks (higher step coverage, fewer tool calls, and increased overall success rate), while empirical studies with dual attack/defense agents (CAI framework) indicate defensive advantage is sensitive to operational success metrics (SLA uptime, intrusion avoidance). There is no unconditional superiority for attacker agents under real operational constraints (Challita et al., 11 May 2025, Balassone et al., 20 Oct 2025).
  • Scenario generation: Agentic feedback loops (AgentCyTE) achieve near-100% generation accuracy in schema-constrained environments with only a few rounds of refinement, vastly outperforming unconstrained zero-shot LLM generation (Rodriguez et al., 29 Oct 2025).

Standardized benchmarks (AgentBench, DefenderBench, CyberSOCEval, CyBench, SecEval, AttackSeqBench, AutoPenBench) are evolving; however, comprehensive SOC-end-to-end workflow evaluation and robust multi-agent coordination metrics remain open challenges (Vinay, 7 Dec 2025).

5. Operational, Ethical, and Governance Considerations

Agentic AI requires rigorous governance to mitigate operational, ethical, and legal risks:

  • Ethical governance layers explicitly encode fairness constraints (e.g., upper bounds on P(blocknormal)P(\mathrm{block}|\mathrm{normal}), ECS metrics), soft or hard-penalty functions, and human oversight gates for ambiguous or high-impact actions; these layers are essential in resource-constrained and critical infrastructure settings (Adabara et al., 8 Dec 2025).
  • Policy frameworks such as AAGATE operationalize NIST AI RMF functions, employing explainable policy engines, continuous compliance proof generation (ZK-provers), digital identity rights, behavioral analytics, and cognitive degradation monitors. Red-teaming is institutionalized via agents (Janus SMA) for plan drift detection and response (Huang et al., 29 Oct 2025).
  • System-level principles including least privilege, complete mediation, tamper-resistance, and secure information flow must be adapted to the probabilistic and natural-language controlled environment of agentic AI, requiring new formalisms for security boundaries, dynamic policy inference, and taint tracking in vector spaces (Christodorescu et al., 1 Dec 2025).
  • Audit and compliance: Immutable, append-only logs, cross-agent context handoff, and policy traceability underpin trust in agentic workflows and satisfy ISO, GDPR/CCPA, and zero-trust compliance requirements (Olayinka et al., 25 Sep 2025, Huang et al., 29 Oct 2025).

6. Research Challenges and Future Directions

Several technical frontiers remain unresolved:

  • Unified evaluation frameworks for end-to-end, multi-agent workflows, coordinated tool-use correctness, and high-impact action validation are underdeveloped (Vinay, 7 Dec 2025).
  • Response validation and tool-use correctness demand independent verifier models and formal pre/post-conditions for agent actions.
  • Security by construction requires mechanistic interpretability and real-time “circuit-breaker” interventions at inference, embedded policy languages, and formal IRs between LLM and executor (Christodorescu et al., 1 Dec 2025).
  • Multi-agent coordination and memory: Consensus protocols, contract nets, persistent and replayable context storage are needed for robust, non-redundant, and reproducible multi-agent cyber operations (Vinay, 7 Dec 2025).
  • Long-horizon safety and adversarial adaptation: Metrics and mechanisms for detecting sleeper behaviors, handling adaptive attackers, and preventing cascading action chains or memory poisoning are essential (Ghosh et al., 27 Nov 2025, Datta et al., 27 Oct 2025).
  • Human–agent interfaces must reconcile usability and security, balancing cognitive load and permission solicitation for high-risk or ambiguous actions (Christodorescu et al., 1 Dec 2025).

Continued secure deployment of agentic AI in cybersecurity will require advances across formal methods, distributed architecture, interpretability, multi-agent learning, and interdisciplinary governance. As agentic AI increases its field autonomy and decision bandwidth, thorough operational safeguards, reproducible evaluation, and rigorous risk management processes become indispensable for safe scaling and trustworthiness in digital defense (Datta et al., 27 Oct 2025, Ghosh et al., 27 Nov 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (15)

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Agentic AI in Cybersecurity.