pith. machine review for the scientific record. sign in

arxiv: 2603.09002 · v2 · submitted 2026-03-09 · 💻 cs.CR · cs.AI

Recognition: no theorem link

Security Considerations for Multi-agent Systems

Authors on Pith no claims yet

Pith reviewed 2026-05-15 14:06 UTC · model grok-4.3

classification 💻 cs.CR cs.AI
keywords multi-agent systemsAI securitythreat modelingsecurity frameworkscybersecuritynon-determinismdata leakageOWASP
0
0 comments X

The pith

No reviewed security framework covers a majority of threats in any category for multi-agent AI systems.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Multi-agent systems let autonomous agents share persistent memory, delegate tool use, and coordinate through communication, creating attack surfaces unlike those of single models. The paper constructs 193 concrete threat items across nine categories through expert-validated modeling and then scores sixteen existing frameworks on a three-point scale. No framework reaches majority coverage in even one category, with non-determinism and data leakage showing the lowest average scores. The OWASP Agentic Security Initiative records the highest overall coverage at 65.3 percent. The evaluation supplies the first direct comparison and concrete selection guidance based on measured gaps.

Core claim

Multi-agent AI systems introduce qualitatively distinct security vulnerabilities from those of singular models; a quantitative scoring of sixteen frameworks against 193 validated threats shows that none achieve majority coverage in any of the nine risk categories, while non-determinism and data leakage remain the most under-addressed.

What carries the argument

A set of 193 distinct threat items across nine risk categories, derived from generative AI-assisted modeling and domain-expert validation, then used to score each framework on a three-point scale.

If this is right

  • Framework selection for MAS deployments should favor the OWASP initiative for design-phase coverage while supplementing gaps elsewhere.
  • Non-determinism and data leakage in shared-memory and inter-agent settings require targeted mitigations absent from current frameworks.
  • Development and operational phases show lower average scores than design, indicating where new controls are most needed.
  • The absence of majority coverage across all frameworks implies that custom controls will be required for production MAS.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Organizations may need to combine elements from multiple frameworks rather than relying on any single one.
  • Real-world deployment data could test whether the 193 items capture the most common attack patterns observed in practice.
  • Inter-agent communication protocols may require new standards that existing frameworks do not yet address.

Load-bearing premise

The 193 threat items fully and accurately represent the complete MAS cybersecurity risk landscape.

What would settle it

Discovery of a framework that scores above 50 percent coverage on every one of the nine categories using the same 193-item list, or identification of major MAS threats missing from that list.

Figures

Figures reproduced from arXiv: 2603.09002 by Dheeraj Arremsetty, Moses Ndebugre, Tam Nguyen.

Figure 1
Figure 1. Figure 1: Coverage of 193 agentic AI threat items per framework, stacked by coverage tier. Frameworks sorted by total coverage [PITH_FULL_IMAGE:figures/full_fig_p183_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Mean coverage score (averaged across all 16 frame [PITH_FULL_IMAGE:figures/full_fig_p183_2.png] view at source ↗
read the original abstract

Multi-agent artificial intelligence systems or MAS are systems of autonomous agents that exercise delegated tool authority, share persistent memory, and coordinate via inter-agent communication. MAS introduces qualitatively distinct security vulnerabilities from those documented for singular AI models. Existing security and governance frameworks were not designed for these emerging attack surfaces. This study systematically characterizes the threat landscape of MAS and quantitatively evaluates 16 security frameworks for AI against it. A four-phase methodology is proposed: constructing a deep technical knowledge base of production multi-agent architectures; conducting generative AI-assisted threat modeling scoped to MAS cybersecurity risks and validated by domain experts; structuring survey plans at individual-threat granularity; and scoring each framework on a three-point scale against the cybersecurity risks. The risks were organized into 193 distinct main threat items across nine risk categories. The expected minimal average score is 2. No reviewed framework achieves majority coverage of any single category. Non-Determinism (mean score 1.231 across all 16 frameworks) and Data Leakage (1.340) are the most under-addressed domains. The OWASP Agentic Security Initiative leads overall at 65.3\% coverage and in the design phase; the CDAO Generative AI Responsible AI Toolkit leads in development and operational coverage. These results provide the first empirical cross-framework comparison for MAS security and offer evidence-based guidance for framework selection. Please check back for information on the published journal version.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper proposes a four-phase methodology to characterize the cybersecurity threat landscape of multi-agent AI systems (MAS), which feature autonomous agents with delegated tool authority, persistent shared memory, and inter-agent coordination. It constructs a set of 193 distinct threat items across nine risk categories via generative AI-assisted modeling followed by domain-expert validation, then scores 16 existing AI security frameworks on a three-point scale against these threats. The central findings are that no framework achieves majority coverage of any category, Non-Determinism (mean score 1.231) and Data Leakage (1.340) are the most under-addressed domains, and the OWASP Agentic Security Initiative leads with 65.3% overall coverage.

Significance. If the 193-item threat list is shown to be both exhaustive and accurately classified, the work supplies the first quantitative, cross-framework empirical comparison for MAS security. This would offer concrete, evidence-based guidance on framework selection and identify priority gaps (especially non-determinism and data leakage) that future frameworks must address. The explicit scoring protocol and category-level breakdowns would also enable reproducible follow-on studies.

major comments (2)
  1. [Methodology] Methodology section: the description of generative AI-assisted threat modeling and subsequent domain-expert validation supplies no information on the number of experts, the validation protocol, inter-rater agreement statistics, or any cross-check against existing MAS or AI security taxonomies (e.g., NIST or academic surveys). Because the headline coverage percentages and the claim that 'no reviewed framework achieves majority coverage' rest entirely on the completeness and correctness of the 193-item list, these omissions are load-bearing.
  2. [Results] Results and Evaluation sections: the three-point scoring procedure is presented without examples of how individual threats were scored, how consistency across raters or frameworks was maintained, or sensitivity analysis showing how the reported means (Non-Determinism 1.231, Data Leakage 1.340) would change under plausible reclassifications of even a modest subset of the 193 items.
minor comments (2)
  1. [Abstract] The abstract ends with the sentence 'Please check back for information on the published journal version,' which is atypical for an arXiv preprint and should be removed or clarified.
  2. [Results] Table or figure presenting the per-framework, per-category scores would benefit from an explicit legend for the three-point scale and from reporting the raw counts rather than only means.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We are grateful to the referee for their constructive feedback, which has helped us improve the clarity and rigor of our work. Below we respond to each major comment and indicate the revisions made to the manuscript.

read point-by-point responses
  1. Referee: [Methodology] Methodology section: the description of generative AI-assisted threat modeling and subsequent domain-expert validation supplies no information on the number of experts, the validation protocol, inter-rater agreement statistics, or any cross-check against existing MAS or AI security taxonomies (e.g., NIST or academic surveys). Because the headline coverage percentages and the claim that 'no reviewed framework achieves majority coverage' rest entirely on the completeness and correctness of the 193-item list, these omissions are load-bearing.

    Authors: We agree that the original manuscript omitted key details on the expert validation process. In the revised manuscript, we have expanded the Methodology section to describe the number of domain experts involved, the validation protocol (including iterative review and consensus resolution), inter-rater agreement statistics, and explicit cross-checks against the NIST AI Risk Management Framework as well as relevant academic surveys on AI threats. These additions directly support the completeness of the 193-item list and the validity of the coverage claims. revision: yes

  2. Referee: [Results] Results and Evaluation sections: the three-point scoring procedure is presented without examples of how individual threats were scored, how consistency across raters or frameworks was maintained, or sensitivity analysis showing how the reported means (Non-Determinism 1.231, Data Leakage 1.340) would change under plausible reclassifications of even a modest subset of the 193 items.

    Authors: We concur that greater transparency in the scoring procedure is warranted. The revised manuscript now includes concrete examples of how representative threats were scored, details the standardized rubric and reviewer process used to maintain consistency, and presents a sensitivity analysis evaluating the effect of plausible reclassifications on the mean scores. This analysis confirms that the identification of Non-Determinism and Data Leakage as the most under-addressed domains remains robust. revision: yes

Circularity Check

0 steps flagged

No significant circularity in the evaluation methodology

full rationale

The paper constructs an independent set of 193 threat items via generative AI-assisted modeling and domain-expert validation, then applies a three-point scoring process to 16 external frameworks. This produces the reported means (e.g., Non-Determinism at 1.231) and coverage percentages (e.g., OWASP at 65.3%) directly from the application step. No equation or definition equates the output scores to the construction process itself, no parameter is fitted and then relabeled as a prediction, and no load-bearing premise rests on a self-citation chain. The derivation remains self-contained against external benchmarks.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The central claims rest on the domain assumption that MAS introduce qualitatively distinct vulnerabilities and that the constructed threat model is exhaustive.

axioms (1)
  • domain assumption Multi-agent systems introduce qualitatively distinct security vulnerabilities from those documented for singular AI models.
    Explicit premise stated in the opening of the abstract that justifies the entire study.

pith-pipeline@v0.9.0 · 5545 in / 1190 out tokens · 60550 ms · 2026-05-15T14:06:37.699413+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. A Formal Security Framework for MCP-Based AI Agents: Threat Taxonomy, Verification Models, and Defense Mechanisms

    cs.CR 2026-04 unverdicted novelty 6.0

    MCPSHIELD offers a threat taxonomy of 23 attack vectors, a labeled transition system verification model, and a defense-in-depth architecture claiming 91% coverage for MCP-based AI agents.

Reference graph

Works this paper leans on

300 extracted references · 300 canonical work pages · cited by 1 Pith paper · 19 internal anchors

  1. [1]

    The road to agentic ai navigating architecture, threats, and solutions,

    V . Ciancaglini, M. Balduzzi, S. Gariuolo, R. V osseler, and F. Tucci, “The road to agentic ai navigating architecture, threats, and solutions,” Trend Micro Research, 2025. [Online]. Available: https: //www.trendmicro.com/vinfo/us/security/news/security-technology/ the-road-to-agentic-ai-navigating-architecture-threats-and-solutions

  2. [2]

    Microsoft 365 copilot,

    M. Corporation, “Microsoft 365 copilot,” Software application, 2026, aI-powered productivity tool integrating Word, Excel, PowerPoint, Outlook, and more. Available at: https://apps.microsoft.com/detail/ 9WZDNCRD29V9

  3. [3]

    Artificial intelligence risk management framework (AI RMF 1.0),

    National Institute of Standards and Technology, “Artificial intelligence risk management framework (AI RMF 1.0),” National Institute of Standards and Technology, Tech. Rep. NIST AI 100-1, 2023. [Online]. Available: https://doi.org/10.6028/NIST.AI.100-1

  4. [4]

    Adversarial machine learning: A taxonomy and terminology of attacks and mitigations,

    A. Vassilev, A. Oprea, A. Fordyce, H. Anderson, X. Davies, and M. Hamin, “Adversarial machine learning: A taxonomy and terminology of attacks and mitigations,” National Institute of Standards and Technology, Tech. Rep. NIST AI 100-2e2025, 2025. [Online]. Available: https://doi.org/10.6028/NIST.AI.100-2e2025

  5. [5]

    MITRE ATLAS: Adversarial threat landscape for artificial-intelligence systems,

    MITRE Corporation, “MITRE ATLAS: Adversarial threat landscape for artificial-intelligence systems,” The MITRE Corporation, Tech. Rep., 2025, living knowledge base of adversarial ML tactics and techniques, Spring 2025 release. [Online]. Available: https: //atlas.mitre.org/

  6. [6]

    OW ASP Top 10 for Agen- tic Applications 2026,

    OW ASP GenAI Security Project, “OW ASP Top 10 for Agen- tic Applications 2026,” OW ASP Foundation, Tech. Rep., 2026, agentic Security Initiative (ASI), https://genai.owasp.org/resource/ owasp-top-10-for-agentic-applications/

  7. [7]

    Securing agentic AI: A comprehensive threat model and mitigation framework for generative AI agents,

    V . S. Narajala and O. Narayan, “Securing agentic AI: A comprehensive threat model and mitigation framework for generative AI agents,” Amazon Web Services, Proactive Security, Technical Report, 2025, presents the ATFAA threat taxonomy (9 threats, 5 domains) and the SHIELD mitigation framework (6 control strategies). [Online]. Available: https://arxiv.org/...

  8. [8]

    Generative AI responsible AI toolkit, version 1.0,

    Chief Digital and Artificial Intelligence Office, “Generative AI responsible AI toolkit, version 1.0,” U.S. Department of Defense, Chief Digital and Artificial Intelligence Office, Tech. Rep., Dec

  9. [9]

    Available: https://www.ai.mil/Portals/137/Documents/ Resources%20Page/2024-12GenAI-Responsible-AI-Toolkit.pdf

    [Online]. Available: https://www.ai.mil/Portals/137/Documents/ Resources%20Page/2024-12GenAI-Responsible-AI-Toolkit.pdf

  10. [10]

    Taxonomy of failure mode in agentic ai systems,

    P. Bryan, G. Severi, J. D. Gruyter, D. Jones, B. Bullwinkel, A. Min- nich, S. Chawla, G. Lopez, M. Pouliot, A. Fourney, W. Maxwell, K. Pratt, S. Qi, N. Chikanov, R. Lutz, R. Sekhar, R. Dheekonda, B.- E. Jagdagdorj, E. Kim, J. Song, K. Hines, R. Lundeen, S. Vaughan, V . Westerhoff, Y . Zunger, C. Kawaguchi, M. Russinovich, R. Shankar, and S. Kumar, “Taxono...

  11. [11]

    On the regulatory potential of user interfaces for ai agent governance,

    K. J. K. Feng, T. S. Kim, R. Y . Pang, F. Huq, T. August, and A. X. Zhang, “On the regulatory potential of user interfaces for ai agent governance,” 2025. [Online]. Available: https://arxiv.org/abs/ 2512.00742

  12. [12]

    Formal Policy Enforcement for Real-World Agentic Systems

    N. Palumbo, S. Choudhary, J. Choi, P. Chalasani, and S. Jha, “Policy compiler for secure agentic systems,” 2026. [Online]. Available: https://arxiv.org/abs/2602.16708

  13. [13]

    Agentrim: Tool risk mitigation for agentic ai,

    R. Betser, S. Bose, A. Giloni, C. Picardi, S. Padakandla, and R. Vainshtein, “Agentrim: Tool risk mitigation for agentic ai,” 2026. [Online]. Available: https://arxiv.org/abs/2601.12449

  14. [14]

    Overseeing agents without constant oversight: Challenges and opportunities,

    M. Grunde-McLaughlin, H. Mozannar, M. Murad, J. Chen, S. Amershi, and A. Fourney, “Overseeing agents without constant oversight: Challenges and opportunities,” 2026. [Online]. Available: https://arxiv.org/abs/2602.16844

  15. [15]

    Multi-agent systems execute arbitrary malicious code,

    H. Triedman, R. Jha, and V . Shmatikov, “Multi-agent systems execute arbitrary malicious code,” 2025. [Online]. Available: https://arxiv.org/abs/2503.12188

  16. [16]

    On the suitability of llm-driven agents for dark pattern audits,

    C. Sun, Y . Vekaria, and R. Nithyanand, “On the suitability of llm-driven agents for dark pattern audits,” 2026. [Online]. Available: https://arxiv.org/abs/2603.03881

  17. [17]

    Emergent dark patterns in ai-generated user interfaces,

    D. Pandey, “Emergent dark patterns in ai-generated user interfaces,”

  18. [18]

    Available: https://arxiv.org/abs/2602.18445

    [Online]. Available: https://arxiv.org/abs/2602.18445

  19. [19]

    Human-in-the-loop interactive report generation for chronic disease adherence,

    X. Zhang, J. Yu, P. Yan, L. Jiang, X. Shen, M. Cheng, and X. Liu, “Human-in-the-loop interactive report generation for chronic disease adherence,” 2026. [Online]. Available: https: //arxiv.org/abs/2601.06364

  20. [20]

    Agenticcyops: Securing multi-agentic ai integration in enterprise cyber operations,

    S. Mitra, R. Patel, S. Mittal, M. R. Rahman, and S. Rahimi, “Agenticcyops: Securing multi-agentic ai integration in enterprise cyber operations,” 2026. [Online]. Available: https://arxiv.org/abs/ 2603.09134

  21. [21]

    Masfactory: A graph-centric framework for orchestrating llm-based multi-agent systems with vibe graphing,

    Y . Liu, J. Cai, Y . Li, Q. Meng, Z. Liu, X. Li, C. Qian, C. Shi, and C. Yang, “Masfactory: A graph-centric framework for orchestrating llm-based multi-agent systems with vibe graphing,” 2026. [Online]. Available: https://arxiv.org/abs/2603.06007

  22. [22]

    AgentDoG: A Diagnostic Guardrail Framework for AI Agent Safety and Security

    D. Liu, Q. Ren, C. Qian, S. Shao, Y . Xie, Y . Li, Z. Yang, H. Luo, P. Wang, Q. Liu, B. Hu, L. Tang, J. Mei, D. Guo, L. Yuan, J. Yang, G. Chen, Q. Lin, Y . Yu, B. Zhang, J. Guo, J. Zhang, W. Shao, H. Deng, Z. Xi, W. Wang, W. Wang, W. Shen, Z. Chen, H. Xie, J. Tao, J. Dai, J. Ji, Z. Ba, L. Zhang, Y . Liu, Q. Zhang, L. Zhu, Z. Wei, H. Xue, C. Lu, J. Shao, a...

  23. [23]

    Think-augmented function calling: Improving llm parameter accuracy through embedded reasoning,

    L. Wei, X. Peng, J. Ou, and B. Wang, “Think-augmented function calling: Improving llm parameter accuracy through embedded reasoning,” 2026. [Online]. Available: https://arxiv.org/abs/2601. 18282

  24. [24]

    A resource-rational principle for modeling visual attention control,

    Y . Bai, “A resource-rational principle for modeling visual attention control,” 2026. [Online]. Available: https://arxiv.org/abs/2603.02056

  25. [25]

    Security Considerations for Artificial Intelligence Agents

    N. Li, K. Zhang, K. Polley, and J. Ma, “Security considerations for artificial intelligence agents,” 2026. [Online]. Available: https: //arxiv.org/abs/2603.12230

  26. [26]

    Mm-condchain: A programmatically verified benchmark for visually grounded deep compositional reasoning,

    H. Shen, S. Yan, H. Xue, S. Lu, X. Tang, G. Zhang, T. Zhao, and J. Yin, “Mm-condchain: A programmatically verified benchmark for visually grounded deep compositional reasoning,” 2026. [Online]. Available: https://arxiv.org/abs/2603.12266

  27. [27]

    Language model teams as distributed systems,

    E. Mieczkowski, K. M. Collins, I. Sucholutsky, N. V ´elez, and T. L. Griffiths, “Language model teams as distributed systems,” 2026. [Online]. Available: https://arxiv.org/abs/2603.12229

  28. [28]

    Sceneassistant: A visual feedback agent for open-vocabulary 3d scene generation,

    J. Luo, J. Tang, R. Lu, and G. Zeng, “Sceneassistant: A visual feedback agent for open-vocabulary 3d scene generation,” 2026. [Online]. Available: https://arxiv.org/abs/2603.12238

  29. [29]

    Sentinel agents for secure and trustworthy agentic ai in multi-agent systems,

    D. Gosmar and D. A. Dahl, “Sentinel agents for secure and trustworthy agentic ai in multi-agent systems,” 2025. [Online]. Available: https://arxiv.org/abs/2509.14956

  30. [30]

    A blockchain-monitored agentic ai architecture for trusted perception- reasoning-action pipelines,

    S. Jan, H. A. Razzaqi, A. Akarma, and M. R. Belgaum, “A blockchain-monitored agentic ai architecture for trusted perception- reasoning-action pipelines,” 2025. [Online]. Available: https://arxiv. org/abs/2512.20985

  31. [31]

    Toolsafe: Enhancing tool invocation safety of llm-based agents via proactive step-level guardrail and feedback,

    Y . Mou, Z. Xue, L. Li, P. Liu, S. Zhang, W. Ye, and J. Shao, “Toolsafe: Enhancing tool invocation safety of llm-based agents via proactive step-level guardrail and feedback,” 2026. [Online]. Available: https://arxiv.org/abs/2601.10156

  32. [32]

    Mindguard: Intrinsic decision inspection for securing llm agents against metadata poisoning,

    Z. Wang, H. Du, G. Shi, J. Zhang, H. Cheng, Y . Yao, K. Guo, and X.-Y . Li, “Mindguard: Intrinsic decision inspection for securing llm agents against metadata poisoning,” 2025. [Online]. Available: https://arxiv.org/abs/2508.20412

  33. [33]

    Miniscope: A least privilege framework for authorizing tool calling agents,

    J. Zhu, K. Tseng, G. Vernik, X. Huang, S. G. Patil, V . Fang, and R. A. Popa, “Miniscope: A least privilege framework for authorizing tool calling agents,” 2025. [Online]. Available: https://arxiv.org/abs/2512.11147

  34. [34]

    Trustworthy agentic ai requires deterministic architectural boundaries,

    M. Bhattarai and M. Vu, “Trustworthy agentic ai requires deterministic architectural boundaries,” 2026. [Online]. Available: https://arxiv.org/abs/2602.09947

  35. [35]

    Towards verifiably safe tool use for llm agents,

    A. Doshi, Y . Hong, C. Xu, E. Kang, A. Kapravelos, and C. K ¨astner, “Towards verifiably safe tool use for llm agents,” 2026. [Online]. Available: https://arxiv.org/abs/2601.08012

  36. [36]

    Silent egress: When implicit prompt injection makes llm agents leak without a trace,

    Q. Lan, A. Kaul, S. Jones, and S. Westrum, “Silent egress: When implicit prompt injection makes llm agents leak without a trace,”

  37. [37]

    Available: https://arxiv.org/abs/2602.22450

    [Online]. Available: https://arxiv.org/abs/2602.22450

  38. [38]

    AgentCrypt: Advancing Privacy and (Secure) Computation in AI Agent Collaboration

    H. Karthikeyan, Y . Guo, L. de Castro, A. Polychroniadou, U. M. Sehwag, L. Ardon, S. Ganesh, and M. Veloso, “Agentcrypt: Advancing privacy and (secure) computation in ai agent collaboration,” 2025. [Online]. Available: https://arxiv.org/abs/2512.08104

  39. [39]

    Prompt injection attack to tool selection in llm agents,

    J. Shi, Z. Yuan, G. Tie, P. Zhou, N. Z. Gong, and L. Sun, “Prompt injection attack to tool selection in llm agents,” 2025. [Online]. Available: https://arxiv.org/abs/2504.19793

  40. [40]

    Agent drift: Quantifying behavioral degradation in multi-agent llm systems over extended interactions,

    A. Rath, “Agent drift: Quantifying behavioral degradation in multi-agent llm systems over extended interactions,” 2026. [Online]. Available: https://arxiv.org/abs/2601.04170

  41. [41]

    Security Threat Modeling for Emerging AI-Agent Protocols: A Comparative Analysis of MCP, A2A, Agora, and ANP

    Z. Anbiaee, M. Rabbani, M. Mirani, G. Piya, I. Opushnyev, A. Ghorbani, and S. Dadkhah, “Security threat modeling for emerging ai-agent protocols: A comparative analysis of mcp, a2a, agora, and anp,” 2026. [Online]. Available: https://arxiv.org/abs/2602.11327

  42. [42]

    Agent tools orchestration leaks more: Dataset, benchmark, and mitigation,

    Y . Qiao, D. Liu, H. Yang, W. Zhou, and S. Hu, “Agent tools orchestration leaks more: Dataset, benchmark, and mitigation,” 2025. [Online]. Available: https://arxiv.org/abs/2512.16310

  43. [43]

    Xagen: An explainability tool for identifying and correcting failures in multi- agent workflows,

    X. Wang, M. Yin, E. Koh, and M. D. Dogan, “Xagen: An explainability tool for identifying and correcting failures in multi- agent workflows,” 2025. [Online]. Available: https://arxiv.org/abs/ 2512.17896

  44. [44]

    Adaptiflow: An extensible framework for event-driven autonomy in cloud microservices,

    B. A. Z. Ndadji, S. Bliudze, and C. Quinton, “Adaptiflow: An extensible framework for event-driven autonomy in cloud microservices,” 2025. [Online]. Available: https://arxiv.org/abs/2512. 23499

  45. [45]

    A2p-vis: an analyzer- to-presenter agentic pipeline for visual insights generation and reporting,

    S. Gan, R. Wang, J. Mooney, and D. Kang, “A2p-vis: an analyzer- to-presenter agentic pipeline for visual insights generation and reporting,” 2025. [Online]. Available: https://arxiv.org/abs/2512.22101

  46. [46]

    The 2025 AI Agent Index: Documenting Technical and Safety Features of Deployed Agentic AI Systems

    L. Staufer, K. Feng, K. Wei, L. Bailey, Y . Duan, M. Yang, A. P. Ozisik, S. Casper, and N. Kolt, “The 2025 ai agent index: Documenting technical and safety features of deployed agentic ai systems,” 2026. [Online]. Available: https://arxiv.org/abs/2602.17753

  47. [47]

    A survey of agentic ai and cybersecurity: Challenges, opportunities and use-case prototypes,

    S. J. Lazer, K. Aryal, M. Gupta, and E. Bertino, “A survey of agentic ai and cybersecurity: Challenges, opportunities and use-case prototypes,”

  48. [48]

    Available: https://arxiv.org/abs/2601.05293

    [Online]. Available: https://arxiv.org/abs/2601.05293

  49. [49]

    Agentic AI Security: Threats, Defenses, Evaluation, and Open Challenges

    A. Chhabra, S. Datta, S. K. Nahin, and P. Mohapatra, “Agentic ai security: Threats, defenses, evaluation, and open challenges,” 2025. [Online]. Available: https://arxiv.org/abs/2510.23883

  50. [50]

    Open Challenges in Multi-Agent Security: Towards Secure Systems of Interacting AI Agents

    C. S. de Witt, “Open challenges in multi-agent security: Towards secure systems of interacting ai agents,” 2025. [Online]. Available: https://arxiv.org/abs/2505.02077

  51. [51]

    Every keystroke you make: A tech-law measurement and analysis of event listeners for wiretapping,

    S. Munir, N. Demir, Q. Li, K. Kollnig, and Z. Shafiq, “Every keystroke you make: A tech-law measurement and analysis of event listeners for wiretapping,” 2025. [Online]. Available: https://arxiv.org/abs/2508.19825

  52. [52]

    Secure development of a hooking-based deception framework against keylogging techniques,

    M. S. I. Sajid, S. Ahmed, and R. Sosnoski, “Secure development of a hooking-based deception framework against keylogging techniques,”

  53. [53]

    Available: https://arxiv.org/abs/2508.04178

    [Online]. Available: https://arxiv.org/abs/2508.04178

  54. [54]

    Authenticated delegation and authorized ai agents,

    T. South, S. Marro, T. Hardjono, R. Mahari, C. D. Whitney, D. Greenwood, A. Chan, and A. Pentland, “Authenticated delegation and authorized ai agents,” 2025. [Online]. Available: https: //arxiv.org/abs/2501.09674

  55. [55]

    Towards automating data access permissions in ai agents,

    Y . Wu, K. Yang, F. Roesner, T. Kohno, N. Zhang, and U. Iqbal, “Towards automating data access permissions in ai agents,” 2025. [Online]. Available: https://arxiv.org/abs/2511.17959

  56. [56]

    Policy-aware generative ai for safe, auditable data access governance,

    S. A. Mandalawi, M. A. Mohammed, H. Maclean, M. C. Cakmak, and J. R. Talburt, “Policy-aware generative ai for safe, auditable data access governance,” 2025. [Online]. Available: https://arxiv.org/abs/2510.23474

  57. [57]

    Improving large language models function calling and interpretability via guided-structured templates,

    H. Dang, T. Liu, Z. Wu, J. Yang, H. Jiang, T. Yang, P. Chen, Z. Wang, H. Wang, H. Li, B. Yin, and M. Jiang, “Improving large language models function calling and interpretability via guided-structured templates,” 2025. [Online]. Available: https: //arxiv.org/abs/2509.18076

  58. [58]

    Autotool: Dynamic tool selection and integration for agentic reasoning,

    J. Zou, L. Yang, Y . Qi, S. Chen, M. Ai, K. Shen, J. He, and M. Wang, “Autotool: Dynamic tool selection and integration for agentic reasoning,” 2025. [Online]. Available: https://arxiv.org/abs/ 2512.13278

  59. [59]

    Toolrm: Outcome reward models for tool-calling large language models,

    M. Agarwal, I. Abdelaziz, K. Basu, M. Unuvar, L. A. Lastras, Y . Rizk, and P. Kapanipathi, “Toolrm: Outcome reward models for tool-calling large language models,” 2025. [Online]. Available: https://arxiv.org/abs/2509.11963

  60. [60]

    Where llm agents fail and how they can learn from failures,

    K. Zhu, Z. Liu, B. Li, M. Tian, Y . Yang, J. Zhang, P. Han, Q. Xie, F. Cui, W. Zhang, X. Ma, X. Yu, G. Ramesh, J. Wu, Z. Liu, P. Lu, J. Zou, and J. You, “Where llm agents fail and how they can learn from failures,” 2025. [Online]. Available: https://arxiv.org/abs/2509.25370

  61. [61]

    Tool preferences in agentic llms are unreliable,

    K. Faghih, W. Wang, Y . Cheng, S. Bharti, G. Sriramanan, S. Balasubramanian, P. Hosseini, and S. Feizi, “Tool preferences in agentic llms are unreliable,” 2025. [Online]. Available: https: //arxiv.org/abs/2505.18135

  62. [62]

    From allies to adversaries: Manipulating llm tool-calling through adversarial injection,

    H. Wang, R. Zhang, J. Wang, M. Li, Y . Huang, D. Wang, and Q. Wang, “From allies to adversaries: Manipulating llm tool-calling through adversarial injection,” 2024. [Online]. Available: https://arxiv.org/abs/2412.10198

  63. [63]

    From prompt injections to protocol exploits: Threats in llm-powered ai agents workflows,

    M. A. Ferrag, N. Tihanyi, D. Hamouda, L. Maglaras, A. Lakas, and M. Debbah, “From prompt injections to protocol exploits: Threats in llm-powered ai agents workflows,” 2025. [Online]. Available: https://arxiv.org/abs/2506.23260

  64. [64]

    Attractive metadata attack: Inducing llm agents to invoke malicious tools,

    K. Mo, L. Hu, Y . Long, and Z. Li, “Attractive metadata attack: Inducing llm agents to invoke malicious tools,” 2025. [Online]. Available: https://arxiv.org/abs/2508.02110

  65. [65]

    Bridging ai and software security: A comparative vulnerability assessment of llm agent deployment paradigms,

    T. Gasmi, R. Guesmi, I. Belhadj, and J. Bennaceur, “Bridging ai and software security: A comparative vulnerability assessment of llm agent deployment paradigms,” 2025. [Online]. Available: https://arxiv.org/abs/2507.06323

  66. [66]

    Ai ides or autonomous agents? measuring the impact of coding agents on software development,

    S. Agarwal, H. He, and B. Vasilescu, “Ai ides or autonomous agents? measuring the impact of coding agents on software development,”

  67. [67]

    Available: https://arxiv.org/abs/2601.13597

    [Online]. Available: https://arxiv.org/abs/2601.13597

  68. [68]

    The attack and defense landscape of agentic ai: A comprehensive survey,

    J. Kim, X. Liu, Z. Wang, S. Qiu, B. Li, W. Guo, and D. Song, “The attack and defense landscape of agentic ai: A comprehensive survey,” 2026. [Online]. Available: https://arxiv.org/abs/2603.11088

  69. [69]

    Mind the metrics: Patterns for telemetry-aware in-ide ai application development using the model context protocol (mcp),

    V . Koc, J. Verre, D. Blank, and A. Morgan, “Mind the metrics: Patterns for telemetry-aware in-ide ai application development using the model context protocol (mcp),” 2025. [Online]. Available: https://arxiv.org/abs/2506.11019

  70. [70]

    Exploring the challenges and opportunities of ai-assisted codebase generation,

    P. Eibl, S. Sabouri, and S. Chattopadhyay, “Exploring the challenges and opportunities of ai-assisted codebase generation,” 2025. [Online]. Available: https://arxiv.org/abs/2508.07966

  71. [71]

    Ai agent systems: Architectures, applications, and evaluation,

    B. Xu, “Ai agent systems: Architectures, applications, and evaluation,”

  72. [72]

    Available: https://arxiv.org/abs/2601.01743

    [Online]. Available: https://arxiv.org/abs/2601.01743

  73. [73]

    A safety and security framework for real-world agentic systems,

    S. Ghosh, B. Simkin, K. Shiarlis, S. Nandi, D. Zhao, M. Fiedler, J. Bazinska, N. Pope, R. Prabhu, D. Rohrer, M. Demoret, and B. Richardson, “A safety and security framework for real-world agentic systems,” 2025. [Online]. Available: https: //arxiv.org/abs/2511.21990

  74. [74]

    Task-aware delegation cues for llm agents,

    X. Gu, “Task-aware delegation cues for llm agents,” 2026. [Online]. Available: https://arxiv.org/abs/2603.11011

  75. [75]

    Verified multi-agent orchestration: A plan-execute-verify-replan framework for complex query resolution,

    X. Zhang, Y . Cui, G. Wang, W. Qiu, Z. Li, F. Han, Y . Huang, H. Qiu, B. Zhu, and P. He, “Verified multi-agent orchestration: A plan-execute-verify-replan framework for complex query resolution,”

  76. [76]

    Available: https://arxiv.org/abs/2603.11445

    [Online]. Available: https://arxiv.org/abs/2603.11445

  77. [77]

    Agenttrace: A structured logging framework for agent system observability,

    A. AlSayyad, K. Y . Huang, and R. Pal, “Agenttrace: A structured logging framework for agent system observability,” 2026. [Online]. Available: https://arxiv.org/abs/2602.10133

  78. [78]

    Agentsight: System-level observability for ai agents using ebpf,

    Y . Zheng, Y . Hu, T. Yu, and A. Quinn, “Agentsight: System-level observability for ai agents using ebpf,” 2025. [Online]. Available: https://arxiv.org/abs/2508.02736

  79. [79]

    Zero-trust runtime verification for agentic payment protocols: Mitigating replay and context-binding failures in ap2,

    Q. Lan, A. Kaul, S. Jones, and S. Westrum, “Zero-trust runtime verification for agentic payment protocols: Mitigating replay and context-binding failures in ap2,” 2026. [Online]. Available: https://arxiv.org/abs/2602.06345

  80. [80]

    Trajectory-informed memory generation for self-improving agent systems,

    G. Fang, V . Isahagian, K. R. Jayaram, R. Kumar, V . Muthusamy, P. Oum, and G. Thomas, “Trajectory-informed memory generation for self-improving agent systems,” 2026. [Online]. Available: https://arxiv.org/abs/2603.10600

Showing first 80 references.