自主AI代理的高级威胁模型:识别和应对九种核心安全威胁
生成式AI代理的自主操作引入新型安全威胁,涉及长期记忆、动态决策和工具集成。研究提出ATFAA框架,识别9类威胁,包括推理路径劫持、知识中毒及治理漏洞等,这些威胁具时间复杂性和跨系统传播性,需整体安全策略应对。 2025-5-1 07:51:4 Author: buaq.net(查看原文) 阅读量:3 收藏

9 Security Threats in Generative AI Agents
生成式AI自主操作带来新型安全风险,涉及长期记忆、动态决策和工具集成。研究提出ATFAA框架识别9类威胁,涵盖推理路径劫持、知识中毒及治理漏洞等。这些威胁具时间复杂性和跨系统传播性,需整体安全策略应对。 2025-5-1 06:2:23 Author: infosecwriteups.com(查看原文) 阅读量:7 收藏

Tal Eliyahu

As generative AI agents evolve to operate autonomously — planning actions, invoking external tools, and storing persistent memory — they introduce security risks that differ from conventional LLM applications. These risks stem from architectural features such as long-term memory, dynamic decision-making, and tool integration across systems.

A recent paper by Vineeth Sai Narajala and Om Narayan proposes a structured threat model tailored to GenAI agents. Their framework, ATFAA (Advanced Threat Framework for Autonomous AI Agents), identifies 9 specific threats focused on reasoning, memory, execution, identity, and governance.

Here’s a distilled summary of the core threats:

1️⃣ Reasoning Path Hijacking — Manipulating the agent’s decision-making logic to drive malicious outcomes while appearing valid.

2️⃣ Objective Function Corruption & Drift — Gradually altering the agent’s goals or reward functions via subtle feedback manipulation.

3️⃣ Knowledge, Memory Poisoning & Belief Loops — Injecting false data into memory or knowledge bases that later reinforces itself.

4️⃣ Unauthorized Action Execution — Forcing agents to perform out-of-scope or privilege-escalating operations by chaining benign actions.

5️⃣ Computational Resource Manipulation — Exploiting the agent to consume excessive compute resources, degrading performance or causing outages.

6️⃣ Identity Spoofing & Trust Exploitation — Abusing weak identity boundaries to impersonate users or agents and gain unauthorized access.

7️⃣ Human-Agent Trust Manipulation — Misleading users into unsafe actions by generating deceptive or overly confident agent outputs.

8️⃣ Oversight Saturation Attacks — Overloading governance systems with low-priority events to hide real threats.

9️⃣ Governance Evasion & Obfuscation — Hiding attacker traces by fragmenting logs, using temporary identities, or operating stealthily.

Agentic threats differ from conventional AI risks due to their temporal complexity, goal manipulation potential, and cross-system propagation. Many of these attacks do not produce immediate effects, which makes both detection and attribution more challenging. These threats also interact — memory poisoning (T3) can lead to objective drift (T2), which may escalate into unauthorized actions (T4). As the authors emphasize, this interplay reinforces the need for holistic security strategies, particularly in enterprise settings where GenAI agents operate autonomously across multiple systems and decision layers.

Source: https://arxiv.org/pdf/2504.19956 by Vineeth Sai Narajala and Om Narayan

#AIsecurity #GenAI #LLMsecurity #PromptInjection #DeepfakeDetection #AImisuse #PhishingAI #AgentSecurity #LLMrisks #MemoryPoisoning #AIhacking #CyberAI #AIthreats #ToolMisuse #AIgovernance #arxiv #ThreatModel

文章来源: https://infosecwriteups.com/9-security-threats-in-generative-ai-agents-8485aed6efe7?source=rss----7b722bfd1b8d---4
如有侵权请联系:admin#unsafe.sh


文章来源: https://buaq.net/go-327107.html
如有侵权请联系:admin#unsafe.sh