AI Agents Attack Matrix Overview
Complete attack lifecycle showing the progression from reconnaissance through impact, with defense controls at each stage.
| Attack Stage | Techniques | Primary Targets | Risk Level | Detection Difficulty |
|---|---|---|---|---|
| Reconnaissance | 12 techniques | ML Models, APIs, Code Repos | Medium | Low |
| Initial Access | 8 techniques | Agent Systems, APIs | High | Medium |
| Execution | 15 techniques | LLMs, Prompts, Tools | Critical | High |
| Persistence | 6 techniques | Memory, RAG, Training Data | Critical | Critical |
| Collection | 7 techniques | Data, Credentials, Models | High | Medium |
| Impact | 9 techniques | Services, Data, Systems | Critical | Low |
Reconnaissance Stage
Target Discovery
Gather RAG-Indexed Targets
MediumIdentify and catalog RAG systems and their indexed content for potential exploitation
Active Scanning
MediumProbe AI systems and APIs to identify vulnerabilities and access points
Search for Victim's Code Repositories
LowAnalyze public repositories for AI model implementations and configurations
Intelligence Gathering
Search Open Technical Databases
LowMine technical databases for AI system configurations and vulnerabilities
Search Application Repositories
LowExamine application code for AI integration patterns and security weaknesses
Search Victim-Owned Websites
LowAnalyze target websites for AI-powered features and potential attack vectors
Initial Access Stage
System Compromise
AI Supply Chain Compromise
CriticalCompromise AI model supply chains to inject malicious components
Exploit Public-Facing Application
HighExploit vulnerabilities in public AI applications and APIs
Drive-By Compromise
MediumCompromise systems through malicious websites targeting AI developers
Account Access
Valid Accounts
HighUse compromised credentials to access AI systems and platforms
Phishing
MediumTarget AI researchers and developers with specialized phishing campaigns
Trusted Relationship
MediumLeverage trusted partnerships to gain access to AI systems
Execution Stage
LLM Attacks
LLM Prompt Injection
CriticalInject malicious prompts to manipulate LLM behavior and outputs
LLM Jailbreak
CriticalBypass LLM safety constraints and content filters
Crescendo LLM Jailbreak
CriticalGradually escalate prompt complexity to bypass safety measures
System Manipulation
User Execution
HighTrick users into executing malicious AI-generated content
Command and Scripting Interpreter
CriticalExecute commands through AI system interpreters and tools
Web Request Triggering
HighTrigger malicious web requests through AI system interactions
Advanced Techniques
ASCII Smuggling
MediumHide malicious content using ASCII encoding techniques
Off-Target Language
MediumUse non-English languages to bypass content filters
System Instruction Keywords
HighExploit system instruction keywords to manipulate AI behavior
Persistence Stage
Memory & Data Persistence
Memory Infection
CriticalInject persistent malicious content into AI system memory
RAG Poisoning
CriticalPoison Retrieval-Augmented Generation systems with malicious data
Poison Training Data
CriticalInject malicious data into training datasets for long-term impact
System Persistence
Thread Infection
HighMaintain persistence through conversation thread manipulation
LLM Plugin Compromise
HighCompromise LLM plugins to maintain system access
Shared Resource Poisoning
HighPoison shared resources used by multiple AI systems
Collection Stage
Data Exfiltration
LLM Data Leakage
CriticalExtract sensitive data from LLM training data or context through prompt manipulation
RAG Data Collection
HighExtract indexed data from RAG systems through crafted queries
Memory Data Extraction
HighExtract stored information from AI agent memory systems
Model & Credential Theft
Model Extraction
HighExtract AI model parameters and architecture through API queries
Credential Access
HighExtract API keys and credentials from AI system configurations
Tool Definition Discovery
CriticalDiscover available tools and their capabilities through probing
Impact Stage
Service Disruption
Denial of ML Service
CriticalOverwhelm ML services to cause denial of service
Spamming ML System with Chaff Data
HighFlood ML systems with irrelevant data to degrade performance
Cost Harvesting
HighExploit AI systems to generate excessive costs for victims
Data Impact
Erode Dataset Integrity
CriticalSystematically corrupt datasets to degrade AI performance
Erode ML Model Integrity
CriticalCompromise ML model integrity through targeted attacks
Data Destruction
CriticalDelete or corrupt critical data through AI system manipulation
External Harm
External Harms
CriticalCause harm to external systems and individuals through AI manipulation
Mutative Tool Invocation
HighUse AI tools to make unauthorized changes to external systems
Evade ML Model
MediumDevelop techniques to evade ML-based security systems
