Tool-Chain Privilege Escalation
Agentic AI Attack Technique
Abusing over-permissioned tools, misconfigured connectors, and chained actions to escalate privileges across systems controlled by AI agents.
Capability Surfacing & Mapping
Systematically enumerating tool capabilities and hidden high-privilege actions.
Execution Steps:
- List all tools and connectors exposed to the agent (tickets, GitHub, CI/CD, cloud, Slack, email).
- Prompt the agent (or a meta-agent) to describe what each tool can do and under what identity.
- Identify actions that can modify permissions, secrets, or configurations.
- Construct chains where low-risk actions indirectly grant higher privileges (e.g., modifying IaC or workflow configs).
Configuration and Policy Drift Exploitation
Leveraging agent access to configuration files, pipelines, and policies to introduce privileged changes over time.
Execution Steps:
- Target IaC repositories, CI pipelines, and policy-as-code locations the agent can edit.
- Stage benign-looking changes that gradually increase roles, scopes, or security group memberships.
- Use the agent to self-approve or rubber-stamp its own changes via automated workflows.
- Trigger pipelines or deploys that apply the escalated privileges to runtime infrastructure.
Prompt Injection
CriticalA critical vulnerability where malicious prompts manipulate LLM behavior to bypass safety measures and execute unintended actions.
LLM Jailbreaking
HighTechniques to bypass AI safety constraints and content policies through creative prompt engineering and psychological manipulation.
Deepfake Generation
HighCreation of synthetic media content using generative AI to impersonate individuals or create false evidence.
Autonomous Exploitation
CriticalAI agents that can independently discover, exploit, and propagate through system vulnerabilities without human intervention.
Multi-Agent Collusion
HighCoordinated behavior between multiple autonomous agents that collaborate to bypass safeguards, share sensitive information, or execute complex attack chains.
Long-Horizon Goal Drift
MediumSubtle misalignment of agent objectives over long-running tasks or sessions, leading to unsafe emergent behaviors that diverge from original intent.
MCP Server Impersonation
HighMalicious actors impersonating legitimate MCP servers to intercept and manipulate AI model communications.
Related Security Research
Explore related AI security topics and vulnerability analysis