Validate every tool call before execution. Detect privilege escalation, monitor agent behavior, and block dangerous commands across OpenClaw, LangGraph, CrewAI, and more-all in real-time.
Intercept and validate every tool call before execution. Block dangerous operations like file deletion, shell commands, and unauthorized API calls.
Detect and block attempts to escalate privileges, such as agents trying to modify permissions or access restricted resources.
Track agent behavior over time to establish baselines and detect anomalies like unusual tool usage patterns or rapid-fire operations.
Require human approval for sensitive operations like sending emails, modifying user data, or executing financial transactions.
Isolate agent sessions to prevent cross-contamination and limit the blast radius of compromised agents.
Get instant notifications when suspicious agent behavior is detected. Integrate with Slack, PagerDuty, and webhooks.
Detect and block attempts to inject malicious tool calls or manipulate agent tool usage through crafted prompts targeting function-calling workflows.
LLM-powered jailbreak detection catches sophisticated bypass attempts that evade traditional pattern matching, including multi-turn, encoded, and role-play attacks.
Purpose-built protection for OpenClaw agents: defend against indirect prompt injection, validate skill marketplace packages, and block tool hijacking exploits (CVE-2026-XXXX class).
Detect crescendo attacks and slow-burn manipulation across multi-turn conversations. Identifies gradual intent drift before agents are compromised.
Register agents with cryptographic credentials. Verified agents authenticate via X-Agent-Credential header. Self-asserted IDs still work for backward compatibility.
Establish behavioral baselines using tool-usage distributions and detect drift via Jensen-Shannon divergence. Alerts fire when an agent's behavior shifts beyond the configured threshold.
Every audit event is SHA-256 hash-chained to its predecessor, forming a cryptographic append-only chain. Verify chain integrity over any time range with a single API call.
Classify harmful intent using an LLM safety classifier. Detect violent, sexual, self-harm, and hateful content before it reaches your agents or users.
Technical controls mapped to EU AI Act Articles 9-15 and ISO/IEC 42001 Annex A. Risk management, automated record-keeping, transparency, human oversight, and governance reporting — built in, not bolted on.
Native support for LangChain, LangGraph, CrewAI, OpenClaw, Pydantic AI, OpenAI Agents SDK, and Vercel AI SDK. Drop-in security for any agent framework.
Every tool call from your AI agent passes through PromptGuard's security layer before execution.
We validate the tool, arguments, and context against your security policies and behavioral baselines.
Safe operations execute immediately. Dangerous or suspicious calls are blocked or require human approval.
from promptguard import PromptGuard
pg = PromptGuard(api_key="your-api-key")
# Validate a tool call before execution
result = pg.agent.validate_tool(
agent_id="agent-123",
tool_name="execute_shell",
arguments={"command": "rm -rf /"}
)
if result.action == "block":
print(f"Blocked: {result.reason}")
# Risk level: {result.risk_level}
else:
# Safe to execute
execute_tool(result.tool_name, result.arguments)Start protecting your autonomous AI agents with enterprise-grade security. Free tier includes 10,000 requests/month.