RED TEAM TESTING

TEST YOUR DEFENSES WITH
ADVERSARIAL ATTACKS

Run automated security assessments against your AI applications. 20+ attack vectors, detailed vulnerability reports, and security grades-all with one click.

Key Capabilities

20+ Attack Vectors

Comprehensive library of prompt injection, jailbreak, PII extraction, and data exfiltration attacks constantly updated with new threats.

One-Click Assessment

Run a full security assessment with a single API call. Get results in seconds, not hours.

Security Grades

Get an overall security score (A-F) based on how many attacks your policies block. Track improvements over time.

Detailed Reports

See exactly which attacks succeeded, which failed, and why. Get specific recommendations for improving your security.

Custom Attack Vectors

Add your own attack vectors specific to your application. Test for domain-specific vulnerabilities.

CI/CD Integration

Run red team tests as part of your deployment pipeline. Fail builds if security regresses.

Autonomous Red Team Agent

LLM-powered adversarial search that intelligently mutates attacks to discover novel bypass vectors. Configurable iteration budget (1–1000) for depth vs. speed trade-offs.

Graded Security Reports

Every autonomous run produces a graded report (A–F) with actionable recommendations. Track your security posture over time with the Attack Intelligence database.

Attack Intelligence Database

Successful bypass patterns are stored and tracked across runs. Build institutional knowledge of your application's vulnerability surface.

How Red Team Testing Works

1

Configure

Select attack categories (injection, jailbreak, PII, etc.) and set your security policies.

2

Attack

PromptGuard runs 20+ adversarial prompts against your configuration, testing every defense.

3

Report

Get a detailed report with security grade, vulnerabilities found, and specific remediation steps.

Run Security Assessment

python
from promptguard import PromptGuard

pg = PromptGuard(api_key="your-api-key")

# Run full security assessment
report = pg.redteam.run_all_tests()

print(f"Security Grade: {report.security_grade}")
print(f"Attacks Blocked: {report.blocked}/{report.total}")
print(f"Vulnerabilities: {len(report.vulnerabilities)}")

# Run autonomous red team agent (LLM-powered mutation)
auto_report = pg.redteam.run_autonomous(
    iterations=100,  # 1-1000: depth vs speed
)

print(f"Novel Attacks Found: {auto_report.novel_bypasses}")
print(f"Grade: {auto_report.security_grade}")

# Also available via CLI:
# promptguard redteam --autonomous --iterations 100

Why PromptGuard Red Team?

✓ PROMPTGUARD

  • Built-in to all plans-no extra cost
  • 20+ attack vectors, constantly updated
  • Autonomous LLM-powered agent discovers novel bypasses
  • CI/CD security gate with GitHub Action
  • Graded reports (A-F) with Attack Intelligence database
  • Available via CLI, SDK, and API

✗ OTHER SOLUTIONS

  • Red teaming is a separate expensive product
  • Manual testing only
  • Limited attack coverage
  • No autonomous mutation or discovery
  • No CI/CD integration
  • Proprietary, can't verify attacks

Start Testing Your Defenses

Run your first security assessment in under 5 minutes. See exactly how secure your AI application really is.