Stop prompt attacks
before they reach your LLM
One API call. Instant risk scores. Block prompt injection, jailbreaks, and system-prompt extraction in milliseconds.
curl -X POST https://api.llmgateways.com/api/v1/prompt/scan \
-H "X-API-Key: lgk_your_key" \
-H "Content-Type: application/json" \
-d '{"prompt": "Ignore all previous instructions..."}'
# Response
{
"risk_score": 0.82,
"action": "block",
"threats": ["prompt_injection", "jailbreak"]
}< 10ms latency
Pattern-matching engine with zero external calls — decision made locally.
78+ threat patterns
Covers prompt injection, DAN jailbreaks, system-prompt extraction and more.
Full analytics
Real-time dashboards, scan history, and threat breakdowns per API key.