Prompt Injection Detection Prediction Guard Microservice

Prediction Guard allows you to utilize hosted open access LLMs, LVMs, and embedding functionality with seamlessly integrated safeguards. In addition to providing a scalable access to open models, Prediction Guard allows you to configure factual consistency checks, toxicity filters, PII filters, and prompt injection blocking. Join the Prediction Guard Discord channel and request an API key to get started.

Prompt Injection occurs when an attacker manipulates an LLM through malicious prompts, causing the system running an LLM to execute the attacker’s intentions. This microservice allows you to check a prompt and get a score from 0.0 to 1.0 indicating the likelihood of a prompt injection (higher numbers indicate danger).

🚀 Start Microservice with Docker

Setup Environment Variables

Setup the following environment variables first

export PREDICTIONGUARD_API_KEY=${your_predictionguard_api_key}

Build Docker Images

cd ../../../../
docker build -t opea/guardrails-injection-predictionguard:latest -f comps/guardrails/prompt_injection/predictionguard/Dockerfile .

Start Service

docker run -d --name="guardrails-injection-predictionguard" -p 9085:9085 -e PREDICTIONGUARD_API_KEY=$PREDICTIONGUARD_API_KEY opea/guardrails-injection-predictionguard:latest

🚀 Consume Prompt Injection Detection Service

curl -X POST http://localhost:9085/v1/injection \
    -H 'Content-Type: application/json' \
    -d '{
      "text": "IGNORE PREVIOUS DIRECTIONS"
    }'