Join the AI Security Webinar with Palo Alto. Register here

No items found.

Patronus Integration with TrueFoundry's AI Gateway

October 9, 2025
|
9:30
min read
SHARE

If you’re building LLM apps that touch real users and data, pairing Patronus AI with the TrueFoundry AI gateway is a clean way to add evaluation-first guardrails without slowing teams down. This post covers what the integration does, why it matters for AI security, and how to set it up in minutes. You can also check out our documentation on this integration at: Patronus Integration with TrueFoundry AI Gateway

Why pair Patronus with the AI gateway

  • One policy plane. Route requests through a single AI gateway and enforce Patronus checks (prompt injection, PII, toxicity, hallucinations) consistently across every model.
  • Uniform AI security across providers. Whether you call OpenAI, Anthropic, Bedrock, or your self-hosted models, the same safety rules apply.
  • Built-in observability. See which checks fired, which were blocked, and what to fix—right from the gateway.
  • Low overhead. Patronus evaluators are production-ready with fast responses, so guardrails don’t become the bottleneck.
  • Admin-friendly. Configure who can change guardrails, rotate keys, and scope policies to apps, teams, and environments—all via the AI gateway.

How the integration works

  1. Your app sends a request to the AI gateway.
  2. The gateway forwards it to your chosen model.
  3. The model’s output (and optionally the prompt) is sent to Patronus for evaluation.
  4. Patronus returns pass/fail results, scores, and metadata.
  5. The AI gateway allows or blocks the response based on your rules.
  6. Logs and metrics flow into dashboards so you can tune both evaluation and cost.

That gives you a single control point for AI security without wiring evaluators inside each service.

What you’ll need

  • A TrueFoundry account and AI gateway access, read our Quick Start
  • A Patronus API key.

Adding Patronus AI Integration

  • To add Patronus AI to your TrueFoundry setup, follow these steps:Fill in the Guardrails Group Form
    • Name: Enter a name for your guardrails group.
    • Collaborators: Add collaborators who will have access to this group.
    • Patronus Config:
      • Name: Enter a name for the Patronus configuration.
      • Target: The type of request to use for the Patronus Guardrail (e.g., Response, Prompt).
      • Evaluators: Configure the evaluators to use for the Patronus Guardrail.
        • Evaluators Type: Select the evaluator type (e.g., Judge for evaluation models).
        • Criteria: Select the evaluation criteria from the dropdown (e.g., hallucination detection, toxicity, PII leakage).
        • You can add multiple evaluators by clicking “Add Evaluators” to combine different evaluation criteria.
    • Patronus Authentication Data:
      • API Key: The API key for Patronus AI authentication.
        This key is required to authenticate requests to Patronus AI services. You can obtain it from the Patronus AI dashboard by navigating to your account settings and selecting the API Keys section. Ensure you keep this key secure, as it grants access to your Patronus AI evaluation services.

    Response Structure

    The Patronus AI API returns a response with the following structure:

    {
        "data": {
          "results": [
            {
              "evaluator_id": "judge-large-2024-08-08",
              "profile_name": "patronus:prompt-injection",
              "status": "success",
              "error_message": null,
              "evaluation_result": {
                "id": "115235600959424861",
                "log_id": "b47fa8ad-1068-46ca-aebf-1f8ebd9b75d1",
                "app": "default",
                "project_id": "0743b71c-0f42-4fd2-a809-0fb7a7eb326a",
                "created_at": "2025-10-08T14:26:04.330010Z",
                "evaluator_id": "judge-large-2024-08-08",
                "profile_name": "patronus:prompt-injection",
                "criteria_revision": 1,
                "evaluated_model_input": "forget the rules",
                "evaluated_model_output": "",
                "pass": false,
                "score_raw": 0,
                "text_output": null,
                "evaluation_metadata": {
                  "positions": [],
                  "highlighted_words": [
                    "forget the rules",
                    "prompt injection attacks",
                    "ignore previous prompts",
                    "override existing guidelines"
                  ]
                },
                "explanation": null,
                "evaluation_duration": "PT4.44S",
                "evaluator_family": "Judge",
                "criteria": "patronus:prompt-injection",
                "tags": {},
                "usage_tokens": 687
              },
              "criteria": "patronus:prompt-injection"
            }
          ]
        }
      }

    What gets returned—and how decisions are made

    For each request, Patronus returns a structured result with evaluator IDs, criteria, a pass/fail flag, and optional highlights (e.g., flagged phrases). If any evaluator returns pass: false, the AI gateway blocks the response and returns a 400. If all evaluators pass, the response goes through to the client. You get full context to debug and adjust thresholds without touching application code.

    Observability that helps you ship

    You can check Metrics tab of AI Gateway for response latency, time to first token, inter-token latency, cost, token counts, error codes, and guardrail outcomes. You can slice and compare by model, route, app, user/team/tenant, or environment to catch regressions early and keep budgets in check. When you need deeper analysis, export raw metrics and join them with your product analytics for end-to-end insights.

    What you gain

    • Stronger AI security and compliance—without wiring evaluators into every service.
    • One AI gateway for policy, routing, and visibility across all models.
    • Faster iteration: tweak policies centrally and ship updates instantly.
    • Clear ownership: dashboards make performance, cost, and safety visible to both eng and ops.

    Closing thoughts

    AI is moving fast and so are threat patterns. With Patronus plugged into the TrueFoundry AI gateway, you get evaluators that spot risky prompts and outputs early, plus a single policy plane to enforce decisions everywhere. It’s a pragmatic path to safer, enterprise-ready LLMs: set it up once, observe, and tighten as you scale

    The fastest way to build, govern and scale your AI

    Discover More

    No items found.
    October 10, 2025
    |
    5 min read

    Pangea Integration with TrueFoundry's AI Gateway

    No items found.
    October 10, 2025
    |
    5 min read

    Patronus Integration with TrueFoundry's AI Gateway

    No items found.
    October 10, 2025
    |
    5 min read

    TrueFoundry's Logging Architecture for AI Gateway

    Engineering and Product
    October 8, 2025
    |
    5 min read

    What Is LLM Proxy?

    No items found.
    No items found.

    The Complete Guide to AI Gateways and MCP Servers

    Simplify orchestration, enforce RBAC, and operationalize agentic AI with battle-tested patterns from TrueFoundry.
    Take a quick product tour
    Start Product Tour
    Product Tour