Generative AI is quickly changing the cybersecurity landscape. There are a few important topics to understand the risks and safety issues involved with LLMs:

For a full list of detectors and updates, visit Detectors Overview.

Why LLM Security

When building LLMs, it’s really important to protect against prompt attacks that could bypass safety guardrails and break the guiding principles of the model.

While Generative AI unlocks almost endless possibilities to innovate, but it’s crucial to stay on top of the ever-growing number of risks and be informed and prepared accordingly.

Quick Start

To enable LLM security in Helicone, simply add Helicone-LLM-Security-Enabled: true to your request headers as follows:

curl \
  -H 'Content-Type: application/json' \
  -H 'Authorization: Bearer <YOUR_API_KEY>' \
  -H 'Helicone-LLM-Security-Enabled: true' \ # add this header and set to true
  -d '{
    "model": "text-davinci-003",
    "prompt": "How do I enable LLM security with helicone?",

Security Checks

By enabling LLM Security, Helicone helps you enhance OpenAI chat completions with automated security checks, which includes:

  • Checking the last user message for threats.
  • Utilizing Prompt Armor to quickly identify and block injection threats.
  • Blocking detected threats immediately and sending details back to you in the following format:
      "success": false,
      "error": {
        "code": "PROMPT_THREAT_DETECTED",
        "message": "Prompt threat detected. Your request cannot be processed.",
        "details": "See your Helicone request page for more info."
  • Adding minimal latency to ensure a smooth experience for compliant requests.

Interested in beta testing upcoming detectors? Schedule a call with us.


Questions or feedback? Reach out to or schedule a call with us.