LLM Security
Enable robust security measures in your LLM applications to protect against prompt injections, detect anomalies, and prevent data exfiltration.
Introduction
Generative AI is quickly changing the cybersecurity landscape. There are a few important topics to understand the risks and safety issues involved with LLMs:
For a full list of detectors and updates, visit Detectors Overview.
Why LLM Security
When building LLMs, it’s really important to protect against prompt attacks that could bypass safety guardrails and break the guiding principles of the model.
While Generative AI unlocks almost endless possibilities to innovate, but it’s crucial to stay on top of the ever-growing number of risks and be informed and prepared accordingly.
Quick Start
To enable LLM security in Helicone, simply add Helicone-LLM-Security-Enabled: true
to your request headers as follows:
Security Checks
By enabling LLM Security, Helicone helps you enhance OpenAI chat completions with automated security checks, which includes:
- Checking the last user message for threats.
- Utilizing Prompt Armor to quickly identify and block injection threats.
- Blocking detected threats immediately and sending details back to you in the following format:
- Adding minimal latency to ensure a smooth experience for compliant requests.
Interested in beta testing upcoming detectors? Schedule a call with us.
Was this page helpful?