Skip to content
Home AI Security Tools Lakera Guard
Lakera Guard

Lakera Guard

NEW ACQUIRED
Category: AI Security
License: Commercial (with Free tier)
Suphi Cankurt
Suphi Cankurt
AppSec Enthusiast
Updated February 10, 2026
4 min read
Key Takeaways
  • Real-time API that blocks prompt injection with 98%+ detection and sub-50ms latency
  • Supports 100+ languages; acquired by Check Point in 2025
  • Creators of Gandalf, the prompt injection game played by 1M+ people
  • Commercial with free tier; single API endpoint integration

Lakera Guard is an AI security API that protects LLM applications against prompt injection, jailbreaks, and data leakage in real time.

Lakera Guard real-time visibility dashboard showing threat detection across applications

Lakera was founded in 2021 in Zurich by David Haber (CEO), Dr. Mateo Rojas-Carulla, and Dr. Matthias Kraft — AI researchers with backgrounds at Google and Meta. The company has 11 PhDs on staff. Lakera gained widespread recognition for creating Gandalf, an educational game where players try to extract a secret password from an AI through prompt injection. Gandalf has attracted over 1 million players and generated 80M+ adversarial prompts that feed directly into Lakera’s threat intelligence.

In September 2025, Check Point announced the acquisition of Lakera to form a Global Center of Excellence for AI Security in Zurich. The integration brings Lakera’s technology into Check Point’s Infinity Platform, CloudGuard WAF, and GenAI Protect products.

What is Lakera Guard?

Lakera Guard sits between users and LLMs as a security layer. Every input and output passes through Guard’s detection engine before reaching the model. If a threat is detected, it flags or blocks the request before the LLM processes it.

The system delivers 98%+ detection rates with sub-50ms latency and false positive rates below 0.5%. It screens content across 100+ languages and scripts. The detection models learn from 100K+ new adversarial samples each day, drawn partly from the Gandalf community’s 80M+ prompts.

Prompt Attack Detection
Detects and blocks direct prompt injection, indirect prompt injection, jailbreak attempts, and system prompt extraction across 100+ languages in real time.
Data Leakage Prevention
PII detection and redaction, secrets detection, and custom data pattern matching for both inputs and outputs. Prevents sensitive information from reaching or leaving the LLM.
Content Moderation
Filters toxic, hateful, violent, and inappropriate content. Supports custom content policies and profanity detection. Identifies suspicious URLs outside approved domain lists.

Key Features

FeatureDetails
Prompt Injection DetectionDirect injection, indirect injection, jailbreak, system prompt extraction
Detection Rate98%+ across all attack types
LatencySub-50ms per request
False Positive RateBelow 0.5% in production
Language Support100+ languages and scripts
PII DetectionIdentifies and redacts personal data in inputs and outputs
Content ModerationToxicity, hate speech, violence, custom policies
Link ScanningFlags suspicious URLs outside approved domain lists
API FormatOpenAI-compatible chat completions message format
Scale1M+ secured transactions per app per day

How the API works

Lakera Guard uses a single endpoint: POST https://api.lakera.ai/v2/guard. Requests follow the OpenAI chat completions message format with roles (system, user, assistant). Guard screens the last interaction in the message chain and returns a flagged boolean indicating whether a threat was detected.

You can configure projects with specific policies that determine which detectors run. The breakdown parameter returns per-detector flagging details, and the payload parameter returns match locations for PII and profanity.

The four main detection categories are:

  • Prompt attacksprompt injections, jailbreaks, and manipulation attempts
  • Data leakage — PII and sensitive information exposure
  • Content violation — offensive, hateful, sexual, violent, or vulgar material
  • Unknown links — suspicious URLs outside approved domain lists

Lakera Red

Lakera Red is the company’s AI red teaming product. It runs automated attack simulations against your LLM applications to identify vulnerabilities before they reach production. Red teaming results feed back into Guard’s detection models.

Lakera Guard threat detection and response interface

Gandalf

Gandalf is Lakera’s interactive game where players try to extract a secret password from an AI through increasingly sophisticated prompt injection techniques. It demonstrates real-world attack patterns and has been used by security researchers, AI engineers, educational institutions, and CTF competitions.

The 80M+ adversarial prompts collected through Gandalf form a unique dataset that informs Lakera’s threat intelligence. The game is free to play at gandalf.lakera.ai.

Check Point acquisition
Check Point acquired Lakera in 2025. Lakera’s technology is being integrated into Check Point CloudGuard WAF (for AI-enabled applications) and Check Point GenAI Protect (for user traffic to GenAI apps). Lakera Guard remains available as a standalone API.

Getting Started

1
Create an account — Sign up at platform.lakera.ai. A free tier is available to get started.
2
Set up a project — Create a project to get a project ID and configure which detectors and policies to apply. Lakera recommends setting up a project rather than using the default policy.
3
Integrate the API — Add a single API call to your application. Send a POST request to https://api.lakera.ai/v2/guard with your messages in OpenAI chat completions format. If flagged is true, block the request.
4
Monitor and tune — Use the Security Center dashboard to monitor threats, review analytics, and adjust policies. Feed log data into your SIEM via Grafana, Splunk, or similar integrations.

When to use Lakera Guard

Lakera Guard is built for teams deploying LLM-powered applications that need real-time input/output screening. The API-first design means integration takes minutes rather than weeks. It works with any LLM — OpenAI, Anthropic, Google, Azure OpenAI, AWS Bedrock, or self-hosted models.

The platform handles high-volume production traffic (1M+ transactions per app per day) with sub-50ms latency, making it practical for customer-facing chatbots and real-time applications.

Best for
Teams deploying customer-facing LLM applications that need prompt injection detection with low latency and low false positive rates across multiple languages.

For more on prompt injection and LLM threats, see our AI security guide. For open-source prompt injection detection, consider LLM Guard. For broader AI/ML model security (scanning, runtime defense), look at HiddenLayer or Protect AI Guardian. For LLM red teaming, see Garak or Promptfoo. For custom guardrail logic, explore NeMo Guardrails.

Note: Acquired by Check Point in 2025 to form the Global Center of Excellence for AI Security. Includes Lakera Guard, Lakera Red, and Gandalf Agent Breaker.

Frequently Asked Questions

What is Lakera Guard?
Lakera Guard is a real-time AI security API that protects LLM applications against prompt injection, jailbreaks, and data leakage. It delivers 98%+ detection rates with sub-50ms latency across 100+ languages. Lakera was acquired by Check Point in 2025.
Is Lakera Guard free or commercial?
Lakera Guard is commercial with a free tier available at platform.lakera.ai. Enterprise plans support higher volumes and additional features like custom policies and on-prem deployment.
Does Lakera Guard protect against prompt injection?
Yes, prompt injection detection is Lakera Guard’s core capability. It blocks direct injection, indirect injection, jailbreak attempts, and system prompt extraction in real time with sub-50ms latency.
What is the Gandalf game?
Gandalf is an educational game created by Lakera where players try to extract a secret password from an AI through prompt injection. It has attracted 1M+ players and generated 80M+ adversarial prompts that feed back into Lakera’s threat intelligence.
How does Lakera Guard integrate with LLM applications?
Lakera Guard uses a single API endpoint (POST to /v2/guard) that follows the OpenAI chat completions message format. It screens the last interaction in a message chain and returns a flagged boolean. Integration takes one API call.