Build and Deploy Production-Ready Prompts
The Helicone Prompt Editor enables you to:
- Design prompts collaboratively in a UI
- Create templates with variables and track real production inputs
- Connect to any major AI provider (Anthropic, OpenAI, Google, Meta, DeepSeek and more)
Version Control for Your Prompts
Take full control of your prompt versions:
- Track versions automatically in code or manually in UI
- Switch, promote, or rollback versions instantly
- Deploy any version using just the prompt ID
Prompt Editor Copilot
Write prompts faster and more efficiently:
- Get auto-complete and smart suggestions
- Add variables (⌘E) and XML delimiters (⌘J) with quick shortcuts
- Perform any edits you describe with natural language (⌘K)
Real-Time Testing
Test and refine your prompts instantly:
- Edit and run prompts side-by-side with instant feedback
- Experiment with different models, messages, temperatures, and parameters
Auto-Improve (Beta)
We’re excited to launch Auto-Improve, an intelligent prompt optimization tool that helps you write more effective LLM prompts. While traditional prompt engineering requires extensive trial and error, Auto-Improve analyzes your prompts and suggests improvements instantly.
How it Works
- Click the Auto-Improve button in the Helicone Prompt Editor
- Our AI analyzes each sentence of your prompt to understand:
- The semantic interpretation
- Your instructional intent
- Potential areas for enhancement
- Get a new suggested optimized version of your prompt
Key Benefits
- Semantic Analysis: Goes beyond simple text improvements by understanding the purpose behind each instruction
- Maintains Intent: Preserves your original goals while enhancing how they’re communicated
- Time Saving: Skip hours of prompt iteration and testing
- Learning Tool: Understand what makes an effective prompt by comparing your original with the improved version
Using Prompts in Your Code
API Migration Notice: We are actively working on a new Router project that
will include an updated Generate API. While the previous Generate API
(legacy) is still functional (see the notice on
that page for deprecation timelines), here’s a temporary way to import and use
your UI-managed prompts directly in your code in the meantime:
For OpenAI users or Azure
const openai = new OpenAI({
baseURL: "https://generate.helicone.ai/v1",
defaultHeaders: {
"Helicone-Auth": `Bearer ${process.env.HELICONE_API_KEY}`,
OPENAI_API_KEY: process.env.OPENAI_API_KEY,
// For Azure users
AZURE_API_KEY: process.env.AZURE_API_KEY,
AZURE_REGION: process.env.AZURE_REGION,
AZURE_PROJECT: process.env.AZURE_PROJECT,
AZURE_LOCATION: process.env.AZURE_LOCATION,
},
});
const response = await openai.chat.completions.create({
inputs: {
number: "world",
},
promptId: "helicone-test",
} as any);
Using API to pull down the compiled prompt templates
Step 1: Get the compile the prompt template
Bash exmaple
curl --request POST \
--url https://api.helicone.ai/v1/prompt/helicone-test/compile \
--header 'Content-Type: application/json' \
--header 'authorization: sk-helicone-n4vqkhi-gg6exli-teictoi-aw7azyy' \
--data '{
"filter": "all",
"includeExperimentVersions": false,
"inputs": {
"number": "10"
}
}'
Javascript example with openai
const promptTemplate = await fetch(
"https://api.helicone.ai/v1/prompt/helicone-test/compile",
{
method: "POST",
headers: {
authorization: "sk-helicone-n4vqkhi-gg6exli-teictoi-aw7azyy",
"Content-Type": "application/json",
},
body: JSON.stringify({
filter: "all",
includeExperimentVersions: false,
inputs: { number: "10" }, // place all of your inputs here
}),
}
).then((res) => res.json() as any);
const example = (await openai.chat.completions.create({
...(promptTemplate.data.prompt_compiled as any),
stream: false, // or true
})) as any;
Responses are generated using AI and may contain mistakes.