What is Helicone?
Helicone is an open-source AI Gateway and LLM observability platform that helps developers build production AI applications with complete visibility and reliability. We solve the hardest problems in production LLM applications: provider outages that break your app, unpredictable costs, and debugging issues that are impossible to reproduce.AI Gateway
Access 100+ LLM models through a unified OpenAI-compatible API with automatic fallbacks, intelligent routing, and 0% markup
Full Observability
Complete visibility into every request with automatic cost tracking, latency monitoring, and session debugging
Open Source
Apache 2.0 licensed and self-hostable. Your data stays under your control, always accessible
Production Ready
SOC 2 and GDPR compliant with edge deployment for <50ms overhead. Built for your worst day
Key Features
🌐 AI Gateway
Access 100+ AI models with a single API key through our OpenAI-compatible gateway. Switch between OpenAI, Anthropic, Google, Groq, and more by just changing the model name. We handle intelligent routing, automatic fallbacks, and unified billing with 0% markup.🔍 LLM Observability
Inspect and debug every request with automatic tracking of costs, latency, token usage, and errors. Trace multi-step AI workflows and agent sessions with full visibility into the reasoning chain.🎮 Prompt Management
Version control your prompts and deploy changes without code updates. Test iterations in production, A/B test versions, and maintain full history of changes.💰 Cost & Usage Tracking
Understand your unit economics with real-time cost tracking across all providers. Break down spending by user, feature, or any custom dimension. Set budgets and prevent surprise bills.🎛️ Smart Controls
- Caching: Save costs and improve latency with automatic response caching
- Rate Limiting: Protect your budget with per-user or global rate limits
- Automatic Fallbacks: Stay online even when providers go down
- Security: Built-in content moderation and LLM security
Who Uses Helicone?
Startups
Move fast with instant observability. No need to build monitoring infrastructure from scratch.
Enterprises
Self-host for complete data control. SOC 2 compliant with enterprise Helm charts available.
AI Engineers
Debug complex agent workflows. Trace multi-step reasoning with session trees and full context.
Why Helicone?
Best Price AlwaysWe fight for every penny. 0% markup on credits means you pay exactly what providers charge. No hidden fees, no games. Invisible Performance
Your app shouldn’t slow down for observability. Edge deployment keeps us under 50ms overhead. Always. Always Online
Your app stays up, period. Providers fail, we fallback. Rate limits hit, we load balance. We don’t go down. Never Be Surprised
No shock bills. No mystery spikes. See every cost as it happens. We believe in radical transparency. Built for Your Worst Day
When production breaks at 3am and everyone’s panicking, we’re rock solid. Built for when you need us most.
Getting Started
Quick Start
Get your first LLM request logged in under 2 minutes
Quick Start Guide
Send your first request through the AI Gateway and see it logged in real-time
Understand the Platform
Learn how Helicone solves production AI challenges
Platform Overview
Explore the architecture and see how the gateway, observability, and features work together
Explore Features
Go deeper with advanced capabilities
Sessions & Agent Tracing
Debug multi-step AI workflows with full visibility
Prompt Management
Version and deploy prompts without code changes
Gateway Fallbacks
Automatic failover when providers go down
Cost Tracking
Understand your LLM economics by user or feature
Integration Methods
Helicone works in two ways:- AI Gateway (Recommended)
- Bring Your Own Keys
Use Credits for Instant Access
Add credits to your Helicone account and access 100+ models without signing up for each provider. We manage the API keys and handle billing at cost (0% markup).Benefits:- Single integration for all providers
- Switch models by just changing the name
- Automatic observability built-in
- No provider rate limits
- Unified billing
Supported Providers
Helicone integrates with 100+ LLM providers including:OpenAI
GPT-4o, GPT-4o-mini, o1, o3
Anthropic
Claude Sonnet 4, Opus, Haiku
Gemini 2.0, Gemini Pro, Vertex AI
Groq
Llama, Mixtral, Gemma
AWS Bedrock
Claude, Llama, Titan
Azure OpenAI
All Azure-hosted models
Together AI
Open-source models
And More
Mistral, DeepSeek, Perplexity, X.AI, Fireworks, and more
Framework Integrations
Helicone works seamlessly with popular AI frameworks:- LangChain (JS/TS & Python) - Use AI Gateway for unified provider access
- LlamaIndex (Python) - Full observability for data applications
- LangGraph (Python) - Track stateful multi-actor workflows
- Vercel AI SDK (JS/TS) - Build AI-powered applications
- CrewAI (Python) - Monitor role-playing AI agents
Open Source & Self-Hosting
Helicone is Apache 2.0 licensed and fully open source. Self-host for complete data ownership and control.Docker Setup
Get started with Docker Compose in minutes
Helm Chart
Production-ready Kubernetes deployment (contact [email protected])
⭐ Star us if you find Helicone useful!
Community & Support
Discord Community
Join 2000+ developers building with Helicone
Documentation
Comprehensive guides and API references
Email Support
[email protected] for questions and feedback
GitHub Issues
Report bugs and request features
What’s Next?
Ready to get started? Follow our quick start guide to log your first request in under 2 minutes.Quick Start Guide
Send your first LLM request through Helicone and see it logged in real-time
