[ PERFORMANCE AT A GLANCE ]
[ INCLUDED IN OSS ]
Bifrost OSS includes production-ready features that startups typically have to build themselves or pay enterprise prices to access. All free, all open source.
Seamless provider switching when primary systems fail. Zero downtime when OpenAI, Anthropic, or other providers experience outages or hit rate limits.
99.999% uptimeWeighted API key distribution with model-specific filtering. Route requests across multiple keys and providers for maximum throughput and reliability.
Weighted routingIntelligent response caching based on semantic similarity. Reduce costs and latency by serving similar queries from cache instead of hitting the model.
Cost savingsCreate separate credentials for each team or project with independent rate limits and budgets. Hierarchical cost management across virtual keys, teams, users, and customers.
Access controlBuilt-in request tracking, performance analysis, Prometheus metrics, and OpenTelemetry integration. Monitor every AI request in real-time.
OTEL nativeReplace existing AI SDK connections by changing just the base URL. Works with OpenAI SDK, Anthropic SDK, LangChain, and all major frameworks.
Zero refactor[ WHY STARTUPS CHOOSE BIFROST OSS ]
Startups and SMEs use Bifrost OSS to avoid vendor lock-in, reduce costs, and focus on product instead of infrastructure.
Free, open-source, and production-ready. No enterprise license required for core functionality. Deploy on your infrastructure and scale without vendor lock-in.
Built in Go for high-performance workloads. Handles 5,000 requests per second with only 11µs overhead. Scales horizontally as your startup grows.
Switch between 1000+ models without code changes. Compare model quality, cost, and latency across OpenAI, Anthropic, Google, and others from day one.
Semantic caching reduces redundant model calls. Load balancing optimizes API key usage. Virtual key budgets prevent runaway spending.
No need to build custom routing, failover, or monitoring infrastructure. Focus on product features while Bifrost handles LLM reliability and observability.
Open source with active GitHub community. Contribute features, report issues, and learn from other startups building with Bifrost.
[ CORE CAPABILITIES ]
Direct requests to specific models and providers with weighted strategies. Support for OpenAI, Anthropic, AWS Bedrock, Google Vertex, Azure, Groq, Mistral, Cohere, and 12+ additional providers.
Allow-list controls determining which Model Context Protocol tools are available per virtual key. Fine-grained permissions for AI agent capabilities.
Custom plugins via Go or WebAssembly for bespoke business logic. Includes Mocker plugin for simulating provider responses during development.
Visual configuration dashboard with real-time monitoring. Configure virtual keys, view request logs, track costs, and analyze performance without code.
Native metrics collection via Prometheus scraping or Push Gateway. OpenTelemetry integration with Grafana, New Relic, Honeycomb, and any observability stack.
Direct application integration for maximum performance and control. Deploy as HTTP gateway or embed in your Go application.
[ DEPLOYMENT ]
Choose Docker, Kubernetes, standalone binary, or Go SDK. Self-host on any infrastructure.
Pull and run the official Docker image
docker pull bifrost-gateway docker run -p 8080:8080 bifrost
Deploy via Helm chart with auto-scaling
helm repo add bifrost https://charts.getbifrost.ai helm install bifrost bifrost/bifrost
Download and run the standalone binary
wget https://github.com/maximhq/bifrost/releases/latest ./bifrost --port 8080
Embed directly in your application
import "github.com/maximhq/bifrost"
gateway := bifrost.NewGateway()
gateway.Start(":8080")[ OSS VS ENTERPRISE ]
OSS provides production-ready core features. Enterprise adds compliance, advanced governance, and priority support.
Open Source (Free)
Enterprise
Open Source (Free)
Enterprise
Open Source (Free)
Enterprise
Open Source (Free)
Enterprise
[ USE CASES ]
Launch AI features without infrastructure overhead. Free OSS provides production-ready failover, caching, and multi-provider access from day one.
Test models across OpenAI, Anthropic, Google, and others with zero code changes. Find the best quality/cost balance before committing to one provider.
Semantic caching and load balancing reduce redundant LLM spend. Virtual key budgets prevent surprise bills. Free OSS means no gateway license fees.
Small and medium enterprises can deploy enterprise-grade LLM infrastructure without enterprise budgets. Self-host on existing infrastructure.
Integrate AI features into developer tools, IDEs, or CLI applications. Go SDK provides direct embedding. HTTP gateway works with any language.
Add AI capabilities to open source software without vendor dependencies. Community-driven development with transparent roadmap and governance.
[ GROWTH PATH ]
Begin with OSS and upgrade to Enterprise only when you need compliance features.
Deploy OSS with core features
Add features as you grow
Upgrade when you need compliance
[ BIFROST FEATURES ]
Everything you need to run AI in production, from free open source to enterprise-grade features.
01 Governance
SAML support for SSO and Role-based access control and policy enforcement for team collaboration.
02 Adaptive Load Balancing
Automatically optimizes traffic distribution across provider keys and models based on real-time performance metrics.
03 Cluster Mode
High availability deployment with automatic failover and load balancing. Peer-to-peer clustering where every instance is equal.
04 Alerts
Real-time notifications for budget limits, failures, and performance issues on Email, Slack, PagerDuty, Teams, Webhook and more.
05 Log Exports
Export and analyze request logs, traces, and telemetry data from Bifrost with enterprise-grade data export capabilities for compliance, monitoring, and analytics.
06 Audit Logs
Comprehensive logging and audit trails for compliance and debugging.
07 Vault Support
Secure API key management with HashiCorp Vault, AWS Secrets Manager, Google Secret Manager, and Azure Key Vault integration.
08 VPC Deployment
Deploy Bifrost within your private cloud infrastructure with VPC isolation, custom networking, and enhanced security controls.
09 Guardrails
Automatically detect and block unsafe model outputs with real-time policy enforcement and content moderation across all agents.
[ SHIP RELIABLE AI ]
Change just one line of code. Works with OpenAI, Anthropic, Vercel AI SDK, LangChain, and more.
[ FREQUENTLY ASKED QUESTIONS ]
Yes. Bifrost OSS is Apache 2.0 licensed and includes production-ready features like automatic failover, load balancing, semantic caching, and multi-provider routing at zero cost. No enterprise license required for core functionality.
OSS includes all core LLM gateway features. Enterprise adds advanced governance (guardrails, RBAC, SSO), compliance features (audit logging, in-VPC deployment), and priority support. Most startups and SMEs find OSS sufficient for production workloads.
Yes. Bifrost OSS deploys via Docker, Kubernetes (Helm), standalone binary, or embedded Go SDK. Self-host on AWS, GCP, Azure, or any infrastructure. No data leaves your environment unless routing to external LLM providers.
Bifrost OSS is self-hosted (zero gateway fees), open source (full transparency), and community-driven. Hosted alternatives like Helicone or Portkey charge per request or require enterprise contracts. OSS gives you control and flexibility.
Yes. The built-in web UI provides visual configuration, real-time request logs, cost tracking, and performance analysis. Prometheus metrics and OpenTelemetry integration are also included in OSS.
Upgrade to Enterprise when you need compliance features (SOC 2 Type II, HIPAA, GDPR), advanced security (guardrails, SSO, RBAC), audit logging for regulated industries, or in-VPC deployment with priority support. OSS is production-ready for most use cases. Get 14 days free enterprise trial.