Try Bifrost Enterprise free for 14 days.
Request access
[ BIFROST OSS FOR STARTUPS ]

Production-Ready LLM Gateway
Free & Open Source

Built for startups and growing teams to route requests across 1000+ providers, enforce budgets and rate limits, and cache responses. Scale without license fees.

Apache 2.0 License
Active Community
$0 Cost

[ PERFORMANCE AT A GLANCE ]

11µs
Gateway Overhead
Per request at 5K RPS
5K
Requests/Second
Sustained throughput
1000+
AI Models
Unified API access
0
Configuration
Deploy in seconds

[ INCLUDED IN OSS ]

Enterprise Features at Zero Cost

Bifrost OSS includes production-ready features that startups typically have to build themselves or pay enterprise prices to access. All free, all open source.

Automatic failover

Seamless provider switching when primary systems fail. Zero downtime when OpenAI, Anthropic, or other providers experience outages or hit rate limits.

99.999% uptime

Intelligent load balancing

Weighted API key distribution with model-specific filtering. Route requests across multiple keys and providers for maximum throughput and reliability.

Weighted routing

Semantic caching

Intelligent response caching based on semantic similarity. Reduce costs and latency by serving similar queries from cache instead of hitting the model.

Cost savings

Virtual keys & budgets

Create separate credentials for each team or project with independent rate limits and budgets. Hierarchical cost management across virtual keys, teams, users, and customers.

Access control

Real-time observability

Built-in request tracking, performance analysis, Prometheus metrics, and OpenTelemetry integration. Monitor every AI request in real-time.

OTEL native

Drop-in replacement

Replace existing AI SDK connections by changing just the base URL. Works with OpenAI SDK, Anthropic SDK, LangChain, and all major frameworks.

Zero refactor

[ WHY STARTUPS CHOOSE BIFROST OSS ]

Ship AI Features Faster, Cheaper, Better

Startups and SMEs use Bifrost OSS to avoid vendor lock-in, reduce costs, and focus on product instead of infrastructure.

Zero cost to start

Free, open-source, and production-ready. No enterprise license required for core functionality. Deploy on your infrastructure and scale without vendor lock-in.

Scale without breaking

Built in Go for high-performance workloads. Handles 5,000 requests per second with only 11µs overhead. Scales horizontally as your startup grows.

Avoid provider lock-in

Switch between 1000+ models without code changes. Compare model quality, cost, and latency across OpenAI, Anthropic, Google, and others from day one.

Cut costs immediately

Semantic caching reduces redundant model calls. Load balancing optimizes API key usage. Virtual key budgets prevent runaway spending.

Ship faster

No need to build custom routing, failover, or monitoring infrastructure. Focus on product features while Bifrost handles LLM reliability and observability.

Community-driven

Open source with active GitHub community. Contribute features, report issues, and learn from other startups building with Bifrost.

[ CORE CAPABILITIES ]

Everything You Need in the OSS Version

Multi-provider routing

Direct requests to specific models and providers with weighted strategies. Support for OpenAI, Anthropic, AWS Bedrock, Google Vertex, Azure, Groq, Mistral, Cohere, and 12+ additional providers.

MCP tool filtering

Allow-list controls determining which Model Context Protocol tools are available per virtual key. Fine-grained permissions for AI agent capabilities.

Custom Plugins

Custom plugins via Go or WebAssembly for bespoke business logic. Includes Mocker plugin for simulating provider responses during development.

Built-in web UI

Visual configuration dashboard with real-time monitoring. Configure virtual keys, view request logs, track costs, and analyze performance without code.

Prometheus & OTEL

Native metrics collection via Prometheus scraping or Push Gateway. OpenTelemetry integration with Grafana, New Relic, Honeycomb, and any observability stack.

Go SDK integration

Direct application integration for maximum performance and control. Deploy as HTTP gateway or embed in your Go application.

[ DEPLOYMENT ]

Deploy in Seconds, Your Way

Choose Docker, Kubernetes, standalone binary, or Go SDK. Self-host on any infrastructure.

Docker

Pull and run the official Docker image

docker pull bifrost-gateway
docker run -p 8080:8080 bifrost

Kubernetes

Deploy via Helm chart with auto-scaling

helm repo add bifrost https://charts.getbifrost.ai
helm install bifrost bifrost/bifrost

Binary

Download and run the standalone binary

wget https://github.com/maximhq/bifrost/releases/latest
./bifrost --port 8080

Go SDK

Embed directly in your application

import "github.com/maximhq/bifrost"

gateway := bifrost.NewGateway()
gateway.Start(":8080")

[ OSS VS ENTERPRISE ]

What's Included in Each Version

OSS provides production-ready core features. Enterprise adds compliance, advanced governance, and priority support.

Core Features

Open Source (Free)

  • Automatic failover
  • Load balancing
  • Semantic caching
  • Virtual keys & budgets
  • 1000+ models support
  • MCP tool filtering

Enterprise

  • Everything in OSS
  • Adaptive load balancing
  • Advanced routing strategies
  • Adaptive multi-region failover

Observability

Open Source (Free)

  • Real-time request logs
  • Prometheus metrics
  • OpenTelemetry integration
  • Built-in web UI
  • Cost tracking

Enterprise

  • Everything in OSS
  • Datadog/BigQuery connector
  • Log exports
  • Audit logging
  • Compliance reports

Security & Governance

Open Source (Free)

  • Virtual key access control
  • Rate limiting
  • Budget enforcement
  • MCP tool permissions

Enterprise

  • Everything in OSS
  • Guardrails (PII, content moderation)
  • SSO (like Okta, Entra)
  • RBAC
  • Vault integration
  • In-VPC deployment

Extensibility

Open Source (Free)

  • Go plugins
  • WASM plugins
  • Mocker plugin
  • Custom plugins

Enterprise

  • Everything in OSS
  • Priority support

[ USE CASES ]

Who Uses Bifrost OSS

Early-stage startups

Launch AI features without infrastructure overhead. Free OSS provides production-ready failover, caching, and multi-provider access from day one.

Rapid prototyping

Test models across OpenAI, Anthropic, Google, and others with zero code changes. Find the best quality/cost balance before committing to one provider.

Cost-conscious teams

Semantic caching and load balancing reduce redundant LLM spend. Virtual key budgets prevent surprise bills. Free OSS means no gateway license fees.

SME AI adoption

Small and medium enterprises can deploy enterprise-grade LLM infrastructure without enterprise budgets. Self-host on existing infrastructure.

Developer tools

Integrate AI features into developer tools, IDEs, or CLI applications. Go SDK provides direct embedding. HTTP gateway works with any language.

Open source projects

Add AI capabilities to open source software without vendor dependencies. Community-driven development with transparent roadmap and governance.

[ GROWTH PATH ]

Start Free, Scale as You Grow

Begin with OSS and upgrade to Enterprise only when you need compliance features.

Phase 1

Start Free

Deploy OSS with core features

  • Failover
  • Load balancing
  • Caching
  • Virtual keys
  • Observability
Phase 2

Scale Up

Add features as you grow

  • Horizontal scaling
  • Custom plugins
  • Advanced routing
  • Community support
Phase 3

Go Enterprise

Upgrade when you need compliance

  • Guardrails
  • SSO
  • RBAC
  • Audit logs
  • Priority support

Ready to Deploy Production-Ready LLM Infrastructure?

Join startups and SMEs building with Bifrost OSS. Free, open source, and production-ready.

Apache 2.0 License • Self-Host Anywhere • No Vendor Lock-In

[ BIFROST FEATURES ]

Open Source & Enterprise

Everything you need to run AI in production, from free open source to enterprise-grade features.

01 Governance

SAML support for SSO and Role-based access control and policy enforcement for team collaboration.

02 Adaptive Load Balancing

Automatically optimizes traffic distribution across provider keys and models based on real-time performance metrics.

03 Cluster Mode

High availability deployment with automatic failover and load balancing. Peer-to-peer clustering where every instance is equal.

04 Alerts

Real-time notifications for budget limits, failures, and performance issues on Email, Slack, PagerDuty, Teams, Webhook and more.

05 Log Exports

Export and analyze request logs, traces, and telemetry data from Bifrost with enterprise-grade data export capabilities for compliance, monitoring, and analytics.

06 Audit Logs

Comprehensive logging and audit trails for compliance and debugging.

07 Vault Support

Secure API key management with HashiCorp Vault, AWS Secrets Manager, Google Secret Manager, and Azure Key Vault integration.

08 VPC Deployment

Deploy Bifrost within your private cloud infrastructure with VPC isolation, custom networking, and enhanced security controls.

09 Guardrails

Automatically detect and block unsafe model outputs with real-time policy enforcement and content moderation across all agents.

[ SHIP RELIABLE AI ]

Try Bifrost Enterprise with a 14-day Free Trial

[quick setup]

Drop-in replacement for any AI SDK

Change just one line of code. Works with OpenAI, Anthropic, Vercel AI SDK, LangChain, and more.

1import os
2from anthropic import Anthropic
3
4anthropic = Anthropic(
5 api_key=os.environ.get("ANTHROPIC_API_KEY"),
6 base_url="https://<bifrost_url>/anthropic",
7)
8
9message = anthropic.messages.create(
10 model="claude-3-5-sonnet-20241022",
11 max_tokens=1024,
12 messages=[
13 {"role": "user", "content": "Hello, Claude"}
14 ]
15)
Drop in once, run everywhere.

[ FREQUENTLY ASKED QUESTIONS ]

Common Questions

Is Bifrost OSS truly free for production use?

Yes. Bifrost OSS is Apache 2.0 licensed and includes production-ready features like automatic failover, load balancing, semantic caching, and multi-provider routing at zero cost. No enterprise license required for core functionality.

What are the limitations of the OSS version?

OSS includes all core LLM gateway features. Enterprise adds advanced governance (guardrails, RBAC, SSO), compliance features (audit logging, in-VPC deployment), and priority support. Most startups and SMEs find OSS sufficient for production workloads.

Can I self-host Bifrost OSS on my infrastructure?

Yes. Bifrost OSS deploys via Docker, Kubernetes (Helm), standalone binary, or embedded Go SDK. Self-host on AWS, GCP, Azure, or any infrastructure. No data leaves your environment unless routing to external LLM providers.

How does Bifrost OSS compare to hosted LLM gateways?

Bifrost OSS is self-hosted (zero gateway fees), open source (full transparency), and community-driven. Hosted alternatives like Helicone or Portkey charge per request or require enterprise contracts. OSS gives you control and flexibility.

Does OSS include the web UI and monitoring dashboard?

Yes. The built-in web UI provides visual configuration, real-time request logs, cost tracking, and performance analysis. Prometheus metrics and OpenTelemetry integration are also included in OSS.

When should I upgrade from OSS to Enterprise?

Upgrade to Enterprise when you need compliance features (SOC 2 Type II, HIPAA, GDPR), advanced security (guardrails, SSO, RBAC), audit logging for regulated industries, or in-VPC deployment with priority support. OSS is production-ready for most use cases. Get 14 days free enterprise trial.