[ SUPPORTED CLI AGENTS ]
Each CLI agent brings unique capabilities. Bifrost provides the enterprise controls, multi-provider access, and observability that all of them need at scale.
Official CLI Agent for Anthropic Models
Provider: Anthropic
OpenAI's powerful code generation CLI
Provider: OpenAI
Google Gemini powered coding agent
Provider: Google
Open source coding agent framework
Provider: Multi-provider
[ CORE CAPABILITIES ]
Whether you use Claude Code, Codex CLI, Gemini CLI, or OpenCode, routing through Bifrost gives you enterprise-grade controls without changing developer workflows.
Every CLI agent gets access to 1000+ models through Bifrost. Switch between Anthropic, OpenAI, Google, AWS Bedrock, and more without reconfiguring your agent.
1000+ modelsAll agent requests route through Bifrost, applying guardrails, PII redaction, content policies, and role-based access controls transparently.
SOC 2 Type II readyTrack LLM spend across all CLI agents from a single dashboard. Virtual keys enforce per-team budgets. Export via the logs API or OTEL for custom cost analysis.
Per-request costsEvery request from Claude Code, Codex CLI, Gemini CLI, and OpenCode is logged with full metadata. Export to any observability stack via OpenTelemetry.
OTEL nativeWhen a provider hits rate limits or experiences an outage, Bifrost automatically routes agent requests to healthy providers. Adaptive load balancing keeps throughput stable.
99.999% uptimeRepeat or near-identical queries across agents resolve instantly from cache, cutting costs and reducing latency for common coding operations.
Cost savings[ SETUP ]
Deploy Bifrost once, configure your agents, and get instant enterprise controls across all CLI coding tools.
Run Bifrost as a standalone service in your VPC or use managed hosting. Configure your LLM providers and optional guardrails through the dashboard.
# pull and start bifrost docker pull bifrost-gateway docker run -p 8080:8080 bifrost
Set the API base URL environment variable for your chosen agent. Each agent has slightly different configuration requirements.
# Claude Code export ANTHROPIC_BASE_URL="http://localhost:8080" # Codex CLI (must end in /v1) export OPENAI_API_BASE="http://localhost:8080/openai/v1" # Gemini CLI export GEMINI_API_BASE="http://localhost:8080" # OpenCode export API_BASE_URL="http://localhost:8080"
Launch your agent and start coding. All requests automatically route through Bifrost with full observability, cost tracking, and guardrails applied.
# Use Bifrost CLI for guided setup npx -y @maximhq/bifrost-cli # Or launch agents directly claude code codex generate gemini chat opencode
[ AGENT CONFIGURATION ]
Each agent requires slightly different configuration. Here are the specifics for connecting to Bifrost.
Environment Variable:
ANTHROPIC_BASE_URLValue:
http://localhost:8080Supports MCP tool injection via Bifrost
Example command:
claude mcp add --transport http bifrost http://localhost:8080/mcpEnvironment Variable:
OPENAI_API_BASEValue:
http://localhost:8080/openai/v1Must end in /v1 to avoid 404 errors
Example command:
codex config set api-base http://localhost:8080/openai/v1Environment Variable:
GEMINI_API_BASEValue:
http://localhost:8080Standard Google AI endpoint format
Example command:
gemini config set base-url http://localhost:8080Environment Variable:
API_BASE_URLValue:
http://localhost:8080Flexible multi-provider configuration
Example command:
opencode config base http://localhost:8080[ BIFROST CLI ]
Bifrost CLI is an interactive terminal tool that launches coding agents through a Bifrost gateway without manual configuration. Run a single command, select your preferred agent and model through a guided setup, and start coding with enterprise controls.
npx -y @maximhq/bifrost-cli[ COMPARISON ]
| Feature | Standalone Agents | Agents + Bifrost |
|---|---|---|
| Multi-provider access | No | 20+ providers |
| Automatic failover | No | Across all providers |
| Cost tracking | No | Real-time per request |
| Team budgets | No | Virtual keys + limits |
| Security guardrails | No | PII redaction + policies |
| Request observability | No | Full logs + OTEL export |
| Semantic caching | No | Reduce costs + latency |
| MCP tool integration | Claude Code only | All agents via gateway |
[ USE CASES ]
Teams use different CLI agents for different tasks: Claude Code for architecture, Codex for quick scripts, Gemini for code review, OpenCode for deployments. Bifrost provides unified governance.
Run the same coding task through Claude Code, Codex CLI, and Gemini CLI to compare quality, speed, and cost. Bifrost logs performance metrics for each provider.
Organizations route all CLI agent traffic through Bifrost to enforce PII redaction, content policies, and audit trails for SOC 2 Type II, HIPAA, and GDPR compliance.
Platform teams set budgets for each engineering department. Real-time dashboards show which teams, projects, and agents are driving LLM spend.
When Anthropic hits rate limits, Claude Code requests automatically route to Bedrock. When OpenAI hits rate limits or errors, Codex CLI switches to Azure OpenAI.
Security teams enforce organization-wide policies across all CLI agents from a single control plane. No need to configure guardrails per agent.
[ ENTERPRISE FEATURES ]
Bifrost provides the enterprise controls platform teams need before rolling out CLI agents organization-wide.
Create separate credentials for each team with independent rate limits and budgets.
Detect and redact PII, sensitive data, and API keys before requests reach models.
Complete tamper-evident logs of every request for compliance and security.
Authenticate via Google, GitHub, Okta, or any SAML/OIDC provider.
Traffic distributes intelligently based on real-time health and performance.
Deploy entirely within your network perimeter for maximum security.
[ BIFROST FEATURES ]
Everything you need to run AI in production, from free open source to enterprise-grade features.
01 Governance
SAML support for SSO and Role-based access control and policy enforcement for team collaboration.
02 Adaptive Load Balancing
Automatically optimizes traffic distribution across provider keys and models based on real-time performance metrics.
03 Cluster Mode
High availability deployment with automatic failover and load balancing. Peer-to-peer clustering where every instance is equal.
04 Alerts
Real-time notifications for budget limits, failures, and performance issues on Email, Slack, PagerDuty, Teams, Webhook and more.
05 Log Exports
Export and analyze request logs, traces, and telemetry data from Bifrost with enterprise-grade data export capabilities for compliance, monitoring, and analytics.
06 Audit Logs
Comprehensive logging and audit trails for compliance and debugging.
07 Vault Support
Secure API key management with HashiCorp Vault, AWS Secrets Manager, Google Secret Manager, and Azure Key Vault integration.
08 VPC Deployment
Deploy Bifrost within your private cloud infrastructure with VPC isolation, custom networking, and enhanced security controls.
09 Guardrails
Automatically detect and block unsafe model outputs with real-time policy enforcement and content moderation across all agents.
[ SHIP RELIABLE AI ]
Change just one line of code. Works with OpenAI, Anthropic, Vercel AI SDK, LangChain, and more.
[ FREQUENTLY ASKED QUESTIONS ]
Yes. Bifrost provides 100% compatible endpoints for OpenAI, Anthropic, and Gemini APIs, allowing you to route Claude Code, Codex CLI, Gemini CLI, and OpenCode through the same gateway simultaneously. Each agent gets access to all configured providers. Read more about Bifrost CLI.
You can configure model routing through Bifrost virtual keys, which specify which models each key can access. Alternatively, use Bifrost CLI (npx @maximhq/bifrost-cli) for guided setup that lets you select models before launching agents. The gateway handles provider routing automatically.
Bifrost adds approximately 11µs of gateway overhead per request at 5,000 RPS. This is effectively invisible compared to LLM response times that typically range from hundreds of milliseconds to several seconds. Read more about Bifrost becnhmarks.
Claude Code has native MCP support that integrates directly with Bifrost MCP endpoints. Other agents can access MCP-connected tools through Bifrost's gateway, though the integration may vary by agent. The Bifrost CLI automatically registers MCP servers for Claude Code.
Bifrost logs every request with model, provider, token counts, and cost. Use virtual keys to tag requests by agent, team, or environment. Export via the logs API or OTEL to build cost breakdowns in your BI tool of choice.
Bifrost automatically fails over to healthy providers. If Claude Code is configured to use Anthropic but hits rate limits, requests transparently route to AWS Bedrock or Google Vertex AI. Developers experience zero downtime.