Complete reference for Bifrost HTTP transport API endpoints and usage patterns.
http://localhost:8080
(configurable)
All endpoints and request/response formats are OpenAI compatible.
"stream": true
in your request. The response will be a text/event-stream
of Server-Sent Events (SSE).
Request with Streaming:
data:
. The stream is terminated by a [DONE]
message.
Parameter | Type | Description | Example |
---|---|---|---|
model | string | Provider and model name | "openai/gpt-4o-mini" |
params | object | Model parameters | {"temperature": 0.7} |
fallbacks | array | Fallback model names | ["anthropic/claude-3-sonnet-20240229"] |
Parameter | Type | Default | Description |
---|---|---|---|
temperature | float | 1.0 | Randomness (0.0-2.0) |
max_tokens | integer | Provider default | Maximum tokens to generate |
top_p | float | 1.0 | Nucleus sampling (0.0-1.0) |
frequency_penalty | float | 0.0 | Frequency penalty (-2.0-2.0) |
presence_penalty | float | 0.0 | Presence penalty (-2.0-2.0) |
stop | array | null | Stop sequences |
Status | Code | Description |
---|---|---|
400 | invalid_request_error | Bad request format |
401 | authentication_error | Invalid API key |
403 | permission_error | Access denied |
429 | rate_limit_error | Rate limit exceeded |
500 | internal_error | Server error |
503 | service_unavailable | Provider unavailable |