Base URL
All URLs referenced in the documentation have the following base:Authentication
The Edgee API uses bearer authentication. When making requests, you must include your API Key in theAuthorization header in the format Bearer <token>. For more details, please refer to the Authentication page.
Errors
When an error occurs, the Edgee API responds with a conventional HTTP response code and a JSON object containing more details about the error. For more information, please refer to the Errors page.Rate Limiting
Please note that the Edgee has its own rate limit technology to prevent abuse and ensure service stability. If you exceed these limits, your requests will be throttled and you will receive a429 Too Many Requests response.
Additionally, usage limits may be enforced based on your API key configuration.
API Formats
Edgee supports two API formats to accommodate different use cases and provider preferences:OpenAI Format
Use
/v1/chat/completions for maximum flexibility. Works with all providers (OpenAI, Anthropic, Google, Meta, etc.).Model format: provider/model (e.g., anthropic/claude-sonnet-4.5)Recommended for: New integrations, multi-provider applications, and maximum flexibility.Anthropic Format
Use
/v1/messages for native Anthropic Messages API format. Only works with Anthropic provider.Model format: model (e.g., claude-sonnet-4.5)Recommended for: Migrating from Anthropic’s API, using Anthropic SDK, or requiring Anthropic-specific features.- For new projects: Use the OpenAI format (
/v1/chat/completions) for maximum flexibility and multi-provider support. - For Anthropic migrations: Use the Anthropic format (
/v1/messages) for seamless compatibility with existing code using Anthropic’s SDK.
Features
OpenAI-Compatible APIFully compatible with the OpenAI API format, making it easy to switch between providers or use multiple providers through a single interface.
Multi-Provider SupportAccess models from multiple providers (OpenAI, Anthropic, etc.) through a single API endpoint. Simply specify the model using the format
{author_id}/{model_id}.Streaming SupportBoth streaming and non-streaming responses are supported. Enable streaming by setting
stream: true to receive Server-Sent Events (SSE) with partial message deltas.Function CallingThe API supports function calling (tools) that allows models to call external functions, enabling more interactive and powerful applications.
Usage TrackingEvery response includes detailed usage statistics: token counts (prompt, completion, total), cached tokens, and reasoning tokens.