curl --request POST \
--url https://api.edgee.ai/v1/responses \
--header 'Authorization: Bearer <token>' \
--header 'Content-Type: application/json' \
--data '
{
"model": "openai/gpt-4o",
"input": "What is the capital of France?"
}
'{
"id": "resp_abc123",
"object": "response",
"status": "completed",
"created_at": 1677652288,
"model": "openai/gpt-4o",
"output": [
{
"id": "msg_1",
"type": "message",
"status": "completed",
"role": "assistant",
"content": [
{
"type": "output_text",
"text": "Paris"
}
]
}
],
"usage": {
"input_tokens": 8,
"output_tokens": 1,
"total_tokens": 9
}
}Create responses using the OpenAI Responses API format
curl --request POST \
--url https://api.edgee.ai/v1/responses \
--header 'Authorization: Bearer <token>' \
--header 'Content-Type: application/json' \
--data '
{
"model": "openai/gpt-4o",
"input": "What is the capital of France?"
}
'{
"id": "resp_abc123",
"object": "response",
"status": "completed",
"created_at": 1677652288,
"model": "openai/gpt-4o",
"output": [
{
"id": "msg_1",
"type": "message",
"status": "completed",
"role": "assistant",
"content": [
{
"type": "output_text",
"text": "Paris"
}
]
}
],
"usage": {
"input_tokens": 8,
"output_tokens": 1,
"total_tokens": 9
}
}Creates a response using the OpenAI Responses API format (Documentation Index
Fetch the complete documentation index at: https://www.edgee.ai/docs/llms.txt
Use this file to discover all available pages before exploring further.
POST /v1/responses). This endpoint is compatible with OpenAI’s Responses API, making it easy to use with tools like the Codex CLI.Comma-separated list of tags for categorizing and filtering requests in analytics and logs. Example: production,agent,codex
Enable debug mode to include additional debugging information in the response.
Compression bundle to apply.
claude, opencode, cursor, codex ID of the model to use, with provider prefix. Format: {author_id}/{model_id}.
"openai/gpt-4o"
The input to the model. Either a plain string (treated as a single user message) or a flat array of typed input items (messages, function calls, function call outputs).
System-level instruction prepended to the conversation. An alternative to including a system role message in the input array.
If set, the response is streamed as Server-Sent Events (SSE).
Maximum number of tokens to generate.
x >= 1Tools available to the model. Uses the Responses API flat format (no nested function key).
Show child attributes
Controls tool selection.
auto, none Sampling temperature between 0 and 2. Higher values produce more random outputs.
0 <= x <= 2Nucleus sampling probability. Alternative to temperature.
List of string tags for categorizing and filtering requests in analytics and logs. Can also be sent via the X-Edgee-Tags header.
Enable debug mode to include additional information in the response.
List of Edgee-managed tool IDs to include automatically (e.g. edgee_current_time, edgee_generate_uuid). Each ID must be activated for your API key.
["edgee_current_time", "edgee_generate_uuid"]Pending operation ID when continuing a conversation after Edge Tool execution. The gateway injects stored Edge Tool results into the conversation history.
Selects the compression bundle to apply to the request. Equivalent to the X-Edgee-Compression-Model header.
claude, opencode, codex, cursor Response created successfully
Unique identifier for the response, prefixed with resp_.
"resp_abc123"
response completed for non-streaming responses; in_progress is emitted on the initial response.created SSE event.
completed, in_progress Unix timestamp (as a float) of when the response was created.
1677652288
The model used to generate the response.
"openai/gpt-4o"
Array of output items produced by the model.
Show child attributes
Token usage statistics for the response.
Show child attributes
Was this page helpful?