Skip to main content

Documentation Index

Fetch the complete documentation index at: https://www.edgee.ai/docs/llms.txt

Use this file to discover all available pages before exploring further.

Edgee is an Agent Gateway, the infrastructure layer between your coding agent (Claude Code, Codex, OpenCode, Cursor) and the LLM provider APIs (Anthropic, OpenAI, GLM, others). It applies three things to every request: compression of context, intelligent routing across providers, and observability of token consumption. Edgee Agent Gateway

The three pillars

Compress

Surgical removal of redundancy from what enters and leaves the model. Two layers: Input (~99% of token volume) and Output (~1%, high ROI).

Route

Per-request fallback on provider 5xx and timeouts. Plan-cap continuity for Claude Pro/Max users when quota is hit. Configurable provider chain.

Observe

Every request, every compression event, every cost delta — at session level locally and at team level in the managed console.

Coding agents, start in seconds

1

Install the CLI

curl -fsSL https://install.edgee.ai | bash
2

Launch your coding agent

edgee launch claude
Your coding agent now runs through Edgee with compression, routing, and metering active. The CLI prints a session-analytics link on exit.

Use the API directly

import Edgee from 'edgee';

const edgee = new Edgee("your-api-key");

const response = await edgee.send({
  model: 'gpt-5.2',
  input: 'What is the capital of France?',
});

console.log(response.text);
if (response.compression) {
  console.log(`Tokens saved: ${response.compression.saved_tokens}`);
}
Edgee works with any OpenAI or Anthropic-compatible client by setting baseURL to https://api.edgee.ai.

Next steps

Why Edgee

The longer pitch — three pillars, OSS ecosystem, all the receipts.

Book a call

Talk to us about the managed Edgee console for teams.