Introducing unorouter: one API for every AI model
Why we built unorouter, how it routes across providers and what's next.
We built unorouter because managing a dozen AI provider keys, each with its own rate limits, pricing model, and outage window, is the opposite of productive. One endpoint, every major model, intelligent failover.
What it does
unorouter exposes a single OpenAI-compatible endpoint that fronts every model we support (Claude, GPT, Gemini, DeepSeek, Mistral, and more). Your application picks the model by name. We handle routing, authentication, rate-limit retries, and provider failover transparently.
Getting started in 30 seconds
Point any OpenAI SDK at our base URL and pass your unorouter API key. The rest of your code stays unchanged.
import OpenAI from "openai";
const client = new OpenAI({
baseURL: "https://api.unorouter.ai/v1",
apiKey: process.env.UNOROUTER_API_KEY,
});
const res = await client.chat.completions.create({
model: "claude-sonnet-4-6",
messages: [{ role: "user", content: "Hello!" }],
});Why it's different
- Automatic failover. When one provider is degraded or rate-limited, we transparently retry against the next healthy upstream.
- Per-token pricing, no subscription required. Top up once, use any model. Subscriptions are optional for volume discounts.
- All major endpoint shapes. OpenAI Chat Completions, Anthropic Messages, Google Gemini. Use the shape you already have.
What's next
We're shipping weekly. Expect: more models, smarter routing heuristics, better dashboards for cost and latency, and deeper integrations with the tooling developers already use (Claude Code, Codex CLI, Gemini CLI).
Ready to try it? Grab an API key or browse the model catalog.