Quickstart
WorldRouter is a unified OpenAI-compatible endpoint for top AI models. This page walks you through the four things you need to start using it: an API key, the base URL, a connection test, and the model IDs you can call.
Get an API key
- Open the API Keys dashboard.
- Sign in or create a WorldRouter account.
- Click Create API key, name it, and copy the generated key (format:
sk-...). - Add credits on the Credits page so the key can make paid calls.
Keys are shown only once. Store them in a password manager or server-side secret store. Never commit them to source control or paste them into client-side code.
Base URL
Every WorldRouter request goes to the same OpenAI-compatible endpoint. Use this value for the base_url / baseURL field in any SDK or the target URL for direct HTTP calls:
https://inference-api.worldrouter.ai/v1The URL above is resolved for the environment you are signed in to, so copy-paste works without further editing.
Verify the connection
Run one of the snippets below with your key in place of your_api_key. A successful call returns a chat completion JSON payload, which confirms the key, the base URL, and your network are all working:
curl https://inference-api.worldrouter.ai/v1/chat/completions \
-H "Authorization: Bearer your_api_key" \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-5.4",
"messages": [{ "role": "user", "content": "Hello" }]
}'A successful response looks like this — the choices[0].message.content is the model’s reply, and the usage block shows the tokens billed against your credits:
{
"id": "chatcmpl-9f2abc123",
"object": "chat.completion",
"created": 1716240000,
"model": "gpt-5.4",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 9,
"completion_tokens": 11,
"total_tokens": 20
}
}A 401 means the key is wrong or missing. A 404 on /chat/completions almost
always means the base URL is missing the /v1 suffix; copy it from the block
above.
Supported models
These are the model IDs you can use to test the connection — drop any one of them into the model field of the verification request above to check whether that specific model is currently callable on your account. The full catalog with live pricing lives on the Models page.
| Vendor | Model ID | Input (Credits / 1M) | Output (Credits / 1M) | Cache read (Credits / 1M) | Cache write (Credits / 1M) |
|---|---|---|---|---|---|
| Anthropic | claude-opus-4-7 | 350 | 1,750 | 35 | 437.5 |
claude-opus-4-6 | 350 | 1,750 | 35 | 437.5 | |
claude-sonnet-4-6 | 210 | 1,050 | 21 | 262.5 | |
claude-haiku-4-5 | 70 | 350 | 7 | 87.5 | |
| OpenAI | gpt-5.5 | 700 | 3,150 | 70 | 700 |
gpt-5.4 | 350 | 1,575 | 35 | 350 | |
gpt-5.4-mini | 52.5 | 315 | 5.25 | 52.5 | |
gpt-oss-120b | 3.9 | 19 | 3.9 | 3.9 | |
gpt-oss-20b | 3 | 14 | 3 | 3 | |
gemini-3.1-pro-preview | 280 | 1,260 | 28 | 280 | |
gemini-3.1-flash-lite-preview | 17.5 | 105 | 1.75 | 17.5 | |
gemma-4-31b-it | 13 | 38 | 13 | 13 | |
gemma-4-26b-a4b-it | 13 | 40 | 13 | 13 | |
gemma-3-27b-it | 8 | 16 | 8 | 8 | |
gemma-3-12b-it | 4 | 13 | 4 | 4 | |
| Alibaba | qwen3.5-plus | 40.11 | 240.8 | 4.011 | 50.1375 |
qwen3.5-flash | 12.04 | 120.4 | 1.204 | 15.05 | |
qwen3.6-plus | 77.07 | 462.14 | 7.707 | 96.3375 | |
qwen3-coder-plus | 200.76 | 2,006.97 | 20.076 | 250.95 | |
qwen3.5-flash-02-23 | 10 | 40 | 1 | 12.5 | |
qwen3-235b-a22b-2507 | 10 | 60 | 10 | 10 | |
qwen3.5-397b-a17b | 39 | 234 | 19.5 | 39 | |
qwen3.5-9b | 10 | 15 | 10 | 10 | |
qwen3.5-35b-a3b | 25 | 200 | 25 | 25 | |
qwen3.5-27b | 30 | 240 | 30 | 30 | |
qwen3.5-plus-02-15 | 50 | 300 | 5 | 62.5 | |
qwen3-vl-235b-a22b-instruct | 20 | 88 | 11 | 20 | |
qwen3-coder-next | 14 | 80 | 9 | 14 | |
qwen3-next-80b-a3b-instruct | 9 | 110 | 9 | 9 | |
qwen3-coder | 22 | 180 | 22 | 22 | |
qwen3-30b-a3b-instruct-2507 | 9 | 30 | 9 | 9 | |
qwen3-32b | 8 | 24 | 4 | 8 | |
| Moonshot | kimi-k2.6 | 66.5 | 280 | 11.2 | 66.5 |
kimi-k2.5 | 42 | 210 | 7 | 42 | |
| MiniMax | MiniMax-M2.7 | 21 | 84 | 4.2 | 26.25 |
MiniMax-M2.5 | 21 | 84 | 4.2 | 26.25 | |
| Zhipu | glm-5.1 | 112 | 352 | 20.8 | 112 |
glm-5 | 70 | 224 | 14 | 70 | |
glm-5-turbo | 120 | 400 | 24 | 120 | |
glm-4.5-air | 13 | 85 | 2.5 | 13 | |
glm-4.7 | 40 | 175 | 8 | 40 | |
glm-4.6 | 39 | 190 | 39 | 39 | |
glm-4.7-flash | 6 | 40 | 1 | 6 | |
| DeepSeek | deepseek-v4-pro | 121.8 | 243.6 | 10.15 | 121.8 |
deepseek-v4-flash | 9.8 | 19.6 | 1.96 | 9.8 | |
deepseek-v3.2 | 19.6 | 29.4 | 1.96 | 19.6 | |
deepseek-chat-v3-0324 | 20 | 77 | 13.5 | 20 | |
deepseek-chat-v3.1 | 15 | 75 | 15 | 15 | |
deepseek-v3.1-terminus | 21 | 79 | 13 | 21 | |
deepseek-v3.2-exp | 27 | 41 | 27 | 27 | |
| xAI | grok-4.1-fast | 20 | 50 | 5 | 20 |
grok-4-fast | 20 | 50 | 5 | 20 | |
grok-4.20 | 400 | 1,200 | 40 | 400 | |
| StepFun | step-3.5-flash | 10 | 30 | 10 | 10 |
| Xiaomi | mimo-v2-pro | 200 | 600 | 40 | 200 |
mimo-v2-flash | 9 | 29 | 4.5 | 9 | |
mimo-v2.5-pro | 200 | 600 | 40 | 200 | |
mimo-v2-omni | 40 | 200 | 8 | 40 | |
mimo-v2.5 | 80 | 400 | 16 | 80 | |
| Mistral | mistral-nemo | 2 | 4 | 2 | 2 |
mistral-small-3.2-24b-instruct | 7.5 | 20 | 7.5 | 7.5 | |
| Meta | llama-3.1-8b-instruct | 2 | 5 | 2 | 2 |
llama-3.1-70b-instruct | 40 | 40 | 40 | 40 | |
llama-4-maverick | 15 | 60 | 15 | 15 | |
llama-3.3-70b-instruct | 10 | 32 | 10 | 10 | |
| NVIDIA | nemotron-3-nano-30b-a3b | 5 | 20 | 5 | 5 |
nemotron-3-super-120b-a12b | 9 | 45 | 9 | 9 |
Supported tools
WorldRouter works with any client that accepts a custom OpenAI-compatible base URL. Pick the integration path that matches how you use AI, and follow its setup guide:
CLI
Set up terminal agents such as Codex, Claude Code, and OpenCode with environment variables and config files.
VS Code extension
Use the Codex plugin in VS Code with WorldRouter as the OpenAI-compatible endpoint.
API
Call the OpenAI-compatible API directly from Python, Node.js, Go, PHP, or any HTTP client.
- CLI agents: terminal agents including Codex, Claude Code, and OpenCode.
- VS Code extensions: the first-party Codex plugin and other extensions that read a custom OpenAI endpoint.
- Direct API: any application, SDK, or script that speaks the OpenAI
/chat/completionsprotocol. See the API reference for full details.
What’s next
- Base URL for this environment:
https://inference-api.worldrouter.ai/v1. - Manage keys and view usage on the API Keys dashboard.
- Browse the full Models catalog with live pricing.
- Read the API reference for authentication, error codes, and privacy.