
portkeyai/gateway with a tiny footprint (~45kb installed)
✅ Load balance across multiple models, providers, and keys
✅ Fallbacks make sure your app stays resilient
✅ Automatic Retries with exponential fallbacks come by default
✅ Plug-in middleware as needed
✅ Battle tested over 100B tokens
Run using Docker directly:
docker run -d -p 8787:8787 portkeyai/gateway:latest
For more information on the Docker image, check here
wget "[***]"
docker compose up -d
Let's try making a chat completions call to OpenAI through the AI gateway:
bashcurl '127.0.0.1:8787/v1/chat/completions' \ -H 'x-portkey-provider: openai' \ -H "Authorization: Bearer $OPENAI_KEY" \ -H 'Content-Type: application/json' \ -d '{"messages": [{"role": "user","content": "Say this is test."}], "max_tokens": 20, "model": "gpt-4"}'
Full list of supported SDKs
| Provider | Support | Stream | Supported Endpoints | |
|---|---|---|---|---|
| OpenAI | ✅ | ✅ | /completions, /chat/completions,/embeddings, /assistants, /threads, /runs, /images/generations, /audio/* | |
| Azure OpenAI | ✅ | ✅ | /completions, /chat/completions,/embeddings | |
| Anyscale | ✅ | ✅ | /chat/completions | |
| Google Gemini & Palm | ✅ | ✅ | /generateMessage, /generateText, /embedText | |
| Anthropic | ✅ | ✅ | /messages, /complete | |
| Cohere | ✅ | ✅ | /generate, /embed, /rerank | |
| Together AI | ✅ | ✅ | /chat/completions, /completions, /inference | |
| Perplexity | ✅ | ✅ | /chat/completions | |
| Mistral | ✅ | ✅ | /chat/completions, /embeddings | |
| Nomic | ✅ | ✅ | /embeddings | |
| AI21 | ✅ | ✅ | /complete, /chat, /embed | |
| Stability AI | ✅ | ✅ | /generation/{engine_id}/text-to-image | |
| DeepInfra | ✅ | ✅ | /inference | |
| Ollama | ✅ | ✅ | /chat/completions |
View the complete list of 100+ supported models here
Unified API SignatureConnect with 100+ LLM using OpenAI's API signature. The AI gateway handles the request, response and error transformations so you don't have to make any changes to your code. You can use the OpenAI SDK itself to connect to any of the supported LLMs.
|
Automatic RetriesTemporary issues shouldn't mean manual re-runs. AI Gateway can automatically retry failed requests upto 5 times. We apply an exponential backoff strategy, which spaces out retry attempts to prevent network overload.
|
| Language | Supported SDKs |
|---|---|
| Node.js / JS / TS | Portkey SDK OpenAI SDK LangchainJS LlamaIndex.TS |
| Python | Portkey SDK OpenAI SDK Langchain LlamaIndex |
| Go | go-openai |
| Java | openai-java |
| Rust | async-openai |
| Ruby | ruby-openai |
Join our growing community around the world, for help, ideas, and discussions on AI.
!Rubeus Social Share (4)




manifest unknown 错误
TLS 证书验证失败
DNS 解析超时
410 错误:版本过低
402 错误:流量耗尽
身份认证失败错误
429 限流错误
凭证保存错误
来自真实用户的反馈,见证轩辕镜像的优质服务