AI Gateway

Portkey provides the world's fastest multimodal AI gateway that enables working with multiple LLM providers as well as different modalities very easy. The gateway providers features to improve your app's reliability, cost efficiency and accuracy.

Features

Cover

Universal API

Use any of the supported models with a universal API (REST and SDKs)

Cover

Cache (Simple & Semantic)

Save costs and decrease latencies by using a cache

Cover

Fallbacks

Fallback between providers and models for resilience

Cover

Multimodality

Use vision, audio, image generation, and more models

Cover

Automatic Retries

Setup automatic retry strategies

Cover

Load Balancing

Load balance between various API Keys to counter rate-limits

Cover

Canary Testing

Canary test new models in production

Cover

Vault

Manage AI provider keys in a secure vault

Cover

Request Timeout

Easily handle unresponsive LLM requests

Conditional Routing

Route to different targets based on custom conditional checks

Using the Gateway

The various gateway strategies are implemented using Gateway configs. You can read more about configs below.

Configs

Open Source

We've open sourced our battle-tested AI gateway to the community. You can run it locally with a single command:

npx @portkey-ai/gateway

Contribute here.

While you're here, why not give us a star? It helps us a lot!

You can also self-host the gateway and then connect it to Portkey. Please reach out on [email protected] and we'll help you set this up!

Last updated