You can proxy requests to Cohere AI models through AI Gateway using the AI Proxy and AI Proxy Advanced plugins. This reference documents all supported AI capabilities, configuration requirements, and provider-specific details needed for proper integration.
Cohere provider
Upstream paths
AI Gateway automatically routes requests to the appropriate Cohere API endpoints. The following table shows the upstream paths used for each capability.
| Capability | Upstream path or API |
|---|---|
| Chat completions | /v1/chat |
| Completions | /v1/generate |
| Embeddings | /v2/embed |
| Function calling | /v1/chat |
Supported capabilities
The following tables show the AI capabilities supported by Cohere provider when used with the AI Proxy or the AI Proxy Advanced plugin.
Set the plugin’s
route_typebased on the capability you want to use. See the tables below for supported route types.
Text generation
Support for Cohere basic text generation capabilities including chat, completions, and embeddings:
| Capability | Route type | Streaming | Model example | Min version |
|---|---|---|---|---|
| Chat completions | llm/v1/chat |
command-a-03-2025 | 3.6 | |
| Completions | llm/v1/completions |
command-r-plus-08-2024 | 3.6 | |
| Embeddings | llm/v1/embeddings |
embed-english-v3.0 | 3.11 |
Advanced text generation
Support for Cohere function calling to allow Cohere models to use external tools and APIs:
| Capability | Route type | Model example | Min version |
|---|---|---|---|
| Function calling | llm/v1/chat |
command-a-03-2025 | 3.6 |
Cohere base URL
The base URL is https://api.cohere.com:443/{route_type_path}, where {route_type_path} is determined by the capability.
AI Gateway uses this URL automatically. You only need to configure a URL if you’re using a self-hosted or Cohere-compatible endpoint, in which case set the upstream_url plugin option.
Supported native LLM formats for Cohere
By default, the AI Proxy plugin uses OpenAI-compatible request formats. Set config.llm_format to a native format to use Cohere-specific APIs and features.
The following native Cohere APIs are supported:
| LLM format | Supported APIs |
|---|---|
cohere |
|
Configure Cohere with AI Proxy
To use Cohere with AI Gateway, configure the AI Proxy or AI Proxy Advanced.
Here’s a minimal configuration for chat completions:
For more configuration options and examples, see:
Tutorials
- Guide survey classification behavior using the AI Prompt Decorator plugin View →
- Set up AI Proxy Advanced with Cohere in Kong Gateway View →
- Set up AI Proxy with Cohere in Kong Gateway View →
- Enforce responsible AI behavior using the AI Prompt Decorator plugin View →
- Use Cohere rerank API for document-grounded chat with AI Proxy in Kong Gateway View →
FAQs
How do I use Cohere’s document-grounded chat for RAG pipelines?
Configure AI Proxy with the Cohere provider and send queries with candidate documents. The model filters for relevance and returns answers with citations. See Use Cohere rerank API for document-grounded chat.