Related Documentation
Made by
Kong Inc.
Supported Gateway Topologies
hybrid db-less traditional
Supported Konnect Deployments
hybrid cloud-gateways serverless
Compatible Protocols
grpc grpcs http https
Minimum Version
Kong Gateway - 3.8
Tags
#ai

3.10.0.1

Release date 2025/04/15

Bugfix

  • Fixed an issue where AI Proxy and AI Proxy Advanced would use corrupted plugin config.

3.10.0.0

Release date 2025/03/27

Breaking Change

  • Changed the serialized log key of AI metrics from ai.ai-proxy to ai.proxy to avoid conflicts with metrics generated from plugins other than AI Proxy and AI Proxy Advanced. If you are using logging plugins (for example, File Log, HTTP Log, etc.), you will have to update metrics pipeline configurations to reflect this change.

Deprecation

  • Deprecated preserve mode in config.route_type. Use config.llm_format instead. The preserve mode setting will be removed in a future release.

Feature

  • Added support for boto3 SDKs for Bedrock provider, and for Google GenAI SDKs for Gemini provider.

  • Added new priority balancer algorithm, which allows setting apriority group for each upstream model.

  • Added the failover_criteria configuration option, which allows retrying requests to the next upstream server in case of failure.

  • Added cost to tokens_count_strategy when using the lowest-usage load balancing strategy.

  • Added the huggingface, azure, vertex, and bedrock providers to embeddings. They can be used by the ai-proxy-advanced, ai-semantic-cache, ai-semantic-prompt-guard, and ai-rag-injector plugins.

  • Allow authentication to Bedrock services with assume roles in AWS.

  • Added the ability to set a catch-all target in semantic routing.

Bugfix

  • Fixed an issue where AI upstream URL trailing would be empty.

  • Fixed an issue where the ai-proxy-advanced plugin failed to failover between providers of different formats.

  • Fixed an issue where the ai-proxy-advanced plugin identity running failed in retry scenarios.

3.9.0.0

Release date 2024/12/12

Feature

  • Added support for streaming responses to the AI Proxy Advanced plugin.

  • Made the embeddings.model.name config field a free text entry, enabling use of a self-hosted (or otherwise compatible) model.

Bugfix

  • Fixed an issue where stale plugin config was not updated in dbless and hybrid mode.

  • Fixed an issue where lowest-usage and lowest-latency strategy did not update data points correctly.

3.8.1.0

Release date 2024/11/04

Bugfix

  • Fixed an issue where stale plugin config was not updated in dbless and hybrid mode.

  • Fixed an issue where lowest-usage and lowest-latency strategy did not update data points correctly.

3.8.0.0

Release date 2024/09/11

Feature

  • allow AI plugin to read request from buffered file

  • Added the ai-proxy-advanced plugin that supports advanced load balancing between LLM services.

Something wrong?

Help us make these docs great!

Kong Developer docs are open source. If you find these useful and want to make them better, contribute today!