-
AI Gateway Overview
Overview of AI gateway capabilities
-
Quickstart
Get started quickly with AI Gateway setup and usage.
-
AI Gateway Capabilities
Learn about the core capabilities of AI Gateway.
-
AI providers
Learn about the various providers supported by AI Gateway.
-
AI Usage Governance
Understand how to manage and govern AI usage effectively.
-
Data Governance
Explore how AI Gateway helps enforce data governance policies.
-
Prompt Engineering
Best practices and tools for designing effective prompts.
-
Guardrails and Content Safety
Implement safeguards to ensure safe and compliant AI outputs.
-
Request Transformations
Customize and transform AI requests with Gateway features.
-
Streaming
Learn how AI Proxy streaming works.
-
Audit log
Learn about AI Gateway logging capabilities.
-
Monitor AI LLM Metrics
Explore how to monitor AI LLM metrics in AI Gateway.
-
Advanced Analytics
Access advanced analytics features in AI Gateway.
All AI Gateway documentation
Overview
AI Gateway plugins
-
AI Azure Content Safety
Use Azure AI Content Safety to check and audit AI Proxy plugin messages before proxying them to an upstream LLM
-
AI Prompt Decorator
Prepend or append an array of llm/v1/chat messages to a user's chat history
-
AI Prompt Guard
Check llm/v1/chat or llm/v1/completions requests against a list of allowed or denied expressions
-
AI Prompt Template
Provide fill-in-the-blank AI prompts to users
-
AI Proxy
The AI Proxy plugin lets you transform and proxy requests to a number of AI providers and models.
-
AI Proxy Advanced
The AI Proxy Advanced plugin lets you transform and proxy requests to multiple AI providers and models at the same time. This lets you set up load balancing between targets.
-
AI RAG Injector
Create RAG pipelines by automatically injecting content from a vector database
-
AI Rate Limiting Advanced
Provides rate limiting for the providers used by any AI plugins.
-
AI Request Transformer
Use an LLM service to transform a client request body prior to proxying the request to the upstream server
-
AI Response Transformer
Use an LLM service to transform the upstream HTTP(S) prior to forwarding it to the client
-
AI Semantic Prompt Guard
Semantically and intelligently create allow and deny lists of topics that can be requested across every LLM.
-
AI Sanitizer
Protect sensitive information in client request bodies before they reach upstream services
-
AI Prompt Compressor
Compress prompts before they are sent to LLMs to reduce costs, and improve latency
-
AI AWS Guardrails
Use AWS Guardrails to validate requests and/or responses in the AI Proxy plugin before forwarding them between clients and upstream LLMs.
MCP traffic gateway
-
MCP Traffic Gateway
This page is an introduction to MCP Traffic Gateway capabilites in Kong AI Gateway.
-
Secure MCP traffic
Secure GitHub MCP Server traffic with Kong Gateway and Kong AI Gateway
-
Govern MCP traffic
Use Kong AI Gateway to govern GitHub MCP traffic
-
Observe MCP traffic
Observe GitHub MCP traffic with Kong AI Gateway
-
Kong Konnect MCP Server
Get started with Kong Konnect MCP Server
-
Kong Konnect MCP Server tools
Check available Kong Konnect MCP Server tools
AI load balancing
-
Load balancing with AI Proxy Advanced
Overview of load balancing and retry and fallback strategies in the AI Proxy Advanced plugin.
-
Consistent Hashing - AI Proxy Advanced
Set up consistent hashing for load balancing.
-
Lowest Latency - AI Proxy Advanced
Configure load balancing based on the lowest latency.
-
Lowest Usage - AI Proxy Advanced
Set up load balancing based on the lowest usage.
-
Priority - AI Proxy Advanced
Configure priority-based load balancing.
-
Round Robin - AI Proxy Advanced
Set up round-robin load balancing.
-
Semantic - AI Proxy Advanced
Set up semantic load balancing.
How-tos
-
Collect Konnect audit logs
Learn how to configure you SIEM provider to collect Konnect logs and configure a Konnect audit log webhook.
-
Configure dynamic authentication to LLM providers using HashiCorp vault
Use HashiCorp Vault to securely store and reference API keys for OpenAI, Mistral, and other LLM providers in Kong AI Gateway.
-
Control prompt size with the AI Compressor plugin
Learn how to use the AI Compressor plugin alongside the RAG Injector and AI Prompt Decorator plugins to keep prompts lean, reduce latency, and optimize LLM usage for cost efficiency
-
Ensure chatbots adhere to compliance policies with the AI RAG Injector plugin
Learn how to configure the AI RAG Injector plugin.
-
Get started with AI Gateway
Learn how to quickly get started with AI Gateway
-
Observe GitHub MCP traffic with Kong AI Gateway
Learn how to observe MCP traffic within GitHub remote MCP server with the AI Proxy Advanced and Kong Gateway Prometheus plugin
-
Provide AI prompt templates for end users with the AI Prompt Template plugin and Mistral
Configure the AI Proxy plugin to route requests to a model provider like Mistral, then define reusable templates with the AI Prompt Template plugin to enforce consistent prompt formatting for tasks like summarization, code explanation, and Q&A.
-
Route OpenAI chat traffic using semantic balancing and Vault-stored keys
Use the AI Proxy Advanced plugin to route chat requests to OpenAI models based on semantic intent, secured with API keys stored in HashiCorp Vault.
-
Save LLM usage costs with AI Proxy Advanced semantic load balancing
Configure the AI Proxy Advanced plugin to optimize LLM usage and reduce costs by intelligently routing chat requests across multiple OpenAI models based on semantic similarity.
-
Send asynchronous requests to LLMs
Reduce costs by using llm/v1/files and llm/v1/batches route_types to send asynchronous batched requests to LLMs.
-
Set up AI Proxy Advanced with Anthropic in Kong Gateway
Configure the AI Proxy Advanced plugin to create a chat route using Anthropic.
-
Set up AI Proxy Advanced with Ollama
Configure the AI Proxy Advanced plugin to create a chat route using Ollama.
-
Set up AI Proxy Advanced with OpenAI in Kong Gateway
Configure the AI Proxy Advanced plugin to create a chat route using OpenAI.
-
Set up AI Proxy with Anthropic in Kong Gateway
Configure the AI Proxy plugin to create a chat route using Anthropic.
-
Set up AI Proxy with Ollama
Configure the AI Proxy Advanced plugin to create a chat route using Ollama.
-
Set up AI Proxy with OpenAI in Kong Gateway
Configure the AI Proxy plugin to create a chat route using OpenAI.
-
Store a Mistral API key as a secret in Konnect Config Store
Learn how to set up Konnect Config Store as a Vault backend and store a Mistral API key.
-
Store and rotate Mistral API keys as secrets in Google Cloud
Learn how to store and rotate secrets in Google Cloud with Kong Gateway, Mistral, and the AI Proxy plugin.
-
Transform a request body using OpenAI in Kong Gateway
Use the AI Request Transformer plugin with OpenAI to transform a client request body before proxying it.
-
Transform a response using OpenAI in Kong Gateway
Use the AI Response Transformer plugin with OpenAI to transform a response before returning it to the client.
-
Use AI Prompt Guard plugin to govern your LLM traffic
Use the AI Prompt Guard plugin to filter LLM traffic based on regex rules that allow general IT questions and deny unsafe or off-topic content.
-
Use AI Semantic Prompt Guard plugin to govern your LLM traffic
Use the AI Semantic Prompt Guard plugin to enforce topic-level guardrails for LLM traffic, filtering prompts based on meaning.
-
Use AI to protect sensitive information in requests
Use the AI Sanitizer plugin to protect sensitive information in requests.
-
Use Azure Content Safety plugin
Learn how to use the Azure AI Content Safety plugin.
-
Use Kong AI Gateway to govern GitHub MCP traffic
Learn how to govern MCP traffic within GitHub remote MCP server with the AI Proxy Advanced and AI Prompt Guard plugins
-
Use LangChain with AI Proxy in Kong Gateway
Connect your LangChain integrations with Kong Gateway with no code changes.
-
Use the AI AWS Guardrails plugin
Learn how to use the AI AWS Guardrails plugin.
-
Visualize AI Gateway metrics
Use a sample Elasticsearch, Logstash, and Kibana stack to visualize data from the AI Proxy plugin.