The Azure OpenAI SDK constructs request URLs in the format https://{azure_instance}.openai.azure.com/openai/deployments/{deployment_id}/chat/completions. Each deployment has its own URL path.
You can map each deployment to a separate Kong Gateway Route with its own AI Proxy Advanced configuration. The SDK switches between deployments by pointing azure_endpoint at Kong Gateway and changing the model parameter. Kong Gateway matches the request to the correct Route and forwards it to the corresponding Azure deployment. When the SDK sends a request with model="gpt-4o", the AzureOpenAI client constructs the path /openai/deployments/gpt-4o/chat/completions, which matches the first Route. Requests with model="gpt-4.1-mini" match the second Route.
This approach gives you explicit control over each deployment’s configuration, such as different auth keys, model options, or logging settings per deployment.