Skip to main content

Release Notes

Stay up-to-date with the latest improvements, features, and fixes in Envoy AI Gateway.

v0.4.x

November 7, 2025 - Current
Latest
Release introducing Model Context Protocol (MCP) Gateway, OpenAI Image Generation, Anthropic support (direct and AWS Bedrock), guided output decoding for GCP Vertex AI/Gemini, cross-namespace references, enhanced authentication, and comprehensive observability improvements.
MCP GatewayImage GenerationAnthropic SupportGuided Output DecodingCross-Namespace SupportEnhanced Upstream AuthInferencePool v1
Releases: v0.4.0

v0.3.x

August 21, 2025 - November 7, 2025
Stable
Release version introducing intelligent inference routing with Endpoint Picker Provider,enhanced observability features, Google Vertex AI support, and enhanced provider integrations.
Endpoint Picker SupportGoogle Vertex AIExpanded Provider EcosystemArchitecture ImprovementsInferencePool SupportGateway API Inference ExtensionOpenInference TracingModel Name VirtualizationProduction Ready ProvidersDynamic Load Balancing
Releases: v0.3.0

v0.2.x

June 5 - August 21, 2025
Stable
Release version introducing Azure OpenAI integration, sidecar architecture, cross-backend failover, and a CLI tool.
Azure OpenAI IntegrationSidecar ArchitecturePerformance ImprovementsCLI ToolsModel Failover and RetryCertificate Manager Integration
Releases: v0.2.0, v0.2.1

v0.1.x

Feb 25 - Apr 3, 2025
First Release
The first official release of Envoy AI Gateway! This milestone includes unified API for LLM providers, token-based rate limiting, traffic management, and Kubernetes native integration.
OpenAIAWS BedrockRate LimitingUnified APIUpstream AuthenticationToken-Based Rate Limiting
Releases: v0.1.0, v0.1.1, v0.1.2, v0.1.3, v0.1.4, v0.1.5

Release Timeline

v0.4.0November 07, 2025

Envoy AI Gateway v0.4.0 is a major release introducing Model Context Protocol (MCP) Gateway with full OAuth and server multiplexing support, OpenAI Image Generation API, first-party Anthropic provider integration, and native Anthropic Messages API support for AWS Bedrock. This release also brings guided output decoding for GCP Vertex AI/Gemini, cross-namespace resource references, enhanced authentication for AWS and Azure, and comprehensive observability improvements across all providers.

v0.3.0August 21, 2025

Envoy AI Gateway v0.3.0 introduces intelligent inference routing, expanded provider support (including Google Vertex AI and Anthropic), and enhanced observability with OpenInference tracing and configurable metrics. Key features include Endpoint Picker Provider with InferencePool for dynamic load balancing, model name virtualization, and seamless Gateway API Inference Extension integration.

v0.2.1June 9, 2025

Quick bug fix release addressing an AWS authentication issue discovered in the v0.2.0 release.

v0.2.0June 5, 2025

Envoy AI Gateway v0.2.0 builds upon the solid foundation of v0.1.0 with focus on expanding provider ecosystem support, improving reliability and performance through architectural changes, and enterprise-grade authentication support for Azure OpenAI.

v0.1.5April 3, 2025

Final patch in the v0.1.x series addressing ExtProc image synchronization.

v0.1.4March 20, 2025

Additional AWS validation and error handling improvements.

v0.1.3March 14, 2025

Streaming improvements and enhanced observability with GenAI metrics.

v0.1.2March 5, 2025

Container image and deployment fixes for improved reliability.

v0.1.1February 28, 2025

Quick bug fix release addressing AWS tooling issues discovered in the initial release.

v0.1.0February 25, 2025

🎉 The inaugural release of Envoy AI Gateway! This milestone release establishes the foundation for unified AI service management with support for major LLM providers, comprehensive rate limiting, and Kubernetes-native integration.


Community and Support

Envoy AI Gateway is a community-driven project. We welcome your contributions and feedback!