Skip to main content

3 posts tagged with "Features"

Articles about Envoy AI Gateway Features

View All Tags

Announcing Model Context Protocol Support in Envoy AI Gateway

· 8 min read
Ignasi Barrera
Founding Engineer - Tetrate
Takeshi Yoneda
Envoy AI Gateway Maintainer - Tetrate

Hero feature image of title.

We’re excited to announce that the next release of Envoy AI Gateway will introduce first-class support for Model Context Protocol (MCP), cementing Envoy AI Gateway (EAIGW) as the universal gateway for modern production AI workloads.

Envoy AI Gateway started in close collaboration with Bloomberg and Tetrate to meet production-scale AI workload demands, combining real-world expertise and innovation from some of the industry’s largest adopters. Built upon the battle-tested Envoy Proxy data plane as the AI extension of Envoy Gateway, it is trusted for critical workloads by thousands of enterprises worldwide. EAIGW already provides unified LLM access, cost and quota enforcement, credential management, intelligent routing, resiliency, and robust observability for mission-critical AI traffic.

With the addition of MCP, we have brought these features to the communication between Agents and external tools, making EAIGW even more versatile for enterprise-scale AI deployments. For a deeper look at the collaborative story and technical vision, see the Bloomberg partnership announcement, their official release coverage, and previous project announcements.

Enhancing AI Gateway Observability - OpenTelemetry Tracing Arrives in Envoy AI Gateway

· 6 min read
Erica Hughberg
Envoy AI Gateway Maintainer - Tetrate
Adrian Cole
Principal Engineer - Tetrate

Hero feature image of title.

Aggregated metrics like latency, error rates, and throughput on their own won't reveal the source of why a system's output was wrong, slow, or expensive.

The v0.3 release of Envoy AI Gateway brings comprehensive OpenTelemetry tracing support with OpenInference semantic conventions, extending the existing metrics foundation to provide complete visibility into LLM application behavior.

This enables you to improve the quality and safety of your AI-integrated applications by allowing you to understand the full context of a request journey, as your LLM traces will inform application improvements and guardrail needs.

Envoy AI Gateway Introduces Endpoint Picker Support

· 7 min read
Erica Hughberg
Envoy AI Gateway Maintainer - Tetrate
Xunzhuo (Bit) Liu
Envoy AI Gateway Maintainer - Tencent

Reference Architecture for Envoy AI Gateway

Introduction

Envoy AI Gateway now supports Endpoint Picker Provider (EPP) integration as per the Gateway API Inference Extension.

This feature enables you to leverage intelligent, dynamic routing for AI inference workloads through intelligent endpoint selection based on real-time metrics, including KV-cache usage, queued requests, and LoRA adapter information.

When running AI inference at scale, this means your system can automatically select the optimal inference endpoint for each request, thereby optimizing resource utilization.

An overview of Endpoint Picker together with Envoy AI Gateway