diff --git a/website/docs/overview/architecture/gateway-integrations.md b/website/docs/overview/architecture/gateway-integrations.md
new file mode 100644
index 000000000..921f085f5
--- /dev/null
+++ b/website/docs/overview/architecture/gateway-integrations.md
@@ -0,0 +1,85 @@
+---
+id: gateway-integrations
+title: Gateway Integrations
+sidebar_label: Gateway Integrations
+description: How the Semantic Router plugs into Envoy AI Gateway, Istio, AIBrix, LLM-D, and the vLLM Production Stack, plus what each integration adds.
+---
+
+The Semantic Router ships with multiple gateway profiles. This page shows **Which gateway plugs in**, **What SR adds**, and **What’s already validated**.
+
+## High-level topology
+
+import ZoomableMermaid from '@site/src/components/ZoomableMermaid';
+
+
+{`
+flowchart LR
+ C[Client / SDK]
+ GW["Gateway
(Envoy | Istio | AIBrix | LLM-D | Prod Stack)"]
+ SR["Semantic Router
(ExtProc gRPC)"]
+ SC["Semantic Cache
(Milvus)"]
+ OBS["Telemetry
(OTel → Prom/Grafana)"]
+ B1["Cloud LLMs
(OpenAI, Anthropic, ...)"]
+ B2["Self-hosted
vLLM workers"]
+
+ C --> GW
+ GW -- ExtProc
Inference Extension --> SR
+ SR -->|headers: model, safety| GW
+ SR --> SC
+ SR --> OBS
+ GW --> B1
+ GW --> B2
+ B1 --> OBS
+ B2 --> OBS
+
+ style SR fill:#1f2937,stroke:#0ea5e9,stroke-width:2,color:#e5e7eb
+ style GW fill:#0f172a,stroke:#a855f7,stroke-width:2,color:#e5e7eb
+`}
+
+
+## Supported Profiles
+
+| Gateway profile | Integration path | SR adds | CI status | Manifests / config |
+| -------------------- | ------------------------------------------ | ----------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------------------------------------ |
+| **Envoy AI Gateway** | ExtProc gRPC (Envoy AI Gateway → SR) | Classification → model header, PII/jailbreak, semantic cache, observability headers | [](https://github.com/vllm-project/semantic-router/actions/workflows/integration-test-k8s.yml)
**Validates:**
• **Features:** Classification, Cache, PII, Jailbreak
• **Routing:** Priority, Fallback, Keyword
• **Traffic:** Chat API, Stress tests
[](https://github.com/vllm-project/semantic-router/actions/workflows/integration-test-helm.yml)
**Validates:** Install, Upgrade, Rollback | [`deploy/kubernetes/ai-gateway`](https://github.com/vllm-project/semantic-router/tree/main/deploy/kubernetes/ai-gateway) |
+| **Istio Gateway** | Gateway API Inference Extension + ExtProc | Same as above; demo with dual vLLM backends | Manual guide | [`deploy/kubernetes/istio`](https://github.com/vllm-project/semantic-router/tree/main/deploy/kubernetes/istio) |
+| **AIBrix Gateway** | Envoy Gateway API resources + ExtProc | SR intelligence in front of AIBrix autoscaler and distributed KV | Helm + AIBrix manifests;
follows Envoy ExtProc;
Planned E2E | [`deploy/kubernetes/aibrix`](https://github.com/vllm-project/semantic-router/tree/main/deploy/kubernetes/aibrix) |
+| **LLM-D Gateway** | Istio Gateway + LLM-D schedulers + ExtProc | Semantic routing feeds pool selection in LLM-D | Covered by Istio flow;
Planned E2E | [`deploy/kubernetes/llmd-base`](https://github.com/vllm-project/semantic-router/tree/main/deploy/kubernetes/llmd-base) |
+
+> **Reading map**: pick your gateway, open the install guide, then jump to the manifests to see the exact resources the diagram refers to.
+
+## Request Flow
+
+
+{`
+sequenceDiagram
+ autonumber
+ participant Client
+ participant Gateway
+ participant SR as Semantic Router
+ participant Cache as Semantic Cache
+ participant Upstream as LLM Backends
+
+ Client->>Gateway: OpenAI-compatible request
+ Gateway->>SR: ExtProc gRPC (headers/body)
+ SR->>SR: PII / jailbreak / category classification
+ SR->>Cache: Semantic lookup
+ alt cache hit
+ SR-->>Gateway: Headers + cached response
+ else miss
+ SR-->>Gateway: Route headers (model, policy flags)
+ Gateway->>Upstream: Forward to chosen backend
+ Upstream-->>Gateway: LLM response
+ Gateway-->>SR: Response headers/body (optional)
+ SR->>Cache: Write entry
+ end
+ Gateway-->>Client: Final response
+`}
+
+
+## Where to go next
+
+- **Envoy AI Gateway install**: [installation/k8s/ai-gateway](../../installation/k8s/ai-gateway)
+- **Istio Gateway install**: [installation/k8s/istio](../../installation/k8s/istio)
+- **AIBrix Gateway install**: [installation/k8s/aibrix](../../installation/k8s/aibrix)
+- **LLM-D Gateway install**: [installation/k8s/llm-d](../../installation/k8s/llm-d)
diff --git a/website/sidebars.ts b/website/sidebars.ts
index 4509c36cd..790b265fd 100644
--- a/website/sidebars.ts
+++ b/website/sidebars.ts
@@ -27,6 +27,7 @@ const sidebars: SidebarsConfig = {
label: 'Architecture',
items: [
'overview/architecture/system-architecture',
+ 'overview/architecture/gateway-integrations',
'overview/architecture/envoy-extproc',
'overview/architecture/router-implementation',
],