Together, the following components represent the three layers of the new AI platform:
- - AI Gateway: Kong AI Gateway (including MCP support) controls both GenAI and MCP flow and orchestrates the existing services like Vector Databases, Event Streaming, etc.
- - Compute Layer: Akamai Linode Kubernetes Engine (LKE)
- - Security Layer: Akamai Firewall for AI
Here's the fundamental diagram with the three main components:

A basic LLM request flow goes through the following steps:
- - User or Agent sends a request with a prompt.
- - Request passes through Akamai Firewall for AI. Prompt is inspected and sanitized.
- - Request reaches the infrastructure on Akamai LKE.
- - Request goes through Kong AI Gateway: The request is Authenticated and routed to the best model.
- - Model executes inference.
- - Response flows back to Kong which logs it and generates metrics and traces.
- - Filtered again by Akamai Firewall for AI.
- - Return to user.
Multiple variations also exist. Depending on the request, Kong AI Gateway may send queries to the Vector Database to solve, for example, Semantic Routing policies.
Similarly, the Agent might send MCP requests which will be also handled by the Data Plane which is responsible for enforcing policies specifically defined for the MCP Servers and Tools, like, for example OAuth2 Authorization.