Resources
  • eBooks
  • Reports
  • Case Studies
  • Demos
  • Videos
  • Value Calculator
  • Home
  • Resources
  • Videos
  • Efficient Inference with Advanced Semantic Caching
Video

Kong AI Gateway: Advanced Semantic Caching, Routing, and Firewall for LLM

Revisit one of our sessions from API Summit 2024!

To reach a more desirable and precise response, it's highly beneficial to provide context to the LLM you're working with. Furthermore, rapid responses can be made possible by caching existing contextual queries. In this session, we'll explore the advanced Semantic Caching, Routing, and Firewall capabilities of the Kong AI Gateway.

Topics:AI Gateway
|
Performance
|
API Summit
Powering the API world

Increase developer productivity, security, and performance at scale with the unified platform for API management, AI gateways, service mesh, and ingress controller.

Sign up for Kong newsletter

Platform
Kong KonnectKong GatewayKong AI GatewayKong InsomniaDeveloper PortalGateway ManagerCloud GatewayGet a Demo
Explore More
Open Banking API SolutionsAPI Governance SolutionsIstio API Gateway IntegrationKubernetes API ManagementAPI Gateway: Build vs BuyKong vs PostmanKong vs MuleSoftKong vs Apigee
Documentation
Kong Konnect DocsKong Gateway DocsKong Mesh DocsKong AI GatewayKong Insomnia DocsKong Plugin Hub
Open Source
Kong GatewayKumaInsomniaKong Community
Company
About KongCustomersCareersPressEventsContactPricing
  • Terms•
  • Privacy•
  • Trust and Compliance
  • © Kong Inc. 2025