Resources
  • eBooks
  • Reports
  • Demos
  • Videos
|
  • Value Calculator
  1. Home
  2. Resources
  3. Videos
  4. Efficient Inference with Advanced Semantic Caching
Video

Kong AI Gateway: Advanced Semantic Caching, Routing, and Firewall for LLM

Revisit one of our sessions from API Summit 2024!

To reach a more desirable and precise response, it's highly beneficial to provide context to the LLM you're working with. Furthermore, rapid responses can be made possible by caching existing contextual queries. In this session, we'll explore the advanced Semantic Caching, Routing, and Firewall capabilities of the Kong AI Gateway.

Topics
AI GatewayPerformanceAPI Summit
Share on Social
Powering the API world

Increase developer productivity, security, and performance at scale with the unified platform for API management, AI gateways, service mesh, and ingress controller.

Sign up for Kong newsletter

    • Platform
    • Kong Konnect
    • Kong Gateway
    • Kong AI Gateway
    • Kong Insomnia
    • Developer Portal
    • Gateway Manager
    • Cloud Gateway
    • Get a Demo
    • Explore More
    • Open Banking API Solutions
    • API Governance Solutions
    • Istio API Gateway Integration
    • Kubernetes API Management
    • API Gateway: Build vs Buy
    • Kong vs Postman
    • Kong vs MuleSoft
    • Kong vs Apigee
    • Documentation
    • Kong Konnect Docs
    • Kong Gateway Docs
    • Kong Mesh Docs
    • Kong AI Gateway
    • Kong Insomnia Docs
    • Kong Plugin Hub
    • Open Source
    • Kong Gateway
    • Kuma
    • Insomnia
    • Kong Community
    • Company
    • About Kong
    • Customers
    • Careers
    • Press
    • Events
    • Contact
    • Pricing
  • Terms
  • Privacy
  • Trust and Compliance
  • © Kong Inc. 2026