Blog
  • AI Gateway
  • AI Security
  • AIOps
  • API Security
  • API Gateway
|
    • API Management
    • API Development
    • API Design
    • Automation
    • Service Mesh
    • Insomnia
    • View All Blogs
  1. Home
  2. Blog
  3. News
  4. The Age of AI Connectivity
News
December 18, 2025
8 min read

The Age of AI Connectivity

Augusto Marietti
CEO & Co-Founder of Kong

Kong was born to connect. The world is shifting from connecting cloud services with apps to connecting LLMs through agents. API calls and tokens are moving in tandem; a new unit of intelligence is forming. As AI traffic explodes into hypervolumes, speed is all that matters. The same principles of performance, security, and reliability behind Kong are essential in an agentic world. A new connectivity layer for AI is born.

Looking to discover more about AI connectivity? Join us live from the New York Stock Exchange on February 2. Kong's leadership team will show you how to unify your AI data path and move from fragmented experiments to production-ready infrastructure. Register now

From cloud native to agentic AI

A decade ago, we set out to connect the world through APIs, which we saw as fundamental building blocks of software. Before Kong, we founded Mashape as the first API marketplace to provide an assembly line for developers building apps, and then we open-sourced its core API runtime, Kong, the fastest, cloud-native API gateway.

Kong Inc. was officially born in the summer of 2017. A new era of microservices and cloud applications was unleashed. That moment took us from a small open-source gateway to the #1 API repo on GitHub and the leading cloud API management platform used by thousands of organizations worldwide — and recognized by Gartner, Forbes, and many others as a leader and one of the most visionary in cloud infrastructure. 

Today, the world stands at another inflection point: agentic AI. Connectivity is a constant. Builders still need to build cloud-native apps that connect services to humans. But a new consumer entity is emerging in agents, the first true “productization” of AI.

For the first time, the internet is not just about graphical UI but programming interfaces (voice soon). APIs are the UI for an agentic world built for machine-to-machine consumption — not people. Agents will devour those interfaces at a scale we can’t even comprehend now. They discover, learn, and transact faster than humans. New stacks are emerging. Backends will change. For example, in cloud-native apps, you mostly use SQL databases; in the agentic world, we'll use storage/compute-separated databases, vector databases, lakehouses, and LLMs as sources of data gravity. 

Mouse clicks, points, and scrolls will become less relevant. These will be used when things get stuck. An agentic workflow — with a voice overlay for human input — will become the primary method of internet consumption. Even if agent innovations plateau, more specialized agents will be deployed, thus increasing orchestration complexity and Kong's needs as traffic and connectivity challenges grow. 

The next big quest will be the transition from connecting cloud services to connecting intelligence. 

Welcome to the age of AI connectivity. 

API calls + tokens

The connective tissue of the cloud era has been API calls in pure REST fashion. But in the age of AI, we stand at another inflection point: tokens. They reside at the intersection of API traffic and AI traffic.

APIs call. Now they think. 

API traffic is adding an intelligence layer. Before a request leaves an endpoint or an MCP connection is established, it has context, memory, and reasoning. Soon enough, traffic itself will be intelligent. 

The internet's next layer won’t just be made of API calls powering applications; it’ll be made up of models, AI agents, voice assistants, and co-pilots that flow intelligence back and forth via tokens. Every time a system sends a prompt to a model, every context window, every agent-to-agent conversation produces a new type of token traffic. But why does this matter? 

Mountain pass.

I believe AI governance will be won right here. AI, to flow and transact, has to be securely governed. While initial market positioning puts this task near the data platform due to natural gravity, no company has a single data platform. Thus, the best place to manage the heterogeneous landscape of different clouds, data, and models is at the neutral traffic layer. It happened in the microservices revolution we led, and it will happen in AI.

It’s been our long-held belief that the most strategic wedge for building an API platform is to own the traffic-layer mountain pass. In the cloud-native world, API gateways were the broker technology. But in the agentic world, AI gateways are the next connectivity brokers orchestrating traffic between APIs and models – AI traffic will be 1000x larger. It's a machine scale. 

Thus, the mountain pass is shifting. 

Slow at first. Then all of a sudden. As new AI traffic emerges rapidly, new protocols appear (like REST in the cloud). MCP is the first mainstream form, but more AI protocols will emerge to enable AI to connect and talk in new ways. Our principles will always support what’s coming next for the most demanding companies. 

We were born for this moment. The same “connectivity logic” used to manage APIs  — or, better yet, the need for security, governance, acceleration, and observability — is even more critical for AI traffic.

Managing and delivering those tokens efficiently and securely is a new challenge that we solve.

APIs are for developers. MCP for LLMs. Just as Kong became the control plane for APIs, it will now become the control plane for AI traffic. The new unit of proxied value is tokens. Each token carries sensitive context and must be governed. Token throughput defines efficiency, and cost control is the difference between positive and negative margins.

Hypervolumes

As token factories multiply, a new AI scale law emerges. AI is bringing an order of magnitude and sizes that we have never seen before. Doing a billion API calls used to be a significant milestone for a business. Kong Gateway processes trillions of requests per day at ultra-fast speeds without dropping a single error. It's a huge number, yet it is not near what’s coming in the agentic AI era. 

The AI scale race is about who wins at hypervolumes. 

We now have to prepare our infrastructure for quadrillions and soon quintillions of API requests and AI traffic. The volume isn't just about calls and tokens, but also real-time data, events, MCP streamable sessions, and LLM prompts. It’s multi-modal. Its complexity is compounding. We not only have massive traffic volumes, but also heterogeneous ones. Most of the hardware and software infrastructure we’ve built until today will need to be redesigned from the ground up for hypervolumes. But not Kong. Kong was built for multi-modal hypervolumes.      

For example, in May 2025, Google CEO Sundar Pichai presented a chart showing Gemini's monthly token processing had exploded nearly 50x from May 2024 to 480 billion tokens. By October, the company announced reaching 1.3 QUADRILLION monthly tokens across all Google AI products.

Trillions were a unit of scale we thought was hard to reach. Now we have to believe in quadrillions and soon quintillions as AI sets up a new scaling law. Hypervolumes.

Speed is all you need

Tokens are still highly inefficient. Prompt compression, notation optimizations, semantic caching, and intelligent routing will be needed at scale to reduce waste and distribute intelligence abundantly. These are all AI connectivity challenges that Kong helps solve.  

In a world that’s already moving at a staggering pace and where data centers are becoming token-generation factories, revenue will be tied to power. And there are only two things that matter: 

  • Tokens x Watt 
  • Tokens x Second

The former is important because it's directly correlated to revenue and margins. The latter is essential because it's about revenue growth. The more tokens a business can push through, the more revenue it can produce. A half-second delay at the quadrillion-scale is a material revenue delta.

Tokens = Revenue.

We always wanted fast applications and low-latency backends with reliable API calls. Today, it’s not just about that faster digital experience — on someone waiting a few seconds longer to book a flight. This is about economic survival in the AI race. Businesses will have to transform themselves and acquire the fastest, most efficient AI stack from hardware to software. For example, in the CPU world, being 10% slower didn’t really matter, but in the GPU world, it's all about performance. Nvidia saw this a decade ago and captured most of the GPU market by building faster products than everyone else. 

Speed is all you need. 

As previously mentioned, the tokens x-second numbers we see today are still mainly human-driven. When agents go mainstream and consume tokens directly from other agents — to get work done — the throughput load will grow even more. 

But there's a catch! To go fast and move tokens with speed, enterprises will have to build new guardrails to be able to move intelligence with security, reliability, and a limited blast radius if things go wrong. Once again, our opportunity is to act as a control tower on the traffic layer.  

Tokens are how intelligence will move. Speed is how fast it will move the world.

Kong is speed.

A new connectivity layer

This unprecedented demand for speed requires new digital railroads. A connectivity layer for hypervolumes. For light-speed. With governance. For agents, API calls, models, and tokens. All wrapped around guardrails. 

A connectivity layer for AI.

We have the opportunity to build the digital railroads to enable every business to flourish, not to be left behind, to speed up, and to unleash intelligence securely. 

A new AI stack is emerging. 

We have new AI cloud “neoclouds,” like CoreWeave, foundation models, LLMOps, and inference platforms, just to name a few. The “clients” of this new wave are not just websites, services, and applications but AI agents, eventually robotics. In sum, machines.

Where's the connectivity layer — the glue that holds it all — and why now?

The big bet is that the same architectural pattern that emerged in cloud-native microservices will reappear in agentic AI. About a decade ago, higher traffic was coming into the infrastructure of enterprises and tech companies; containers, Kubernetes, serverless, YAML files, CI/CD, and more were born to decouple monoliths into microservices to build new scalable systems. 

There was a push to elastic-scale the backend into a self-sufficient, smaller piece of software. The cloud-native era was born.

This created an explosion of APIs to make it all talk, and HTTP RESTful traffic grew as more and more new apps and devices entered our lives. Initially, developers built business connectivity logic — rate-limiting, logging, authentication — directly into their microservices using libraries from their preferred programming languages. Then, as microservices grew into the hundreds, reinventing the wheel over and over again became R&D-intensive, and the API gateway was born.

The connectivity logic suddenly shifted to the proxy layer — abstracted, language-independent, and capable of dispatching the right controls (like rate limits) to each service. Engineering returned to focusing on building core IPs rather than traffic management. The API management market skyrocketed, and a complete API lifecycle was added around API gateways to help with building, running, discovering, and governing APIs. 

This is repeating with AI. At first, there were only a handful of LLMs within an organization, and frameworks like LangChain emerged. As companies mature their AI-native efforts, more large/medium/small language models and agents will enter their infrastructure, driving AI traffic across calls, event streams, and tokens. As before, a new connectivity logic will need to be built. Guardrails won’t be constructed anymore in the backends/LLMops over and over again, but they'll be abstracted to (once again) a new pattern: AI gateway. 

AI gateways aren't going to be the only kind of “AI middleware.” More types of traffic brokers will emerge from the edge of inference to internal agents-to-agents.  

The initial connectivity layer of AI developing is similar to API gateways, but more complex and semantic in nature. We’ll support ingress/egress traffic for LLMs and MCP servers, applying guardrails from API calls to tokens in a unified experience. For example, request rate limiting will intertwine with token budgets. 

The approach is the same; the connectivity logic will continue to live in the traffic layer, like an airport's control tower. Just as the control tower governs safety, routing, and pacing, AI Gateway governs tokens and agents' access and permissions. API and AI traffic commands will be dispatched to the right people, machines, services, and agents at the right moment. Intelligently flowing. 

The future path for every business is AI.

Kong built the only unified API and AI platform — securing, managing, accelerating, governing, and monetizing the flow of intelligence across every data, model, and API — on any cloud. 

A quadrillion tokens at a time. We’re deploying the modern railroads. The connectivity layer of AI to unleash intelligence securely. 

The Age of AI Connectivity has just begun.

Unleash the power of APIs with Kong Konnect

Learn MoreGet a Demo
AI ConnectivityAgentic AI

Table of Contents

  • From cloud native to agentic AI
  • API calls + tokens
  • Hypervolumes
  • Speed is all you need
  • A new connectivity layer

More on this topic

eBooks

The AI Connectivity Playbook: How to Build, Govern & Scale AI

Videos

PEXA’s Resilient API Platform on Kong Konnect

See Kong in action

Accelerate deployments, reduce vulnerabilities, and gain real-time visibility. 

Get a Demo
Topics
AI ConnectivityAgentic AI
Share on Social
Augusto Marietti
CEO & Co-Founder of Kong

Recommended posts

Announcing the Winners of Kong Agentic AI Hackathon

Kong Logo
NewsOctober 31, 2025

1st place: Session history plugin By: Abhishek Chauhan and Preesha Gehlot The Session History Plugin extends the Kong AI Gateway with persistent session objects that preserve conversation history across requests using the x-ai-session-id hea

Juhi Singh

API Summit 2025 Recap: AI Connectivity and the Agentic Era

Kong Logo
NewsOctober 16, 2025

That’s a wrap on API Summit 2025 ! At our eighth annual event, the brightest minds in the worlds of APIs and AI gathered in New York City to define the next chapter of digital innovation. We're entering an era where APIs are not just connecting ser

Augusto Marietti

New Year, New Unit Economics: Konnect Metering & Billing Is Here

Kong Logo
Product ReleasesJanuary 6, 2026

Every January, the same resolutions show up: eat better, exercise more, finally learn that language, finally figure out that production use case for AI agents (OK, this one isn’t so typical unless you operate in our universe).  But if you're respons

Alex Drag

Move More Agentic Workloads to Production with AI Gateway 3.13

Kong Logo
Product ReleasesDecember 18, 2025

MCP ACLs, Claude Code Support, and New Guardrails New providers, smarter routing, stronger guardrails — because AI infrastructure should be as robust as APIs We know that successful AI connectivity programs often start with an intense focus on how

Greg Peranich

The AI Governance Wake-Up Call

Kong Logo
EnterpriseDecember 12, 2025

Companies are charging headfirst into AI, with research around agentic AI in the enterprise finding as many as 9 out of 10 organizations are actively working to adopt AI agents.  LLMs are being deployed, agentic workflows are getting created left

Taylor Hendricks

Announcing the Kong Agentic AI Hackathon

Kong Logo
NewsAugust 12, 2025

Kong-quer the Agentic AI Hackathon 🚀 Calling all builders, tinkerers, and API innovators. The Kong Hackathon is back for  API Summit 2025 ! This year, we’re challenging developers worldwide to create projects that don’t just react, they  think ,  a

Juhi Singh

From Browser to Prompt: Building Infra for the Agentic Internet

Kong Logo
EnterpriseNovember 13, 2025

A close examination of what really powers the AI prompt unveils two technologies: the large language models (LLMs) that empower agents with intelligence and the ecosystem of MCP tools to deliver capabilities to the agents. While LLMs make your age

Amit Dey

Ready to see Kong in action?

Get a personalized walkthrough of Kong's platform tailored to your architecture, use cases, and scale requirements.

Get a Demo
Powering the API world

Increase developer productivity, security, and performance at scale with the unified platform for API management, AI gateways, service mesh, and ingress controller.

Sign up for Kong newsletter

    • Platform
    • Kong Konnect
    • Kong Gateway
    • Kong AI Gateway
    • Kong Insomnia
    • Developer Portal
    • Gateway Manager
    • Cloud Gateway
    • Get a Demo
    • Explore More
    • Open Banking API Solutions
    • API Governance Solutions
    • Istio API Gateway Integration
    • Kubernetes API Management
    • API Gateway: Build vs Buy
    • Kong vs Postman
    • Kong vs MuleSoft
    • Kong vs Apigee
    • Documentation
    • Kong Konnect Docs
    • Kong Gateway Docs
    • Kong Mesh Docs
    • Kong AI Gateway
    • Kong Insomnia Docs
    • Kong Plugin Hub
    • Open Source
    • Kong Gateway
    • Kuma
    • Insomnia
    • Kong Community
    • Company
    • About Kong
    • Customers
    • Careers
    • Press
    • Events
    • Contact
    • Pricing
  • Terms
  • Privacy
  • Trust and Compliance
  • © Kong Inc. 2026