Blog
  • AI Gateway
  • AI Security
  • AIOps
  • API Security
  • API Gateway
    • API Management
    • API Development
    • API Design
    • Automation
    • Service Mesh
    • Insomnia
    • View All Blogs
  1. Home
  2. Blog
  3. Enterprise
  4. The Next Frontier: Container Orchestration
Enterprise
October 27, 2021
4 min read

The Next Frontier: Container Orchestration

Marco Palladino
CTO and Co-Founder of Kong
Topics
KubernetesDigital TransformationThought Leadership
Share on Social

More on this topic

eBooks

Maturity Model for API Management

eBooks

Federated API Management: Accelerating Innovation with Autonomy and Oversight

See Kong in action

Accelerate deployments, reduce vulnerabilities, and gain real-time visibility. 

Get a Demo

In part 1 of this series on Kubernetes, we discussed how companies like VMware offer the necessary tools to launch, monitor, create and destroy virtual machines. In this post, we review how - much like virtual machines - containers need to be created, monitored, destroyed and relaunched to account for the health of the physical or virtual machines on which they run.

As container adoption has exploded, the need for container orchestration, which automates these processes, has become manifest. Container orchestration allows developers to better track, schedule and operationalize containers at scale.

Modern applications are made up of many single-purpose services called microservices, each running in its container. These can be developed and scaled independently based on demand for each particular microservice.

Usually, the containers run in the cloud, are distributed in multiple data centers for resilience in the face of outages and sometimes on a mix of cloud and on-premises servers to optimize for cost and other factors. If DevOps want to run multiple application instances (containers) of numerous applications in more than one location, they need an impossible amount of DevOps manpower and automation. For microservices-based applications, automation tools are an absolute necessity.

Microservices and Containerization

The complexity introduced by microservices and containerization requires DevOps to answer several questions, such as when to start the right containers, how to ensure the containers can talk to each other, what storage requirements each container has and how to ensure high availability across the entire infrastructure.

Container orchestrators let developers define the rules and requirements of each container while abstracting away the ongoing operations required to match containers with appropriate resources and keep them running and communicating with each other. Much like AWS simplifies EC2 instances management, container orchestrators launch containers and automate or abstract away infrastructure issues, such as what physical server will fulfill the containers' requirements.

Container orchestration tools allow DevOps to start new containers without worrying about what underlying virtual machine will handle the workload. For example, a container orchestration tool such as Kubernetes will do things like deciding if a virtual machine is underutilized and then decide to run a container on that virtual machine over another.

Kubernetes: Modern Infrastructure for Modern Applications

Recent survey data from The New Stack suggests that container adoption is the most significant catalyst for orchestration adoption. Sixty percent of respondents who have deployed containers in production report that they also rely on Kubernetes in production. Another 19 percent of respondents with broad container deployments in production were in the initial stages of Kubernetes adoption.

So, what exactly is Kubernetes? Started by Google in 2014, Kubernetes is an open-source project that focuses on building a robust orchestration system for running thousands of containers in production. Kubernetes can automate many of the painful manual tasks and infrastructure complexity required to deploy, scale and manage containerized applications, which would otherwise fall on DevOps teams.

Not every container is the same. A container can be a stateless service like a web app or contain a stateful service like a node of a distributed database. Kubernetes allows DevOps to run individual, stateless containers efficiently and to account for the different requirements of more complex, containerized services, including storage, networking, or interdependencies between containers.

The key to the rise of Kubernetes is its ability to account for the needs of the containerized application as a complete system, not just as individual containers.

Kubernetes addresses common design challenges, including:

  • How to deploy and orchestrate applications that consist of many moving parts
  • How to design applications that can easily move from one cloud to another
  • How to evenly distribute load across different instances of a service
  • How to ensure that stateful applications retain access to required data and storage when they restart

Even with the rise of Kubernetes as the de facto container orchestration system, there are still some misconceptions about Kubernetes. Many misunderstand the relationship between Docker and Kubernetes.

Docker is a company that popularized containerization when it created a standard way of encoding containers (a Docker Image) and the tooling needed to run that container on a host (called Docker Engine).

Kubernetes is an open-source project under the Cloud Native Computing Foundation that orchestrates and manages containers at scale. It automates the container lifecycle, making use of Docker's container format and tooling to encode and run each container while adding functionality to abstract away the infrastructure concerns discussed above.

Another common misconception is that Kubernetes is a platform as a service (PaaS). Although many vendors host Kubernetes as part of a platform, Kubernetes can be deployed anywhere: a laptop, the public cloud or on-prem. This flexibility is part of why Kubernetes has seen so much success.

In the last part of this series, we go into more detail on How Kubernetes Gets Work Done.

Topics
KubernetesDigital TransformationThought Leadership
Share on Social
Marco Palladino
CTO and Co-Founder of Kong

Recommended posts

Kong Mesh 2.12: SPIFFE/SPIRE Support and Consistent XDS Resource Names

Kong Logo
Product ReleasesSeptember 18, 2025

We're very excited to announce Kong Mesh 2.12 to the world! Kong Mesh 2.12 delivers two very important features: SPIFFE / SPIRE support, which provides enterprise-class workload identity and trust models for your mesh, as well as a consistent Kuma R

Justin Davies

You Might Be Doing API-First Wrong, New Analyst Research Suggests

Kong Logo
EnterpriseSeptember 3, 2025

Ever feel like you're fighting an uphill battle with your API strategy? You're building APIs faster than ever, but somehow everything feels harder. Wasn’t  API-first  supposed to make all this easier?  Well, you're not alone. And now industry analys

Heather Halenbeck

72% Say Enterprise GenAI Spending Going Up in 2025, Study Finds

Kong Logo
EnterpriseJune 18, 2025

Survey Says: Google LLMs See Usage Surge, Most OK with DeepSeek in the Workplace Enterprise adoption of large language models (LLMs) is surging. According to Gartner , more than 80% of enterprises will have deployed generative AI (GenAI) applicatio

Eric Pulsifer

5 Steps to Immediately Reduce Kafka Cost and Complexity

Kong Logo
EnterpriseJune 24, 2025

Kafka delivers massive value for real-time businesses — but that value comes at a cost. As usage grows, so does complexity: more clusters, more topics, more partitions, more ACLs, more custom tooling. But it doesn’t have to be that way. If your tea

Umair Waheed

Kong Mesh 2.11: Reduced Privileges, Improved Support for AWS ECS

Kong Logo
Product ReleasesJune 20, 2025

We’re at it again, bringing more incremental improvements to Kong Mesh!  Built on top of Kuma, Kong Mesh brings much-needed simplicity and production-grade tooling. Kong Mesh is built for smooth operations with platform teams in mind, providing secu

Justin Davies

Is Ambient Mesh the Future of Service Mesh?

Kong Logo
EnterpriseJune 30, 2025

A Practical Look at When (and When Not) to Use Ambient Mesh The word on the street is that ambient mesh is the obvious evolution of service mesh technology — leaner, simpler, and less resource-intensive. But while ambient mesh is an exciting develop

Umair Waheed

How to Create a Platform Cross-Charging Model (and Why Not To Do It)

Kong Logo
EnterpriseMay 2, 2025

I'm commonly asked by customers for advice on how they can build a good platform cross-charging model for their organization. And my gut reaction is nearly always "don't." We'll come back to why I think that later, but first let's look at what cross

Steve Roberts

Ready to see Kong in action?

Get a personalized walkthrough of Kong's platform tailored to your architecture, use cases, and scale requirements.

Get a Demo
Powering the API world

Increase developer productivity, security, and performance at scale with the unified platform for API management, AI gateways, service mesh, and ingress controller.

Sign up for Kong newsletter

Platform
Kong KonnectKong GatewayKong AI GatewayKong InsomniaDeveloper PortalGateway ManagerCloud GatewayGet a Demo
Explore More
Open Banking API SolutionsAPI Governance SolutionsIstio API Gateway IntegrationKubernetes API ManagementAPI Gateway: Build vs BuyKong vs PostmanKong vs MuleSoftKong vs Apigee
Documentation
Kong Konnect DocsKong Gateway DocsKong Mesh DocsKong AI GatewayKong Insomnia DocsKong Plugin Hub
Open Source
Kong GatewayKumaInsomniaKong Community
Company
About KongCustomersCareersPressEventsContactPricing
  • Terms•
  • Privacy•
  • Trust and Compliance•
  • © Kong Inc. 2025