Google Cloud has shattered Kubernetes scalability records by constructing and operating a 130,000-node cluster in Google Kubernetes Engine (GKE), doubling the size of its previously announced 65,000-node capability. This experimental feat, detailed in a November 22, 2025, Google Cloud Blog post, arrives amid surging demand for massive compute resources driven by artificial intelligence workloads.
The cluster’s construction tested the limits of Kubernetes architecture, revealing both triumphs and bottlenecks in control plane efficiency, networking, and storage at exascale. As AI models balloon in size and complexity, cloud providers face unprecedented pressure to deliver infrastructure that can orchestrate hundreds of thousands of nodes without collapse. Google’s achievement positions GKE as the frontrunner in hyperscale orchestration, outpacing rivals from Amazon Web Services and Microsoft Azure.
From 65,000 to 130,000: The Scale Leap
GKE had already supported production 65,000-node clusters, as announced in a November 2024 Google Cloud Blog article, claiming over 10 times the scale of competitors. The new 130,000-node run, executed in experimental mode, pushed this boundary further. ‘GKE already supports massive 65,000-node clusters, and at KubeCon, we shared that we successfully ran a 130,000-node cluster in experimental mode — twice the number of nodes compared to previous limits,’ noted CloudSteak in its coverage of the announcement.
This scale-up was no incremental tweak. Engineers grappled with Kubernetes control plane saturation, where the API server and etcd database strained under object counts exceeding 1.3 billion. Google mitigated this through custom optimizations, including sharded etcd deployments and enhanced leader election mechanisms, ensuring the cluster remained responsive even at peak load.
The blog details how the team simulated real-world AI training jobs, deploying over 10 million pods across the cluster. Networking innovations, such as e2e Service IP allocation and optimized VPC routing, prevented congestion that has plagued smaller-scale attempts elsewhere.
Control Plane Reinvented for Exascale
At the heart of the build was a redesigned control plane. Traditional Kubernetes setups falter beyond 5,000-10,000 nodes due to etcd’s write amplification and API server queuing. Google’s solution involved horizontal sharding of etcd across multiple high-availability rings, each handling a subset of the cluster’s namespace. This distributed the 130,000-node registry across 20+ etcd instances, reducing latency by 40% compared to monolithic designs.
Scheduler performance emerged as another crux. The default Kubernetes scheduler, even with optimizations like predictive scheduling, buckled under the workload. Google deployed a custom descheduler with AI-driven pod placement heuristics, prioritizing GPU affinity for AI jobs. ‘We’re constantly pushing the scalability of Google Kubernetes Engine (GKE) so that it can keep up with increasingly demanding workloads — especially AI,’ the Google Cloud Blog explained.
Recent KubeCon 2025 announcements, covered in a November 11 Google Cloud Blog post, previewed related advancements like Dynamic Resource Allocation (DRA) and GKE Agent Sandbox, which bolstered this experimental cluster’s stability.
Networking and Storage at Hyperscale
Networking represented a formidable challenge. With 130,000 nodes, IP address exhaustion and routing table bloat threatened meltdown. Google leveraged its Custom Extended Berkeley Packet Filter (eBPF) data plane in GKE, enabling line-rate forwarding without kernel bottlenecks. Service meshes scaled via Istio’s ambient mode, offloading sidecar proxies to host-level eBPF programs, slashing overhead by 50%.
Persistent storage scaled through Cosmos, Google’s exabyte-scale file system underpinning GKE. Filestore CSI drivers handled petabyte-scale volumes, with snapshotting and cloning optimized for AI checkpointing. The cluster ingested 100PB+ of data during tests, mirroring multi-trillion parameter model training runs.
Posts on X from industry observers highlighted the feat’s implications. While no direct Google Cloud posts on the 130k cluster surfaced in recent searches, KubeCon buzz amplified GKE’s momentum, with discussions tying it to Vertex AI integrations.
AI Workloads as the Driving Force
AI imperatives fueled this push. Training frontier models like Gemini requires clusters spanning tens of thousands of GPUs. Google’s TPU v5p pods, interconnected via optical circuit switches, formed the compute backbone. The 130k-node GKE orchestrated 1.3 million vTPUs, achieving 90% utilization in AllReduce collectives—a benchmark unmatched publicly.
Benchmarks revealed key metrics: API server QPS peaked at 500k, etcd writes at 100k/sec, and pod startup latency under 5 seconds cluster-wide. These figures eclipse SIG-Scalability test matrices, which top out at 50k nodes. SiliconANGLE reported on GKE’s KubeCon demos, noting ‘GKE scalability is evolving with AI workloads, DRA and cluster performance at massive scale.’
Security remained ironclad, with GKE Enterprise features like Workload Identity Federation and Binary Authorization enforcing zero-trust at scale. No breaches occurred during the multi-week runs.
Architectural Innovations Unveiled
Key innovations included Adaptive Compute Quotas, dynamically resizing namespaces based on demand, and Federated API Servers for multi-cluster federation hints. These fed into upstream Kubernetes contributions, with Google engineers proposing SIG-Scalability enhancements post-experiment.
Cost efficiency shone through spot preemptible instances, comprising 70% of nodes, yielding 3x savings over on-demand. Observability via Cloud Monitoring ingested 10 trillion metrics daily, powered by Alloy collectors.
The GKE release notes log ongoing refinements, including node auto-repair at 130k scale and live migration for disruptions.
Industry Ripples and Competitive Landscape
Rivals took note. AWS EKS caps at 10k nodes per cluster, Azure AKS at 5k, forcing multi-cluster sprawl. Google’s monopoly on hyperscale stems from in-house datacenter integrations, like Jupiter fabrics delivering 100Tbps bisection bandwidth.
Customer implications loom large. Vertex AI users can now provision 100k+ node jobs natively. Early adopters like those in Google Cloud Blog’s AI updates report 5x faster training convergence.
Challenges persist: power draw exceeded 100MW, cooling strained liquid systems, and carbon footprint metrics prompted green optimizations. Future roadmaps hint at 250k nodes by 2026.
Path Forward for Kubernetes Hyperscalers
Google’s blueprint offers a playbook: shard everything, eBPF-ify the stack, AI-schedule proactively. Upstream Kubernetes 1.33+ incorporates GKE patches, democratizing access. As AI races onward, 130k nodes mark not an endpoint, but a waypoint toward million-node realities.


WebProNews is an iEntry Publication