← All Articles

Large-Scale Cluster Management at Google with Borg

Large-Scale Cluster Management at Google with Borg by Abhishek Verma, Luis Pedrosa, Madhukar Korupolu, David Oppenheimer, Eric Tune, and John Wilkes is one of the foundational articles on large-scale infrastructure operations. It explains how Google designed and ran a cluster management system capable of supporting massive workloads with high efficiency, reliability, and control.

This article matters because it shows how Google handled infrastructure at a scale where ordinary scheduling and operations practices were no longer enough. Borg was built to admit jobs, place them on machines, restart them after failure, monitor them continuously, and keep services available across clusters that could contain tens of thousands of machines. What makes the paper stand out is that it does not simply describe a system. It explains the thinking behind it, where reliability, efficiency, and scale are treated as one connected problem.

One of the most important ideas in the article is that large infrastructure should not waste capacity simply because workloads behave differently. Borg runs long-lived production services alongside lower-priority batch work, and the paper shows why that matters. Instead of separating these workloads into completely isolated environments, Borg uses scheduling, priorities, and isolation mechanisms so that spare capacity can be used rather than left idle. In the paper’s evaluation, separating production and non-production work into different cells would have required about 20–30% more machines in the median case.

Another especially interesting point is resource reclamation. This is one of the clearest examples in the paper of how careful platform design can create major gains. Many workloads request more CPU or memory than they actually use most of the time, either for safety or to cover rare spikes. Borg measures actual usage, estimates what a task really needs, and reclaims the unused portion for lower-priority work. The important detail is that production tasks remain protected, while batch tasks make use of capacity that would otherwise sit unused. The paper shows that without this mechanism, many more machines would be needed, and that about 20% of the workload in a median cell ran in reclaimed resources.

One of the strongest parts of the paper is its very practical view of failure. Machines fail, tasks crash, agents lose contact, and networks split. Borg was designed around that reality. It automatically reschedules work, spreads tasks across failure domains to reduce correlated failures, and keeps already-running tasks alive even if the master or local agent goes down. The paper reports that Borgmaster achieved about 99.99% availability in practice. That detail alone says a lot about the gap between a system that looks good on paper and one that survives real production pressure over many years.

Another valuable lesson in the article is that orchestration is much more than starting processes. Borg included a declarative job specification model, stable naming, health checking, monitoring, and tools that helped users understand why jobs were pending or how they were behaving at runtime. In other words, the platform was designed not only to run workloads, but also to make them observable and manageable. That is one reason the paper still feels modern. It describes a complete operational system, not just a scheduler.

The paper also challenges a few assumptions that seem reasonable at first glance. Fixed resource buckets, for example, sound simple, but Borg’s workload did not fit neatly into them. The evaluation showed that rounding CPU and memory requests into bucketed sizes would have required around 30–50% more resources in the median case. That is a strong reminder that convenience in platform design can become waste at scale. Fine-grained resource control was not a small detail in Borg; it was one of the reasons the system could operate so efficiently.

What stands out most in this article is how grounded it is in real operations. Borg is not presented as a perfect system. It is presented as a set of trade-offs tested in production over many years: sharing versus isolation, efficiency versus safety margins, centralized control versus resilience, and simplicity for users versus complexity inside the platform itself. That is exactly why the paper remains worth reading. It gives a clear picture of what serious infrastructure engineering looks like when scale, cost, and reliability all matter at the same time.

The audio version beside this text was created as a conversation-style overview of the article, designed to make its main ideas easier to follow while keeping the focus on its most important technical and operational points. It highlights the core concepts, trade-offs, and production lessons that make this paper especially worth reading in full.

Audio Overview of the Article