Back to Blog

You Already Paid for the Lab

Spare HA capacity is not waste — it is optionality.

By Catalin Lichi · Sugau Infrastructure


Every bare-metal Kubernetes cluster built for production carries redundant capacity by design. N+1 nodes. Standby storage. Uplinks with headroom. This is not inefficiency — it is correct architecture. Without it, a single node failure takes down your workload and you have failed at infrastructure basics.

But here is what the cloud vendors will never tell you: that spare capacity you are paying for anyway can run real workloads. Experimental services. Internal tools. Developer sandboxes. Early-stage products. Things that would cost you real money to spin up in AWS or GCP are, on your bare-metal cluster, essentially free at the margin.

You have already paid for the rack, the power, the cooling, and the uplink bandwidth. The marginal cost of an additional workload is close to zero.


The cloud model charges you for curiosity

In cloud, every experiment has a price tag. A new service needs a cluster, a load balancer, a database instance, storage, and egress costs the moment it generates traffic. Before a single line of code is validated, you are filling out budget approval forms and watching your AWS bill climb.

This creates a hidden tax on engineering culture. Teams learn to avoid experiments they cannot justify on a spreadsheet. Developers stop proposing side projects. Innovation narrows to whatever fits the current sprint roadmap and the current cloud budget.

On bare metal, that tax disappears. If your HA headroom has capacity, you deploy. No approval. No incremental cost. The infrastructure is already running.


What you can run on headroom capacity

The scope is broader than most teams initially consider:

Developer sandboxes. Give engineers a real Kubernetes namespace with real networking and real compute. Not a local Minikube instance with 4GB of RAM — actual cluster resources, with ingress, DNS, and persistent storage. The quality of work that comes out of a proper sandbox versus a hobbled local environment is measurable.

Internal tooling. Dashboards, automation scripts, internal APIs, monitoring extensions. Things your team builds for themselves but never prioritises because there is no place to put them. Now there is.

Experimental services. Early-stage ideas that need to run against real data, handle real HTTP traffic, and prove themselves before they deserve a production namespace. Let them live on headroom. If they die, nothing is lost. If they succeed, you promote them.

AI-assisted development. With current code generation tooling, a single developer can stand up a functional service in hours rather than weeks. The bottleneck is no longer the coding — it is having somewhere to run the result. Headroom capacity removes that bottleneck entirely.

Free or community-facing services. If your business model allows it, you can run public-facing services on this capacity. You have already paid for the bandwidth. A well-architected free tier can be a meaningful marketing and community asset at near-zero marginal cost.


This is not “running loose” — it is governed innovation

There is a version of this pitch that gets rejected immediately in regulated or security-conscious environments, and it sounds like: “just let developers do whatever they want on spare hardware.”

That is not what this is.

Your Kubernetes cluster already has RBAC, network policies, namespace isolation, and audit logging. Experimental workloads run in dedicated namespaces with defined resource quotas. They sit on isolated VLANs or behind network policies that prevent lateral movement to production. The blast radius is defined and bounded before anything is deployed.

This is actually a stronger security posture than most cloud development environments, where a developer account often has IAM permissions far broader than intended and the boundary between dev and prod is a naming convention rather than a network control.

Governed innovation capacity means: fast, frictionless experimentation within a defined perimeter. Not chaos. Not a free-for-all. A structured sandbox with real guardrails that happens to cost you nothing extra to operate.


One honest caveat

If an experimental service succeeds and starts attracting real traffic, it will consume headroom that exists to protect your production workloads. You need to notice this before it becomes a problem.

The answer is capacity planning, not panic. Monitor resource utilisation across your cluster. Set alerts when headroom drops below your HA threshold. When an experiment graduates from “side project” to “real service,” have the conversation about whether it deserves its own hardware, a cloud burst node, or a dedicated node pool.

The headroom model works well for low-to-moderate traffic experiments. It is not a permanent free ride for a service that has found product-market fit. Success is a good problem to have, and it has a straightforward solution.


The reframe

Cloud vendors positioned bare metal as rigid and inflexible. The pitch was: cloud gives you elasticity and speed, bare metal locks you in.

The reality is that a correctly architected bare-metal cluster gives you something cloud cannot: infrastructure you have already paid for, with zero marginal cost for incremental workloads. Cloud elasticity is real, but you pay for every unit of it. Bare-metal headroom is elastic in a different direction — it is capacity that sits ready, costs nothing to use, and is available the moment you need it.

The lab is already running. You might as well use it.


Sugau designs and operates bare-metal Kubernetes infrastructure for organisations that have outgrown the cloud economics model. If your team is ready to stop paying for capacity you do not control, get in touch.