Skip to content

On-Premises Kubernetes

Run Kubernetes where the cloud can’t go.

Not every workload belongs in the cloud. Data sovereignty, latency requirements, regulatory constraints, and economics drive enterprises to run Kubernetes on their own hardware. This track covers everything from datacenter planning to day-2 operations — the knowledge that most free resources skip because it’s not glamorous, but it’s where a massive share of production Kubernetes actually runs.


Planning & Economics (5 modules)
Bare Metal Provisioning (4 modules)
├── Networking (6 modules)
├── Storage (5 modules)
└── Multi-Cluster & Platform (5 modules)
Security & Compliance (8 modules)
Day-2 Operations (9 modules)
Resilience & Migration (3 modules)
AI/ML Infrastructure (6 modules)
SectionModulesFocus
Planning & Economics5Server sizing, cluster topology, TCO, cloud vs on-prem, FinOps & chargeback
Bare Metal Provisioning4PXE, MAAS, Talos, Sidero/Metal3
Networking6Spine-leaf, BGP, MetalLB, DNS/certs, cross-cluster, service mesh
Storage5Ceph/Rook, local storage, object storage (MinIO), database operators
Multi-Cluster & Platform5vSphere/OpenStack, vCluster/Kamaji, Cluster API, fleet management, active-active
Security & Compliance8Air-gapped, HSM/TPM, AD/LDAP, SPIFFE, Vault, policy-as-code, zero-trust
Day-2 Operations9Upgrades, firmware, observability, capacity, self-hosted CI/CD & registry, serverless
Resilience & Migration3Multi-site DR, hybrid connectivity, cloud repatriation
AI/ML Infrastructure6GPU nodes, private training, LLM serving, MLOps, AIOps, HPC storage

51 modules total (30 existing + 21 new from #197). From “should we go on-prem?” to “how do we train LLMs on our own GPUs.”


  • Infrastructure engineers building private Kubernetes platforms
  • Platform teams evaluating on-prem vs cloud for their organization
  • SREs operating bare metal or private cloud Kubernetes clusters
  • Architects designing multi-site, air-gapped, or hybrid environments
  • Budget owners calculating TCO and making build-vs-buy decisions