Back to all scenarios
Scenario #403
Scaling & Load
Kubernetes v1.21, on-prem

Overprovisioned Pods Starved the Cluster

Aggressively overprovisioned pod resources led to failed scheduling and throttling.

Find this helpful?
What Happened

Apps were deployed with excessive CPU/memory, blocking HPA and new workloads.

Diagnosis Steps
  • 1kubectl describe node: Insufficient CPU errors.
  • 2Top nodes showed 50% actual usage, 100% requested.
Root Cause

Reserved resources were never used but blocked the scheduler.

Fix/Workaround
• Adjusted requests/limits based on real usage.
Lessons Learned

Resource requests ≠ real consumption.

How to Avoid
  • 1Right-size pods using VPA recommendations or Prometheus usage data.