Back to all scenarios
Scenario #467
Scaling & Load
Kubernetes v1.24, Azure AKS

Node Resource Exhaustion Due to High Pod Density

Node resource exhaustion occurred when too many pods were scheduled on a single node, leading to instability.

Find this helpful?
What Happened

During scaling events, pods were scheduled too densely on a single node, causing resource exhaustion and instability.

Diagnosis Steps
  • 1Reviewed node resource utilization, which showed that the CPU and memory were maxed out on the affected nodes.
  • 2Pods were not distributed evenly across the cluster.
Root Cause

Over-scheduling pods on a single node during scaling events caused resource exhaustion.

Fix/Workaround
• Adjusted pod affinity rules to distribute pods more evenly across the cluster.
• Increased the number of nodes available to handle the pod load more effectively.
Lessons Learned

Resource exhaustion can occur if pod density is not properly managed across nodes.

How to Avoid
  • 1Use pod affinity and anti-affinity rules to control pod placement during scaling events.
  • 2Ensure that the cluster has enough nodes to handle the pod density.