Back to all scenarios
Scenario #467
Scaling & Load
Kubernetes v1.24, Azure AKS
Node Resource Exhaustion Due to High Pod Density
Node resource exhaustion occurred when too many pods were scheduled on a single node, leading to instability.
Find this helpful?
What Happened
During scaling events, pods were scheduled too densely on a single node, causing resource exhaustion and instability.
Diagnosis Steps
- 1Reviewed node resource utilization, which showed that the CPU and memory were maxed out on the affected nodes.
- 2Pods were not distributed evenly across the cluster.
Root Cause
Over-scheduling pods on a single node during scaling events caused resource exhaustion.
Fix/Workaround
• Adjusted pod affinity rules to distribute pods more evenly across the cluster.
• Increased the number of nodes available to handle the pod load more effectively.Lessons Learned
Resource exhaustion can occur if pod density is not properly managed across nodes.
How to Avoid
- 1Use pod affinity and anti-affinity rules to control pod placement during scaling events.
- 2Ensure that the cluster has enough nodes to handle the pod density.