Back to all scenarios
Scenario #500
Scaling & Load
Kubernetes v1.21, Azure Kubernetes Service (AKS)

Pod Evictions Due to Resource Starvation After Scaling

After scaling up the deployment, resource starvation led to pod evictions, resulting in service instability.

Find this helpful?
What Happened

Scaling events resulted in pod evictions due to insufficient resources on nodes to accommodate the increased pod count.

Diagnosis Steps
  • 1Checked eviction logs and identified that the eviction was triggered by resource pressure, particularly memory.
  • 2Reviewed node resources and found that they were under-provisioned relative to the increased pod demands.
Root Cause

Lack of sufficient resources (memory and CPU) on nodes to handle the scaled deployment.

Fix/Workaround
• Increased the size of the node pool to accommodate the new pod workload.
• Adjusted pod memory requests and limits to prevent overcommitment.
Lessons Learned

Properly provisioning nodes for the expected workload is critical, especially during scaling events.

How to Avoid
  • 1Regularly monitor and analyze resource usage to ensure node pools are adequately provisioned.
  • 2Adjust pod resource requests and limits based on scaling needs.