Back to all scenarios
Scenario #460
Scaling & Load
Kubernetes v1.26, Azure AKS

Autoscaler Fails to Handle Node Termination Events Properly

Autoscaler did not handle node termination events properly, leading to pod disruptions.

Find this helpful?
What Happened

When nodes were terminated due to failure or maintenance, the autoscaler failed to replace them quickly enough, leading to pod disruption.

Diagnosis Steps
  • 1Checked autoscaler logs and found that termination events were not triggering prompt scaling actions.
  • 2Node failure events showed that the cluster was slow to react to node loss.
Root Cause

Autoscaler was not tuned to respond quickly enough to node terminations.

Fix/Workaround
• Configured the autoscaler to prioritize the immediate replacement of terminated nodes.
• Enhanced the health checks to better detect node failures.
Lessons Learned

Autoscalers must be configured to respond quickly to node failure and termination events.

How to Avoid
  • 1Implement tighter integration between node health checks and autoscaling triggers.
  • 2Ensure autoscaling settings prioritize quick recovery from node failures.