Back to all scenarios
Scenario #452
Scaling & Load
Kubernetes v1.25, Google Cloud
Load Balancer Overload After Quick Pod Scaling
Load balancer failed to distribute traffic effectively after a large pod scaling event, leading to overloaded pods.
Find this helpful?
What Happened
Pods were scaled up quickly, but the load balancer did not reassign traffic in a timely manner, causing some pods to receive too much traffic while others were underutilized.
Diagnosis Steps
- 1Investigated the load balancer configuration and found that traffic routing did not adjust immediately after the scaling event.
- 2Noticed uneven distribution of traffic in the load balancer dashboard.
Root Cause
Load balancer was not properly configured to dynamically rebalance traffic after pod scaling.
Fix/Workaround
• Reconfigured the load balancer to automatically adjust traffic distribution after pod scaling events.
• Implemented health checks to ensure that only fully initialized pods received traffic.
Lessons Learned
Load balancers must be able to react quickly to changes in the backend pool after scaling.
How to Avoid
- 1Use auto-scaling triggers that also adjust load balancer settings dynamically.
- 2Implement smarter traffic management for faster pod scale-up transitions.