Back to all scenarios
Scenario #468
Scaling & Load
Kubernetes v1.25, Google Cloud
Scaling Failure Due to Node Memory Pressure
Pod scaling failed due to memory pressure on nodes, preventing new pods from being scheduled.
Find this helpful?
What Happened
Memory pressure on nodes prevented new pods from being scheduled, even though scaling events were triggered.
Diagnosis Steps
- 1Checked memory utilization and found that nodes were operating under high memory pressure, causing scheduling failures.
- 2Noticed that pod resource requests were too high for the available memory.
Root Cause
Insufficient memory resources on nodes to accommodate the newly scaled pods.
Fix/Workaround
• Increased memory resources on nodes and adjusted pod resource requests to better match available resources.
• Implemented memory-based autoscaling to handle memory pressure better during scaling events.
Lessons Learned
Memory pressure must be monitored and managed effectively during scaling events to avoid pod scheduling failures.
How to Avoid
- 1Ensure nodes have sufficient memory available, and use memory-based autoscaling.
- 2Implement tighter control over pod resource requests and limits.