Back to all scenarios
Scenario #43
Cluster Management
K8s v1.24, GKE
Node Pool Scaling Impacting StatefulSets
StatefulSet pods were rescheduled across different nodes, breaking persistent volume bindings.
Find this helpful?
What Happened
Node pool scaling in GKE triggered a rescheduling of StatefulSet pods, breaking persistent volume claims that were tied to specific nodes.
Diagnosis Steps
- 1Observed failed to bind volume errors.
- 2Checked StatefulSet configuration for node affinity and volume binding policies.
Root Cause
Lack of proper node affinity or persistent volume binding policies in StatefulSet configuration.
Fix/Workaround
• Added proper node affinity rules and volume binding policies to StatefulSet.
• Rescheduled the pods successfully.
Lessons Learned
StatefulSets require careful management of node affinity and persistent volume binding policies.
How to Avoid
- 1Use pod affinity rules for StatefulSets to ensure proper scheduling and volume binding.
- 2Monitor volume binding status when scaling node pools.