Back to all scenarios
Scenario #43
Cluster Management
K8s v1.24, GKE

Node Pool Scaling Impacting StatefulSets

StatefulSet pods were rescheduled across different nodes, breaking persistent volume bindings.

Find this helpful?
What Happened

Node pool scaling in GKE triggered a rescheduling of StatefulSet pods, breaking persistent volume claims that were tied to specific nodes.

Diagnosis Steps
  • 1Observed failed to bind volume errors.
  • 2Checked StatefulSet configuration for node affinity and volume binding policies.
Root Cause

Lack of proper node affinity or persistent volume binding policies in StatefulSet configuration.

Fix/Workaround
• Added proper node affinity rules and volume binding policies to StatefulSet.
• Rescheduled the pods successfully.
Lessons Learned

StatefulSets require careful management of node affinity and persistent volume binding policies.

How to Avoid
  • 1Use pod affinity rules for StatefulSets to ensure proper scheduling and volume binding.
  • 2Monitor volume binding status when scaling node pools.