-
Notifications
You must be signed in to change notification settings - Fork 1.4k
CAPi doesnt wait for CSI volume unbinding #4707
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
@MaxRink what does signal the csi controller to detach the volumes? Should the vSphereMachine controller ensure detach happens gracefully orthogonally to the csi controller? |
The pod eviction, after the pods are stopped the csi controller will detach the voumes so they can be rebound on another node. |
I'm not sure this is really a CAPI bug, given that CAPI is not aware of the type of storage providers in use in each cluster. |
generally speaking I think CAPI should ensure that volumes are properly detached before deleting the machines. That can be done regardless of the storage provider e.g. @jzhoucliqr. @randomvariable - IIRC CAPA do not have this problem as non-root volumes are detached and preserved on instance termination right ? |
Its not only the vsphere CSI which might have that issue btw. |
@yastij Are you thinking about inspecting a node and relative Volumes' status? |
/lifecycle awaiting-more-evidence |
/milestone v0.4 |
What steps did you take and what happened:
Ive rolled my workers which had volumes provisioned by the vsphere CSI attached.
Some of those volumes did not detach properly as CAPI was too fast and removed the nodes before the csi controller could fully detach the volumes.
What did you expect to happen:
CAPI waits until volumes are detached.
Anything else you would like to add:
We had a small discussion in the CAPV slack ( https://kubernetes.slack.com/archives/CKFGK3SSD/p1622198292045000 )
@jzhoucliqr has already an patch ( spectrocloud@c340e68 )
Environment:
/kind bug
[One or more /area label. See https://github.com/kubernetes-sigs/cluster-api/labels?q=area for the list of labels]
The text was updated successfully, but these errors were encountered: