[英]Kubernetes pods are stuck after scale up AWS. Multi-Attach error for volume
I experiencing some issues when scale down/up ec2 of my k8s cluster.我在缩小/扩大我的 k8s 集群的 ec2 时遇到了一些问题。 It might happen that sometimes I have new nodes, and old are terminated.
有时我有新节点,而旧节点被终止,这可能会发生。 k8s version is 1.22
k8s版本是1.22
Sometimes some pods are in ContainerCreating state. I am trying to describe pod and see something like this:有时一些 pod 在 ContainerCreating state 中。我正在尝试描述 pod 并看到类似这样的内容:
Warning FailedAttachVolume 29m attachdetach-controller Multi-Attach error for volume
Warning FailedMount 33s (x13 over 27m) kubelet....
I am checking that pv exists, pvs exists as well.我正在检查 pv 是否存在,pvs 是否也存在。 However on pvc I see annotation volume.kube.netes.io/selected-node and its value refers to the node that already not exist.
但是在 pvc 上我看到注释volume.kube.netes.io/selected-node并且它的值指的是已经不存在的节点。
When I am editing the pvc and deleting this annotation, everything continue to work.当我编辑 pvc 并删除这个注释时,一切都继续工作。 Another thing that It happens not always, I don't understand why.
另一件事它并不总是发生,我不明白为什么。
I tried to search information, found some couple of links我试图搜索信息,找到了一些链接
https://github.com/kube.netes/kube.netes/issues/100485 and https://github.com/kube.netes/kube.netes/issues/89953 however I am not sure that I properly understand this. https://github.com/kube.netes/kube.netes/issues/100485和https://github.com/kube.netes/kube.netes/issues/89953但是我不确定我是否正确理解了这一点。
Could you please helm me out with this.你能帮我解决这个问题吗?
Well, as you found out in volume.kube.netes.io/selected-node never cleared for non-existent nodes on PVC without PVs #100485 - this is a known issue, with no available fix yet.好吧,正如您在volume.kube.netes.io/selected-node 中发现的那样,在没有 PV 的情况下,PVC 上的不存在节点从未被清除 #100485 - 这是一个已知问题,目前尚无可用的修复方法。
Until the issue is fixed, as a workaroud, you need to remove volume.kube.netes.io/selected-node
annotation manually.在问题解决之前,作为解决方法,您需要手动删除
volume.kube.netes.io/selected-node
注释。
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.