Replies: 5 comments
-
|
When you see the error |
Beta Was this translation helpful? Give feedback.
-
|
yes, that's true, there is a watcher: But the rbd-image is not mapped and mounted on this server. Is there anything else possibly creating the watcher? |
Beta Was this translation helpful? Give feedback.
-
|
Is it possible that the RBD image is mirrored to an other location? A process like |
Beta Was this translation helpful? Give feedback.
-
|
I also encountered the same problem, but sometimes it can automatically mount back after a period of time, and sometimes it keeps getting stuck. I want to investigate whether the problem is in the CEPH cluster or in the CSI plugin. |
Beta Was this translation helpful? Give feedback.
-
|
How to solve it? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Describe the bug
We have a Proxmox cluster with Ceph and k8s running there.
We are using the csi-rbd provider to deploy PVCs.
It is all seems to be working, a pod is started, pvc is created, a rbd image seems to be created, all is mounted and the app inside a pod is able to use it. Until the pod is deleted and recreated. From that moment, it won't start, as the PVC is not ready for the pod:
I think I was able to verify that the volume is really not mounted anywhere. I got a shell inside each
ceph-csi-rbd-nodepluginpod, on each worker node, and from thecsi-rbdplugincontainer I was looking for the rbd device with something likerbd device ls | grep csi-vol-dea80a4e-85b6-46e2-9d08-ffba10ef7825. I double check directly on the worker node OS, by looking at all mount entries. Nothing found.I also tried to mount the device manually. I picked a
ceph-csi-rbd-nodeplugincontainer, prepared ceph and keyring config and I was able to map and mount:rbd -c /tmp/ceph.conf -k /tmp/ceph.keyring --id k8s-dev --pool k8s-dev map csi-vol-dea80a4e-85b6-46e2-9d08-ffba10ef7825, and then mount the block device on the host with like:mount /dev/rbd7 /mnt/. That worked. Now the rbd image is actually being used :)But I can't find out why k8s things it can't be mounted. I need some help here please.
Is there a change that the
storage.kubernetes.io/csiProvisionerIdentitycould play any role in all of this? Does it change when theceph-csi-rbd-provisioneris redeployed?Because I noted that the
csiProvisionerIdentityis different among persistent volumes and I did a couple of redeploys of the provisioner.Environment details
fuseorkernel. for rbd itskrbdorrbd-nbd) :Steps to reproduce
Initphase because ofFailedMount-MountVolume.MountDevice failed for volume "pvc-7e9588ad-19d1-4099-bd49-30fd2b9f0d42" : rpc error: code = Internal desc = rbd image k8s-dev/csi-vol-dea80a4e-85b6-46e2-9d08-ffba10ef7825 is still being usedNothing interesting from logs unfortunately, or at least I have not noticed.
Beta Was this translation helpful? Give feedback.
All reactions