Cannot support existingClaim and different pvc for different vmstorage pods at the same time
Hi Team, I'm trying to setup two VictoriaMetrics clusters in two GCP zones separately.
For the primary cluster, I set existingClaim to empty string to let the statefulset use volumeClaimTemplates.
For the secondary cluster which is used in case of the primary zone is down, I try to make it use the PVC and PV of the primary cluster, but it seems impossible now. If I set existingClaim, all vmstorage pods will use the same PVC, that is not what I want.
Hello,
I try to make it use the PVC and PV of the primary cluster, but it seems impossible now. If I set existingClaim, all vmstorage pods will use the same PVC, that is not what I want.
You want the secondary cluster to use the PVC and PV of the primary cluster, then you don't want them to use the same PVC. Got me confused, could you elaborate?
Sorry for the confusion. Say the primary cluster has 4 vmstorage pods, and their PVC are vmstorage-0, vmstorage-1, vmstorage-2, vmstorage-3. Then I want the secondary cluster's vmstorage pods use these PVC respectively. If I set existingClaim to vmstorage-0, all four pods are using the same PVC which is not I want.
I got a workaround for this issue, though a bit ugly.
- I updated this line to:
name: {{ .Values.vmstorage.persistentVolume.name | default "vmstorage-volume" }}
so that I can customize the PVC name in the values.yaml.
-
For the primary cluster, I set the
vmstorage.persistentVolume.nametovmstorage-volume-standby, so the actual PVC name isvmstorage-volume-standby-victoria-metrics-cluster-vmstorage-{0/1/2/3} -
For the secondary cluster names as
standby-victoria-metrics-cluster, I set thevmstorage.persistentVolume.nametovmstorage-volume, so the PVC name will be the same as the primary cluster's respectively
The reason I try to reuse the PVCs is that I want to use GCP regional persistent disk to improve availability
Oh, okay, I'm afraid you can't do that, two vmstorage process can't use the same volume as their storage. First of all, vmstorage will try to create a flock.lock file under "/vm-data" to make sure it's the only process has exclusive access to "/vm-data", and will panic if it can't acquire the lock. Even if we remove this check, normally you shouldn't have two different processes to write&read the same volume, it causes data duplicated and conflicted. And if one process causes data corruption, you lose two of them, then there is no sense to call it HA.
@Haleygo Thanks for your reply. I understand your concern.
I should have pointed out that the secondary cluster will not write&read to the same volume with the primary cluster at the same time, it is only for disaster recovery. In other words, the secondary cluster will be started when we find the primary cluster is unavailable and not working anymore.
So you have two different available zone or region, the first vmcluster deployed on zoneA, second one on zoneB and components on zoneB are all stopped(replicas=0) at first. If zoneA is completely down, you can increase zoneB vmcluster's replicas(it's the same as creating a new vmcluster) and use zoneB. But if zoneA is not completely down, just several vmstorage nodes are down, e.g storage0 and storage2, in this case, you can't use zoneB to help recovering, since you will get double storage1 and storage3. Am I understand correctly?
If so, then every method that generates same persistentVolume name for different vmcluster will work, like yours here. And I don't see how we can improve chart for this particular case.
Correct, that's what I mean.
Maybe I can raise a PR to make the vmstorage.persistentVolume.name customizable so that I don't need to host the chart separately? Thanks
Maybe I can raise a PR to make the
vmstorage.persistentVolume.namecustomizable so that I don't need to host the chart separately? Thanks
yeah, feel free)
You can also try using the k8s stack chart, VMClusterSpec.VMStorage.storage supports customized VolumeClaimTemplate name as well.
I created a PR https://github.com/VictoriaMetrics/helm-charts/pull/939
The request has been released in victoria-metrics-cluster-0.11.14, close as completed.
Hi @chenlujjj, @Haleygo I have a similar use case. While I can successfully configure a custom name for volumeClaimTemplates using vmstorage.persistentVolume.name, there's an inconsistency in how this name is applied throughout the rendered manifest.
Using the latest helm chart, I'm able to configure a custom name for volumeClaimTemplates, and the volume name and PVC names are changed as expected in the rendered vmstorage StatefulSet pod manifest. However, the volume name reference in volumeMounts is not updated and still uses the default vmstorage-volume.
For example, in values, if I configure vmstorage.persistentVolume.name: restored-vmstorage-volume, the rendered volumeClaimTemplate section in StatefulSet is correct as follows:
volumeClaimTemplates:
- metadata:
name: restored-vmstorage-volume
Consequently, in the provisioned pod, the volume is configured with the name restored-vmstorage-volume as expected:
volumes:
- name: restored-vmstorage-volume
persistentVolumeClaim:
claimName: restored-vmstorage-volume-vmstorage-0
However, the volumeMounts section still references the old name:
volumeMounts:
- mountPath: /storage
name: vmstorage-volume
This causes the deployment to fail with a "volume vmstorage-volume not found" error.
The issue could be resolved by updating the volumeMounts section to use the custom name specified in vmstorage.persistentVolume.name, which is restored-vmstorage-volume in this example.
@Haleygo Created this issue to resolve this.