external-snapshotter
external-snapshotter copied to clipboard
v6.0.1 controller seems to break k8s e2e tests
See https://testgrid.k8s.io/sig-storage-kubernetes#volume-snapshot. The failures started after https://github.com/kubernetes/kubernetes/pull/110204 was merged, which updated the controller from v4 to v6.
Failures for the mock csi driver tests are due to the VolumeSnapshot not being ready, but the snapshot-controller logs show that the VolumeSnapshotContents are being marked as ready. Maybe something changed between v4 and v6 in the processing somehow?
I can continue to dig but if anyone else has any ideas on what's going on please chime in!
/assign /cc @humblec /cc @xing-yang
(@xing-yang it appears the RBAC problem you saw is a read herring, it's happening when things are shut down and there's a race between the controller getting killed and the rbac getting destroyed. The reason for the test failure seems to be this snapshot contents vs snapshot not ready problem)
In v6, v1beta1 is no longer served. We've updated snapshot CRDs and snapshot-controller to v6. We need to update the sidecar images: https://github.com/kubernetes/kubernetes/tree/master/test/e2e/testing-manifests/storage-csi
Ah... I was looking for error messages around that, but I guess the tests just aren't going to find the new resources.
@mattcary @xing-yang shall I drop a PR to fix the same?
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.
This bot triages issues and PRs according to the following rules:
- After 90d of inactivity,
lifecycle/staleis applied - After 30d of inactivity since
lifecycle/stalewas applied,lifecycle/rottenis applied - After 30d of inactivity since
lifecycle/rottenwas applied, the issue is closed
You can:
- Mark this issue or PR as fresh with
/remove-lifecycle stale - Mark this issue or PR as rotten with
/lifecycle rotten - Close this issue or PR with
/close - Offer to help out with Issue Triage
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale
Here's an issue to track this: https://github.com/kubernetes/kubernetes/issues/112694
/remove-lifecycle stale
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.
This bot triages issues and PRs according to the following rules:
- After 90d of inactivity,
lifecycle/staleis applied - After 30d of inactivity since
lifecycle/stalewas applied,lifecycle/rottenis applied - After 30d of inactivity since
lifecycle/rottenwas applied, the issue is closed
You can:
- Mark this issue or PR as fresh with
/remove-lifecycle stale - Mark this issue or PR as rotten with
/lifecycle rotten - Close this issue or PR with
/close - Offer to help out with Issue Triage
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale
/remove-lifecycle stale
/close
Tests are passing now
@mattcary: Closing this issue.
In response to this:
/close
Tests are passing now
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.